In recent weeks, tech pundits have been deleting their Facebook accounts and encouraging others to join them. Most are signing off in response to Facebook's ongoing bait-and-switch tactics that leave them not trusting Facebook. All of this is completely fair, but those who are signing off are not representative of the bulk of Facebook's users. And this act of resistance by a few -- or a thousand, or even a million -- is not going to reshape the company's practices. That said, growing public resentment and distrust creates a fantastic opportunity to begin a deeper dialogue about how these issues should play out.
I've been critiquing moves made by Facebook for a long time and I'm pretty used to them being misinterpreted. When I lamented the development of the News Feed, many people believed that I thought that the technology was a failure and that it wouldn't be popular. This was patently untrue. I was bothered by it precisely because I knew that it would be popular, precisely because people love to gossip and learn about others, often to their own detriment. It was hugely disruptive and, when it launched, users lacked the controls necessary to really manage the situation effectively. Facebook responded with controls and people were able to find a way of engaging with Facebook with the News Feed as a given. But people were harmed in the transition.
Last week, I offered two different critiques of the moves made by Facebook, following up on my SXSW talk. Both have been misinterpreted in fascinating ways. Even news agencies are publishing statements like: "Microsoft wants Facebook to be regulated as a utility." WTF? Seriously? Le sigh. (For the record, I'm not speaking on behalf of my employer nor do I want regulation; I think that it's inevitable and I think that we need to contend with it. Oh, and I don't think that the regulation that we'll see will at all resemble the ways in which utilities are regulated. I was talking about utilities because that's how Facebook frames itself. But clearly, most folks missed that.) Misinterpretations are frustrating because they make me feel as though I'm doing a bad job of communicating what I think is important. For this, I apologize to all of you. I will try to do better.
Let me begin by highlighting that I do not think that those who quit Facebook are fools. There are plenty of good reasons to quit Facebook. But people should quit Facebook because they don't believe the service is working for them. They shouldn't quit Facebook because they believe that doing so will send a message to the company. With this in mind, I want to enumerate six beliefs that I have that I want to flesh out in this post in light of discussions about how "everyone" is leaving Facebook:
- I do not believe that people will (or should) leave Facebook en masse because of privacy issues.
- I do not believe that the tech elites who are publicly leaving Facebook will affect on the company's numbers; they are unrepresentative and were not central users in the first place.
- I do not believe that an alternative will emerge in the next 2-5 years that will "replace" Facebook in any meaningful sense.
- I believe that Facebook will get regulated (regardless of whether or not this is a good idea) and I would like to see an open discussion of what this means and what form this takes.
- I believe that a significant minority of users are at risk because of decisions Facebook has made and I think that those of us who aren't owe it to those who are to work through these issues.
- I believe that Facebook needs to start a public dialogue with users and those who are concerned ASAP (What we've seen to date doesn't count).
As I stated in my last post, I think that Facebook plays a central role in the lives of many and I think that it is unreasonable for anyone to argue that they should "just leave" if they're not happy. This is like saying that people should just leave their apartments if they're not happy with their landlord or just leave their spouse because they're not happy with a decision or just leave their job if they're not happy with their boss. Life is more complicated than a series of simplified choices and we are always making calculated decisions, balancing costs and benefits. We stay with our jobs, apartments, and partners even when things get messy because we hope to rectify problems. Many are staying with Facebook because they hope that the company will wise up. And those with the most to gain from Facebook are the least likely to leave, even if they also have the most to lose.
In the last few weeks, a handful of well known digerati have proudly announced that they've departed from Facebook. Most of these individuals weren't that engaged in Facebook as users in the first place. I say this as someone who would lose very little (outside of research knowledge) from leaving. I am not a representative user. I barely share on the site for a whole host of personal and professional reasons. (And because I don't have a life.) None of my friends would miss me if I did leave. In fact, they'd probably be grateful for the disappearance of my re-posted tweets. If I were to leave Facebook, I will have pretty much no impact on the network. This is true for many of the people who I've watched depart. At best, they're content broadcasters. But people have other ways of consuming their broadcasting. So their departure is meaningless. We are not the people that Facebook is worried about losing.
Those most actively engaged in Facebook will not leave Facebook en masse, even if a new site were to emerge, simply because of privacy issues. Realistically, if that were enough, they could go to MySpace or Orkut or Friendster or Tribe. But they won't. And not just because those sites are no longer "cool." They won't because they've invested in Facebook and they're still hoping that Facebook will get its act together. Changing services is costly, just like moving apartments or changing jobs or breaking up in general. The deeper the relationship, the harder it is to simply walk away. And the relationship that Facebook has built with many of its users is very very very deep. When transition costs are high, people work hard to change the situation so that they don't have to transition. This is why people are complaining, this is why they are speaking up. And it's really important that those in power listen to what it is that people are upset about. The worst thing that those in power can do is ignore what's going on, waiting for it to go away. This is a bad idea, not because people will walk away, but because they will look to greater authorities of power to push back. This is why Facebook's failure to address what's going on invites regulation.
Facebook has gotten quite accustomed to upset users. Too accustomed. In "The Facebook Effect," David Kirkpatrick outlines how Facebook came to expect that every little tweak would set off an internal rebellion. He documented how most of the members of the group "I AUTOMATICALLY HATE THE NEW FACEBOOK HOME PAGE" were employees of Facebook whose frustration with user rebellion was summed up by the group's description: "I HATE CHANGE AND EVERYTHING ASSOCIATED WITH IT. I WANT EVERYTHING TO REMAIN STATIC THROUGHOUT MY ENTIRE LIFE." Kirkpatrick quotes Zuckerberg as saying, "The biggest thing is going to be leading the user base through the changes that need to continue to happen... Whenever we roll out any major product there's some sort of backlash." Unfortunately, Facebook has become so numb to user complaints that it doesn't see the different flavors of them any longer.
What's happening around privacy is not simply user backlash. In fact, users are far less upset about what's going on than most journalists and bloggers. Why? Because even with the New York Times writing article after article, most users have no idea what's happening. I'm reminded of this every time that I sit down with someone who doesn't run in my tech circles. And I'm reminded that they care every time I sit down and walk them through their privacy settings. The disconnect between average users and those in the tech industry is what makes this situation different, what makes this issue messier. Because the issue comes down to corporate transparency, informed consent, and choice. As long as users believe that their content is private and have no idea how public it is, they won't take to the streets. A disappearance of publicity for these issues is to Facebook's advantage. But it's not to user's advantage. Which is precisely why I think that it's important that the bloggers and the journalists keep covering this topic. Because it's important that more people are aware of what's going on. Unfortunately, of course, we also have to contend with the fact that most people being screwed don't speak English and have no idea this conversation is even happening. Especially when privacy features are only explained in English. And it's also important to keep in mind what users do see and where they are screaming. They are really not happy about having to turn their interests into formalized links. Engaging with users on what is most visible to them is the best way to start this conversation.
In documenting Zuckerberg's attitudes about transparency, Kirkpatrick sheds light on one of the weaknesses of his philosophy: Zuckerberg doesn't know how to resolve the positive (and in his head inevitable) outcomes of transparency with the possible challenges of surveillance. As is typical in the American tech world, most of the conversation about surveillance centers on the government. But Kirkpatrick highlights another outcome of surveillance with a throwaway example that sends shivers down my spine: "When a father in Saudi Arabia caught his daughter interacting with men on Facebook, he killed her." This is precisely the kind of unintended consequence that motivates me to speak loudly even though I'm privileged enough to not face these risks. Statistically, death is an unlikely outcome of surveillance. But there are many other kinds of side effects that are more common and also disturbing: losing one's job, losing one's health insurance, losing one's parental rights, losing one's relationships, etc. Sometimes, these losses will be because visibility makes someone more accountable. But sometimes this will occur because of misinterpretation and/or overreaction. And the examples keep on coming.
I am all in favor of people building what they believe to be alternatives to Facebook. I even invested in Diaspora because I'm curious what will come of that system. But I don't believe that Diaspora is a Facebook killer. I do believe that there is a potential for Diaspora to do something interesting that will play a different role in the ecosystem and I look forward to seeing what they develop. I'm also curious about the future of peer-to-peer systems in light of the move towards the cloud, but I'm not convinced that decentralization is a panacea to all of our contemporary woes. Realistically, I don't think that most users around the globe will find a peer-to-peer solution worth the hassle. The cost/benefit analysis isn't in their favor. I'm also patently afraid that a system like Diaspora will be quickly leveraged for child pornography and other more problematic uses that tend to emerge when there isn't a centralized control system. But innovation is important and I'm excited that a group of deeply passionate developers are being given a chance to see what they can pull off. And maybe it'll be even more fabulous than we can possibly imagine, but I'd bet a lot of money that it won't put a dent into Facebook. Alternatives aren't the point.
Facebook has embedded itself pretty deeply into the ecosystem, into the hearts and minds of average people. They love the technology, but they're not necessarily prepared for where the company is taking them. And while I'm all in favor of giving users the choice to embrace the opportunities and potential of being highly visible, of being a part of a transparent society, I'm not OK with throwing them off the boat just to see if they can swim. Fundamentally, my disagreement with Facebook's approach to these matters is a philosophical one. Do I want to create more empathy, more tolerance in a global era? Of course. But I'm not convinced that sudden exposure to the world at large gets people there and I genuinely fear that possible backlash that can emerge. I'm not convinced that this won't enhance a type of extremism that is manifesting around the globe as we speak.
Personally, I believe that screaming about the end of Facebook is futile. And I think that folks are wasting a lot of energy just telling others to quit or boycott simply to send a message that won't be heard as such. It'll just make us technophiles look like we're living on a different planet. Which we are. Instead, I think that we should all be working to help people understand what's going on. Informing people so that they can make their own decisions about what's best for them. I love using Reclaim Privacy to walk through privacy settings with people. While you're helping your family and friends understand their settings, talk to them and record their stories. We would all benefit from hearing average people's stories, their fears, their passions. It's important to hear what privacy means to diverse peoples and why they care about it. We can learn a lot from hearing about the upside and downside of visibility for different people and the challenges introduced by exposure. And I think folks inside Facebook should start listening and stop telling people what they should feel about privacy. Not because this is another user rebellion, but because Facebook's decisions shape the dynamics of so many people's lives. And because how people experience issues related to privacy is much messier and more complex than can be narrated by any individual.
I also want tech folks to think hard and deep about the role that regulation may play and what the consequences may be for all of us. In thinking about regulation, always keep Larry Lessig's arguments in "Code" in mind. Larry argued that there are four points of regulation for all change: the market, the law, social norms, and architecture (or code). Facebook's argument is that social norms have changed so dramatically that what they're doing with code aligns with the people (and conveniently the market). I would argue that they're misreading social norms but there's no doubt that the market and code work in their favor. This is precisely why I think that law will get involved and I believe that legal regulators don't share Facebook's attitudes about social norms. This is not a question of if but a question of when, in what form, and at what cost. And I think that all of us who are living and breathing this space should speak up about how we think this should play out because if we just pretend like it won't happen, not only are we fooling ourselves, but we're missing an opportunity to shape the future.
I realize that Elliot Schrage attempted to communicate with the public through his NYTimes responses. He failed. I was also tremendously disappointed with Mark Zuckberger's op-ed in the Washington Post. In short, he told users that the only mistake he's willing to cop to is making the control settings too complicated. This situation isn't like the previous ones. Facebook needs to stop telling users what they should feel and start engaging with them actively. They've done this in the past. In inviting the users to vote on the changes to the Terms of Service, Facebook mapped out the possibility of networked engagement, of inviting passionate users to speak back and actively listening. This was a huge success for Facebook, even if very few users actually participated in the polls. It was important because people had the choice to participate and those who were most concerned had a mechanism by which they could voice their concerns. Why aren't they doing this now? I find the silence and dismissiveness that's currently happening to be quite eerie. I cannot imagine that Facebook isn't listening. So, Facebook, if you are listening, please start a dialogue with the public rather than broadcasting your philosophy. Please be transparent if you're asking us to be. And please start now, not when you've got a new set of features ready.
Regardless of how the vocal minority feel about Facebook, millions of average people are deeply wedded to the site. They won't leave because it's too important to their social and/or professional lives for them to stay. But that doesn't mean that they aren't suffering because of decisions being made about them and for them. What's at stake now is not whether or not Facebook will become passe, but whether or not Facebook will become evil. I think that we owe it to the users to challenge Facebook to live up to a higher standard, regardless of what we as individuals may gain or lose from their choices. And we owe it to ourselves to make sure that everyone is informed and actively engaged in a discussion about the future of privacy. Zuckerberg is right: "Given that the world is moving towards more sharing of information, making sure that it happens in a bottom-up way, with people inputting their information themselves and having control over how their information interacts with the system, as opposed to a centralized way, through it being tracked in some surveillance system. I think it's critical for the world." Now, let's hold him to it.
Danah Boyd is a Social Media Researcher at Microsoft Research New England and a Fellow at Harvard University's Berkman Center for Internet and Society. Her research examines social media, youth practices, tensions between public and private, social network sites, and other intersections between technology and society. This story originally appeared on her blog.