Anil Dash is giving a Berkman lunchtime talk, titled “The Web We Lost.” He begins by pointing out that the title of his talk implies a commonality that at least once was.
NOTE: Live-blogging. Getting things wrong. Missing points. Omitting key information. Introducing artificial choppiness. Over-emphasizing small matters. Paraphrasing badly. Not running a spellpchecker. Mangling other people’s ideas and words. You are warned, people.
[Light editing on April 3 2013.]
Anil puts up an icon that is a symbol of privately-owned public spaces in New York City. Businesses create these spaces in order to be allowed to build buildings taller than the zoning requirements allow. These are sorta kinda like parks but are not. E.g., Occupy isn’t in Zuccotti Park any more because the space is a privately-own public space, not a park. “We need to understand the distinction” between the spaces we think are public and the ones that are privately owned.
We find out about these when we transgress rules. We expect to be able to transgress in public spaces, but in these privately-owned spaces we cannot. E.g., Improv Everywhere needs to operate anonymously to perform in these spaces. Anil asks us to imagine “a secretive, private ivy league club.” He is the son of immigrants and didn’t go to college. “A space even as welcoming as this one [Harvard Berkman] can seem intimidating.” E.g., Facebook was built as a private club. It welcomes everyone now, but it still doesn’t feel like it’s ours. It’s very hard for a business to get much past its origins.
One result of online privately-owned public spaces is “the wholesale destruction of your wedding photos.” When people lose them in a fire, they are distraught because those photos cannot be replaced. Yet everyday we hear about a startup that “succeeds” by selling out, and then destroying the content that they’d gathered. We’ve all gotten the emails that say: “Good news! 1. We’re getting rich. 2. You’re not. 3. We’re deleting your wedding photos.” They can do this because of the terms of service that none of us read but that give them carte blanche. We tend to look at this as simply the cost of doing business with the site.
But don’t see it that way, Anil urges. “This is actually a battle” against the values of the early Web. In the mid to late 1990s, the social Web arose. There was a time when it was meaningful thing to say that you’re a blogger. It was distinctive. Now being introduced as a blogger “is a little bit like being introduced as an emailer.” “No one’s a Facebooker.” The idea that there was a culture with shared values has been dismantled.
He challenges himself to substantiate this:
“We have a lot of software that forbids journalism.” He refers to the IoS [iphone operating system] Terms of Service for app developers that includes text that says, literally: “If you want to criticize a religion, write a book.” You can distribute that book through the Apple bookstore, but Apple doesn’t want you writing apps that criticize religion. Apple enforces an anti-journalism rule, banning an app that shows where drone strikes have been.
Less visibly, the laws is being bent “to make our controlling our data illegal.” All the social networks operate as common carriers — neutral substrates — except when it comes to monetizing. The boundaries are unclear: I can sing “Happy Birthday” to a child at home, and I can do it over FaceTime, but I can’t put it up at YouTube [because of copyright]. It’s very open-ended and difficult to figure. “Now we have the industry that creates the social network implicitly interested in getting involved in how IP laws evolve.” When the Google home page encourages visitors to call their senators against SOPA/PIPA, we have what those of us against Citizens United oppose: we’re asking a big company to encourage people to act politically in a particular way. At the same time, we’re letting these companies capture our words and works and put them under IP law.
A decade ago, metadata was all the rage among the geeks. You could tag, geo-tag, or machine-tag Flickr photos. Flickr is from the old community. That’s why you can still do Creative Commons searches at Flickr. But you can’t on Instagram. They don’t care about metadata. From an end-user point of view, RSS is out of favor. The new companies are not investing in creating metadata to make their work discoverable and shareable.
At the old Suck.com, hovering on a link would reveal a punchline. Now, with the introduction of Adlinks and AdSense, Google transformed links from the informative and aesthetic, to an economic tool for search engine optimization (SEO). Within less than 6 months, linkspam was spawned. Today Facebook’s EdgeRank is based on the idea that “Likes” are an expression of your intent, which determines how FB charges for ads. We’ll see like-spammers and all the rest we saw with links. “These gestural things that were editorial or indicators of intent get corrupted right away.” There are still little islands, but for the most part these gestures that used to be about me telling you that I like your work are becoming economic actions.
Anil says that a while ago when people clicked on a link from Facebook to his blog, FB popped up a warning notice saying that it might be dangerous to go there. “The assumption is that my site is less trustworthy than theirs. Let’s say that’s true. Let’s say I’m trying to steal all your privacy and they’re not.” [audience laughs] He has FB comments on his site. To get this FB has to validate your page. “I explicitly opted in to the Facebook ecology” in part to prove he’s a moderate and in part as a convenience to his readers. At the same time, FB was letting the Washington Post and The Guardian publish within the FB walls, and FB never gave that warning when you clicked on their links. A friend at FB told Anil that the popup was a bug, which might be. But that means “in the best case, we’re stuck fixing their bugs on our budgets.” (The worst case is that FB is trying to shunt traffic away from other sites.)
And this is true for all things that compete with the Web. The ideas locked into apps won’t survive the company’s acquisition, but this is true when we change devices as well. “Content tied to devices dies when those devices become obsolete.” We have “given up on standard formats.” “Those of us who cared about this stuff…have lost,” overall. Very few apps support standard formats, with jpg and html as exceptions. Likes and follows, etc., all use undocumented proprietary formats. The most dramatic shift: we’ve lost the expectation that they would be interoperable. The Web was built out of interoperability. “This went away with almost no public discourse about the implications of it.”
The most important implication of all this comes when thinking about the Web as a public space. When the President goes on FB, we think about it as a public space, but it’s not, and dissent and transgression are not permitted. “Terms of Service and IP trump the Constitution.” E.g., every single message you put on FB during the election FB could have transformed into its opposite, and FB would be within its ToS rights. After Hurricane Sandy, public relief officials were broadcasting messages only through FB. “You had to be locked into FB to see where public relief was happening. A striking change.”
What’s most at risk are the words of everyday people. “It’s never the Pharaoh’s words that are lost to history.” Very few people opt out of FB. Anil is still on FB because he doesn’t want to lose contact with his in-laws. [See Dan Gillmor's talk last week.) Without these privately-owned public spaces, Anil wouldn't have been invited to Harvard; it's how he made his name.
"The main reason this shift happened in the social web is the arrogance of the people who cared about the social web in the early days...We did sincerely care about enabling all these positive things. But the way we went about it was so arrogant that Mark Zuckerberg's vision seemed more appealing, which is appalling." An Ivy League kid's software designed for a privileged, exclusive elite turned out to be more appealing than what folks like Anil were building. "If we had been listening more, and a little more open in self-criticism, it would have been very valuable."
There was a lot of triumphalism after PIPA/SOPA went down, but it took a huge amount of hyperbole: "Hollywood wants to destroy the First Amendment, etc." It worked once but it doesn't scale. The willingness to pat ourselves on our back uncritically led us to vilify people who support creative industries. That comes from the arrogance that they're dinosaurs, etc. People should see us being publicly critical of ourselves. For something to seem less inclusive than FB or Apple — incredibly arrogant, non-egalitarian cultures — that's something we should look at very self-critically.
Some of us want to say "But it's only some of the Web." We built the Web for pages, but increasingly we're moving from pages to streams (most recently-updated on top, generally), on our phones but also on bigger screens. Sites that were pages have become streams. E.g., YouTube and Yahoo. These streams feel like apps, not pages. Our arrogance keeps us thinking that the Web is still about pages. Nope. The percentage of time we spend online looking at streams is rapidly increasing. It is already dominant. This is important because these streams are controlled access. The host controls how we experience the content. "This is part of how they're controlling the conversation." No Open Web advocate has created a stream that's anywhere near as popular as the sites we're going to. The geeks tend to fight the last battle. "Let's make an open source version of the current thing." Instead, geeks need to think about creating a new kind of stream. People never switch to more open apps. (Anil says Firefox was an exception.)
So, what do we do? Social technologies follow patterns. It's cyclical. (E.g., "mainframes being rebranded as The Cloud.") Google is doing just about everything Microsoft was doing in the late 1990s. We should expect a reaction against their overreach. With Microsoft, "policy really worked." The Consent Decree made IE an afterthought for developers. Public policy can be an important of this change. "There's no question" that policy over social software is coming.
Also, some "apps want to do the right thing." Anil's ThinkUp demonstrates this. We need to be making apps that people actually want, not ones that are just open. "Are you being more attentive to what users want than Mark Zuckerberg is?" We need to shepherd and coach the apps that want to do the right thing. We count on 23 yr olds to do this, but they were in 5th grade when the environment was open. It's very hard to learn the history of the personal software industry and how it impacted culture. "What happened in the desktop office suite wars ?" [Ah, memories!] We should be learning from such things.
And we can learn things from our own data. “It’s much easier for me to check my heart-rate than how often I’m reading Twitter.”
Fortunately, there are still institutions that care about a healthy Web. At one point there was a conflict between federal law and Terms of Service: the White House was archiving coments on its FB wall, whereas FB said you couldn’t archive for more than 24 hrs.
We should remember that ToS isn’t law. Geeks will hack software but treat ToS as sacred. Our culture is negatively impacted by ToS and we should reclaim our agency over them. “We should think about how to organize action around specific clauses in ToS.” In fact, “people have already chosen a path of civil disobedience.” E.g., search YouTube for “no infringement intended.” “It’s like poetry.” They’re saying “I’m not trying to step on your toes, but the world needs to see this.” “I’m so inspired by this.” If millions of teenagers assembled to engage in civil disobedience, we’d be amazed. They do on line. They feel they need to transgress because of a creative urge, or because it’s speech with a friend not an act of publishing. “That’s the opportunity. That’s the exciting part. People are doing this every single day.
[I couldn't capture the excellent Q&A because I was running the microphone around.]
I liked this post by in the Guardian by John Naughton about the future of Web 2.0, and I’m always delighted to be mention in the same paragraph as Paul Graham, but I want to keep insisting that Web 2.0 was not the moment when the Web moved from publishing platform to social platform. One of the main points of Cluetrain (1999) was in fact that the Web from its beginning was thrilling us because it was a social place, a set of conversations, a party.
Now, it is certainly true that with Web 2.0, the Web became more social, easier to socialize in, undeniably social. That’s why Web 2.0 is a useful concept.
My problem is really with the “point” in Web 2 Point Oh, since it can imply a point in time when the Web became social, as if before that the Web was merely a publishing platform. Nah. It’s been social since the moment browsers started appearing.
I’m very happy to say that I’ve been granted a Franklin Fellowship to work with the US State Department for the next year. I’ll be working with the eDiplomacy group that is working on providing Web 2.0 platforms for internal use, with the semi-secret aim of nudging State from a need-to-know to a need-to-share culture. (This is not exactly how eDiplomacy explains its charter, but it’s how I understand it.)
Franklin Fellowships were established by the State Department in 2006 in order to bring in people from the private and non-profit sectors. I’m working as a volunteer, with my travel expenses covered in part by a grant from Craig Newmark, founder of CraigsList. (Thank you, Craig!) Because I’ll be on-site in DC only a few times a month, I’ll be able to continue as a senior researcher at the Berkman Center. (I’ve also begun doing some work for Harvard Law Library’s digital lab.)
I’ve already spent time with the group. They’re, well, wonderful. They’ve already delivered tools for knowledge sharing (e.g., Diplopedia) and for connecting expertise across every boundary (e.g., The Sounding Board), and they’ve got some very interesting projects in the works. These are dedicated State Dept. employees, some with considerable experience under their belts, who are on fire about the possibilities for making State smarter, more innovative and creative, more responsive, more engaged, and more human, but always within the proper security constraints. Fascinating fascinating.
In this five minute interview, done at Supernova, Alec Ross — who reports to Hillary Clinton as Senior Innovation Adviser — talks about how the Internet cadre is doing in the State Department.
[Disclosure: I may have the opportunity to work with the State Dept. (as a volunteer) on the internal use of Web 2.0 tools, pending my getting a security clearance. I believe Alec was instrumental in this. So, thank you Alec. And, of course, that inevitably taints my interview. FWIW, I was a fan of Alec's well before I knew him.]
Terrific post by Euan Semple (responding to a post by Stowe Boyd) about why he does not love the phrase “Enterprise 2.0″: “…it’s too narrow, too corporate and too managerial!”
The name will work itself out, as names do. I have problems with entire “2.0″ meme — I like that it calls attention to important changes, but am uncomfortable about its implication of discontinuity. But, the phrase has stuck, and it has had the advantage of unsticking lots of thinking. The same for “Enterprise 2.0.” I understand Stowe and Euan’s discomfort, but all names are inadequate, and “Enterprise 2.0″ gives some businesses a frame and a justification for thinking about changing. The phrase’s author, Andrew McAfee, probably agrees the name is imperfect, and probably agrees with much of what Euan says about the changes awaiting business. [Disclosure: Andrew is a Berkman Fellow. And Euan, Stowe, and Andrew are all friends of mine. And, while I'm at it, Euan's post positively cites something I once said.]
Beyond Euan’s discussion of the phrase itself, he maintains a Web Exceptionalist and Web Utopian position, albeit he is a Slow Utopian. Not that Euan’s slow. On the contrary. But he believes the changes businesses are going through are deep and will take decades to accomplish. After all, as he says, “‘the Internet has been around for the best part of 30 years and most people don’t know what the back button on their browser is for!”
DBpedia extracts information from Wikipedia, building a database that you can query. This isn’t easy because much of the information in Wikipedia is unstructured. On the other hand, there’s an awful lot that’s structured enough so that an algorithm can reliably deduce the semantic content from the language and the layout. For example, the boxed info on bio pages is pretty standardized, so your algorithm can usually assume that the text that follows “Born: ” is a date and not a place name. As the DBpedia site says:
The DBpedia knowledge base currently describes more than 2.6 million things, including at least 213,000 persons, 328,000 places, 57,000 music albums, 36,000 films, 20,000 companies. The knowledge base consists of 274 million pieces of information (RDF triples). It features labels and short abstracts for these things in 30 different languages; 609,000 links to images and 3,150,000 links to external web pages; 4,878,100 external links into other RDF datasets, 415,000 Wikipedia categories, and 75,000 YAGO categories.
Over time, the site will get better and better at extracting info from Wikipedia. And as it does so, it’s building a generalized corpus of query-able knowledge.
As of now, the means of querying the knowledge requires some familiarity with building database queries. But, the world has accumulated lots of facility with putting front-ends onto databases. DBpedia is working on something differentL accumulating an encyclopedic database, open to all and expressed in the open language of the Semantic Web.
Jeneane’s got a great post for businesses that think they’re playing well in the social media sandbox. She asks: You’re playing, but are you playing jazz?
Pardon the self-bloggery-floggery, but Wired.com has just posted an article of mine that presents Robin “ZipCar” Chase’s argument that the smart grid and smart cars need to be thought about together. Actually, she wants all the infrastructures we’re now building out to adopt open, Net standards, and would prefer that the Internet of Everything be meshed up together. (Time Mag just named Robin as one of the world’s 100 most influential people. We can only hope that’s true.)
The article is currently on Wired’s automotive page, but it may be moved to the main page today or tomorrow.
The Independent calls WolframAlpha “An invention that could change the Internet forever.” It concludes: “Wolfram Alpha has the potential to become one of the biggest names on the planet.”
Nova Spivak, a smart Semantic Web guy, says it could be as important as Google.
Ton Zijlstra, on the other hand, who knows a thing or two about knowledge and knowledge management, feels like it’s been overhyped. After seeing the video of Wolfram talking at Harvard, Ton writes:
No crawling? Centralized database, adding data from partners? Manual updating? Adding is tricky? Manually adding metadata (curating)? For all its coolness on the front of WolframAlpha, on the back end this sounds like it’s the mechanical turk of the semantic web.
(“The mechanical turk of the semantic web.” Great phrase. And while I’m in parentheses, ReadWriteWeb has useful screenshots of WolframAlpha, and here’s my unedited 55-minute interview with Wolfram.)
I am somewhere in between, definitely over in the Enthusiastic half of the field. I think WolframAlpha [WA] will become a standard part of the Internet’s tool set, but is not transformative.
WA works because it’s curated. Real human beings decide what topics to include (geography but not 6 Degrees of Courtney Love), which data to ingest, what metadata is worth capturing, how that metadata is interrelated (= an ontology), which correlations to present to the user when she queries it (daily tonnage of fish captured by the French compared to daily production of garbage in NYC), and how that information should be presented. Wolfram insists that an expert be present in each data stream to ensure the quality of the data. Given all that human intervention, WA then performs its algorithmic computations … which are themselves curated. WA is as curated as an almanac.
Curation is a source of its strength. It increases the reliability of the information, it enables the computations, and it lets the results pages present interesting and relevant information far beyond the simple factual answer to the question. The richness of those pages will be big factor in the site’s success.
Curation is also WA’s limitation. If it stays purely curated, without areas in which the Big Anyone can contribute, it won’t be able to grow at Internet speeds. Someone with a good idea — provide info on meds and interactions, or add recipes so ingredients can be mashed up with nutritional and ecological info — will have to suggest it to WolframAlpha, Inc. and hope they take it up. (You could to this sorta kinda through the API, but not get the scaling effects of actually adding data to the system.) And WA will suffer from the perspectival problems inevitable in all curated systems: WA reflects Stephen Wolfram’s interests and perspective. It covers what he thinks is interesting. It covers it from his point of view. It will have to make decisions on topics for which there are no good answers: Is Pluto a planet? Does Scientology go on the list of religions? Does the page on rabbits include nutritional information about rabbit meat? (That, by the way, was Wolfram’s example in my interview of him. If you look at the site from Europe, a “rabbit” query does include the nutritional info, but not if you log in from a US IP address.) But WA doesn’t have to scale up to Internet Supersize to be supersized useful.
So, given those strengths and limitations, how important is WA?
Once people figure out what types of questions it’s good at, I think it will become a standard part of our tools, and for some areas of inquiry, it may be indispensable. I don’t know those areas well enough to give an example that will hold up, but I can imagine WA becoming the first place geneticists go when they have a question about a gene sequence or chemists who want to know about a molecule. I think it is likely to be so useful within particular fields that it becomes the standard place to look first…Like IMDB.com for movies, except for broad, multiple fields, with the ability to cross-compute.
But more broadly, is WA the next Google? Does it transform the Internet?
I don’t think so. Its computational abilities mean it does something not currently done (or not done well enough for a crowd of users), and the aesthetics of its responses make it quite accessible. But how many computational questions do you have a day? If you want to know how many tons of fish France catches, WA will work as an almanac. But that’s not transformational. If you want to know how many tons divided by the average weight of a French person, WA is for you. But the computational uses that are distinctive of WA and for which WA will frequently be an astounding tool are not frequent enough for WA to be transformational on the order of a Google or Wikipedia.
There are at least two other ways it could be transformational, however.
First, its biggest effect may be on metadata. If WA takes off, as I suspect it will, people and organizations will want to get their data into it. But to contribute their data, they will have to put it into WA’s metadata schema. Those schema then become a standard way we organize data. WA could be the killer app of the Semantic Web … the app that gives people both a motive for putting their data into ontologies and a standardized set of ontologies that makes it easy to do so.
Second, a robust computational engine with access to a very wide array of data is a new idea on the Internet. (Ok, nothing is new. But WA is going to bring this idea to mainstream awareness.) That transforms our expectations, just as Wikipedia is important not just because it’s a great encyclopedia but because it proved the power of collaborative crowds. But, WA’s lesson — there’s more that can be computed than we ever imagined — isn’t as counter-intuitive as Wikipedia’s, so it is not as apple-cart-upsetting, so it’s not as transformational. Our cultural reaction to Wikipedia is to be amazed by what we’ve done. With WA, we are likely to be amazed by what Wolfram has done.
That is the final reason why I think WA is not likely to be as big a deal as Google or Wikipedia, and I say this while being enthusiastic — wowed, even — about WA. WA’s big benefit is that it answers questions authoritatively. WA nails facts down. (Please take the discussion about facts in a postmodern age into the comments section. Thank you.) It thus ends conversation. Google and Wikipedia aim at continuing and even provoking conversation. They are rich with links and pointers. Even as Wikipedia provides a narrative that it hopes is reliable, it takes every opportunity to get you to go to a new page. WA does have links — including links to Wikipedia — but most are hidden one click below the surface. So, the distinction I’m drawing is far from absolute. Nevertheless, it seems right to me: WA is designed to get you out of a state of doubt by showing you a simple, accurate, reliable, true answer to your question. That’s an important service, but answers can be dead-ends on the Web: you get your answer and get off. WA as question-answerer bookends WA’s curated creation process: A relatively (not totally) closed process that has a great deal of value, but keeps it from the participatory model that generally has had the biggest effects on the Net.
Providing solid, reliable answers to difficult questions is hugely valuable. WolframAlpha’s approach is ambitious and brilliant. WolframAlpha is a genius. But that’s not enough to fundamentally alter the Net.