Tuesday, May 13, 2014
What shape is the internet (continued)?
The argument was triggered by @djbressler's observation that some new browsers (including an experimental build of Chrome) were hiding the URL from the user. This is a reflection of the fact that users increasingly type "Amazon" into the browser rather than "amazon.com" let alone "http://www.amazon.com". Presumably, hiding the URL will further encourage this trend.
Google and other search engines appear to benefit from this in two ways. Firstly, it increases the already heavy dependence of the ordinary internet user on the search engine. And secondly, every time an internet user navigates via search rather than via URL or hyperlink, the search engine gets another opportunity to present some advertising, as well as collecting more information about that user.
Obviously, Google itself depends on URLs and hyperlinks. As Nick points out, Google still relies on links to construct its index, and still uses a version of the original PageRank algorithm to influence what you see when you search for a given term. But indexing and search ranking are only loosely coupled to one another.
And nowadays, the search order is not solely determined by PageRank. Instead, the search order is increasingly influenced by browsing behaviour - of others as well as our own. If you ignore the first two items, click briefly on the third item, and then immediately return to Google to look at the fourth item, Google may conclude that the first three items weren't very relevant to you. In other words, this counts as a "vote" against those items.
Meanwhile, Google only had exclusive rights to the original PageRank patent (which belongs to Stanford University) until 2011.
Obviously Google is not completely open about these algorithms, because it is perpetually at war with SEO and spammers who want to get some commercial advantage by "gaming" the system. So there is a degree of speculation involved in working out what exactly Google is up to. Sometimes Google merely seems to appeal to the lowest common denominator, as David Auerbach suggests in his review of Metafilter search results ("Deranked"). However, it is beyond speculation that Google's behaviour has become increasingly sophisticated over the past decade, and that what we see is increasingly "personalized".
Nick accuses me of "confusing the use of behavior IN the ranking algorithm itself with using behavior to verify the quality of the algorithm". However, there is some evidence that Google initially trials new factors in parallel with the existing algorithm, before integrating these factors into the algorithm itself. (See for example, Google Panda.) In any case, the total behaviour of Google can be thought of in terms of the collective intelligence of human brains AND algorithmic software, and it may not be possible for outsider observers to be exactly sure where the boundary lies at any point in time. (We can detect "momentum", but not "position".)
Obviously URLs are not going to disappear entirely. For my part, I have always made an effort to use links and bookmarks rather than pander to the commercial interests and cognitive distortion of search engines. I don't think this undermines my general point - that the Internet-in-use (based on majority habits) is taking on a different shape. Obviously it is still possible to use the Internet in a disciplined and self-conscious manner, which Nick (always) and I (sometimes) practise, but the fact that this requires effort and intelligence makes it likely that it will never become mainstream.
In the long-term, Google may face a paradox. If people stop using URLs, then Google's ability to index and rank pages across the internet might possibly be compromised. But I'm sure that the clever people at Google have thought of this paradox, and already have a cunning plan.
Meanwhile, the internet (as experienced by ordinary users) is gradually becoming less web-shaped and more star-shaped, with your favourite search engine or social network at the centre. (Please note the word "gradually".)
Sources
David Auerbach, Deranked - Why has Google forsaken MetaFilter? (Slate May 2014)
Bill Slawski, The New PageRank, Same as the Old PageRank? (March 2012)
Daniel Sour, It Knows (LRB October 2011)
Related posts
What shape is the internet (September 2010)
What shape is your intranet (May 2014)
Updated 17 May 2014
Saturday, May 10, 2014
What shape is your Intranet?
In my post What shape is the internet (September 2010), I said that shifting the emphasis from URL hotlinks to search undermines the idea of the internet's being web-shaped. This point is also made in a post by @apike, referenced by David and also called Burying the URL (April 2014).
URLs are the essence. They make hypertext hyper. The term “web” is no accident – it refers to this explicitly.See also an excellent Twitter debate following @apike's tweet.
When David talks about bifurcation, he means that "enterprise IT is diverging enterprise technology from consumer technology in a way that’s creating two irreconcilable branches of technology". He observes that most company intranets have a pretty lousy search facility.
But most company intranets have pretty lousy cross-linking as well. They are mostly just pdf graveyards stuffed with documents of indeterminate pedigree, which people are often reluctant to waste time searching (even if the search facility were better) because they don't expect to find anything of value.
Actually, you can't always find what you are looking for on the Internet either, and that has a lot to do with the limitations of search, but there are enough amusing distractions to conceal this fact. Surely we don't want our company intranets to copy the internet too closely?
And remember that the data revealing Enron's problems were cheerfully displayed on the Enron website. But nobody important had bothered to look at these documents properly. (Actually, a bunch of students had analysed them years previously and concluded that Enron was bankrupt. They probably got low marks for that assignment!)
There is an increasingly common belief that the technology used inside companies should work the same way as outside, should provide the same "affordance". This is not Bring Your Own Device but Bring Your Own
Related posts
What shape is the internet (September 2010)
Bring Your Own Expectations (May 2014)
See also Steven Poole, The pdf graveyards can only expect an increase in their undead populations (Guardian 9 May 2014)
Thursday, September 16, 2010
What shape is the Internet?
The current discussion started for me when @ironick quoted @Microsoft "The Web is about sites, and your browser should be, too." and asked "What's the difference between a site and an app?" @Cybersal quoted @dtapscott's alternative "web not about sites but platforms 4 collaboration". So @ironick asked "What's the WWW really a web of? sites, apps, pages, data... "
My first observation is that if the Internet is merely a collection of sites, apps, or even platforms, then it's not exactly a web. The word "web" appears to focus our attention on the connections rather than sites themselves. There are of course two kinds of connection that exist in the Internet, which we could roughly categorize as syntactic and semantic. A syntactic link is a hotlink coded in HTML, while a semantic link involves some kind of content relationship. For example, you'll note that in this blogpost I've gone to the trouble to add hotlinks to the tweets by Nick and Sally: if you wanted, you could go directly to Twitter to check their exact words. (Go directly to Twitter, do not pass Google, do not collect 200 cookies).
But even if I hadn't added the hotlinks, you'd still be able to find Nick and Sally and their Tweets, by copying their names or their words into a search engine. So I'm creating a semantic link just by referencing something that exists somewhere on the Internet, even if I don't tell you its exact location.
The original hypermedia experience was largely dependent upon syntactic links. For @ironick "the web still feels like hypermedia 2me: clicking from context (page, song, video, snippet, site, app) 2 cntxt 2 cntxt 2..." I agree that it often still feels like that, but I find that a lot of my Internet browsing these days involves typing terms into search engines, and I don't find myself following long chains of hotlinks. In other words, I tend to regard the semantic links as more interesting and more useful than the syntactic links.
Here are some of the many problems I experience with syntactic links
- Sometimes the links aren't provided at all.
- Thanks to an aversion to deep linking, many websites only provide a link to the home page.
- Sometimes the links don't go direct but via some tedious aggregator or intermediator page. Spurious links whose sole purpose is to manipulate the search engines or generate advertisement traffic.
- Sometimes the links take you somewhere boring or irrelevant or obvious (like a Wikipedia page), not-safe-for-work, or someone's idea of a joke (not Rick Astley again).
- Often the links are compressed, so you can't see where you are being led. (No, I don't want to watch a YouTube video right now, thank you.)
- Often the links contain all sorts of other coded information, to pass contextual information to the receiving website.
- And then to cap it all, half the time the links don't work for you anyway, because they are out-of-date, or because the person providing the link has a subscription and you don't, or because there is some kind of context or syntax error.
Of course, there are problems with semantic links as well, above all the danger of over-reliance on the chosen search engine. But I still feel I'm more in control of the experience.
When we talk about the Internet as a world wide web (WWW), the word "web" seems to suggest a network stretching endlessly in all directions, allowing and encouraging the kind of browsing experience Nick mentions. But of course the fly's experience of the spider's web is quite different: being caught in one place, trapped for the benefit of the spider. For a long time, it has been the desire of major internet providers to trap users in one place: this desire is now apparently satisfied whenever users do not stray more than one or two clicks away from their favourite search engine or social networking site. Maybe that's what Microsoft is getting at.
Related Posts
What shape is the internet (continued)? (May 2014)
What shape is your intranet (May 2014)
Thursday, July 15, 2010
Social Networks and Received Opinion
The internet has not become the great leveller that it was once thought it could be. The web is now contrary to the original utopian vision and users focus on information from a handful of wealthy countries. It's making us 'imaginary cosmopolitans'.
Social networks make the problem worse with the majority of people sharing information with folk who share their world-view. Our world-view might actually be narrowing.
Tools like Twitter trap people in so-called "filter bubbles". The internet is too big to understand as a whole, so we get a picture of it that's similar to what our friends see. If you turn to your friends, eventually you get the wisdom of the flock.
The term "filter bubbles" is credited to political activist @elipariser. See Ethan's earlier post Eli Pariser on Filter Bubbles (March 2010).
This phenomenon is important from many perspectives. One question that particularly interests me is the way that these networks can create the illusion of improved intelligence, while actually doing no such thing.
wish, illusion | actual | |
information gathering | availability: fast, rich, high quality, unmediated, diverse | homogeneous, filtered |
sense-making & decision-making | open, creative | closed, doctrinaire |
knowledge | complete, consistent, strong, independent | partial, partisan, weak, received opinion |
learning | progressive | pseudo-learning |
communication | authentic | vapid |
Obviously it would be crazy to write off social networking and the internet as an inevitable producer of these effects - that would be the kind of crude technological determinism that gets the tabloid newspapers bewailing the Perils of Facebook.
Instead, the challenge is both to use the available human and technical networks more wisely, and to develop sociotechnical mechanisms that help to realise the original vision of these technologies and contribute to a greater and better distributed intelligence and understanding. Zuckerman talks ambitiously about mechanisms for amplifying underrepresented voices, and for discovering content through serendipity. He also talks about important new roles - for example curators to collect the content, xenophiles to bridge different cultures, working together to put content into context.
But even if we cannot transform the world overnight, we (ourselves and our organizations) can at least start to use these technologies in a more contingent manner, and with greater awareness of their strengths and weaknesses.
See also Polarizing Filters (March 2021)
Monday, September 15, 2008
Rumour on the Internet
As I pointed out in my post on Turbulent Markets, the regulation Mark proposes raises some interesting technical challenges, as well as calling into question the value of some of the rapid-response technology Mark himself is selling.
In any case, control of rumour sounds to me like one of the twelve labours of Hercules. Agatha Christie adopted exactly that interpretation of the Lernaean Hydra when she wrote twelve stories for Hercule Poirot based on the twelve labours).
So imagine my surprise when I heard on the radio news that Tim Berners-Lee was calling for broad controls of Internet rumour [Warning sounded on web's future, BBC News 15 September 2008]. How on earth is that going to work? Is he also planning to clean the Augean stables (internet porn)? What about capturing the Golden Stag of Artemis (Steve Jobs), the Erymanthian Boar (Steve Ballmer), the Cretan Bull (Larry Ellison), the Horse of Diomedes (Jonathan Schwartz) and Cerebus (Henning Kagermann). Have I forgotten anybody?
I put "Berners-Lee rumour" into a well-known search engine and found a page from March 2006 by Aleks Krotoski, called Tim Berners-Lee on the Semantic Web. According to Aleks, Tim graduated from Oxford University in 1989. (Actually it was 1976. As I pointed out in my earlier post Hasta La Vista, Tim is now old enough to read a book called The Internet for the Older Generation "especially written for the over-50s".)
Aren't you impressed that a search engine, tasked with "Berners-Lee rumour", finds a false rumour about Tim's age? Can the Semantic Web do the same?
(Okay, okay, I'm being sarcastic. Obviously the search didn't find the page with the false information on purpose. And it's not exactly difficult to find pages on the Internet with false information, is it?)
Tuesday, August 12, 2008
Listening Post
The work is a collaboration between an artist (Ben Rubin, who teaches at the Yale School of Art) and a statistician (Mark Hansen, formerly of Bell Labs/Lucent and now associate professor of statistics at UCLA).
The work has been evolving since 2001, and provides a real-time audio-visual summary from a vast number of Internet chatrooms, presented on hundreds of tiny screens, and with selected messages spoken using voice synthesis. When my eyes adjusted to the darkened room, I saw some benches in the middle and went to sit down. But much of the text was too small to see from that distance, and I found it more interesting to stand a little closer and try to read as many as possible of the hundreds of messages flashing across the array of screens.
The work is constructed in seven "movements". For example, one movement is entirely constructed from statements of presence ("I am hot", "I am waiting", "I am 32 years old"), while another movement is entirely constructed from random user names.
If the purpose of art is to provide an unfamiliar view on the familiar, then this worked elegantly and brilliantly. I was fascinated, and I shall certainly try to get back a few times to repeat the experience before it closes.
Notices and Reviews
Hannah Redler (Science Museum): "Monument to the present - the sound of 100,000 people chatting"Torin Douglas (BBC News, Feb 2008): "Listening to internet chatter"
Peter Eleey (Frieze Magazine, May 2003)
Interactive Architecture (August 2005)
Wednesday, July 23, 2003
Bleak Future of the Internet?
Innocent newsletters are getting caught in email filters, and newsletter senders are finding this increasingly frustrating. David Sprott of CBDI devoted his July 10th 2003 newsletter to this topic, and Bruce Schneier (CryptoGram) picked up the topic again in his newsletter of July 15th 2003.
Filters may be locally effective - and this encourages some complacency. But the filters are generally ineffective, and generate significant levels of false positives. Furthermore, the existence of filters simply encourages the producers of rubbish to increase their production volumes, at little cost to themselves, in order to maintain the desired levels of dissemination. They are therefore counterproductive for the Internet as a whole.
While many individuals and small businesses have become dependent on the internet, there are growing numbers of old-economy firms where the nuisance and risks of connection to the internet may be perceived to outweigh the advantages. It it may be hard to continue to justify open access, and many firms may be tempted to disconnect themselves from the internet altogether.
Even in the largest firms, there will always be individuals and groups who want to remain connected to the internet for various reasons - including marketing and R&D groups. But the corporate interest may prevail - and it may be a constant effort to keep the lines of communication open.
This scenario should be extremely worrying for decent small firms - as well as large media empires - whose business depends on proper use of the internet. We are currently talking to a number of media and technology firms, to prepare contingency plans against this scenario.