- Web 2.0 is next week in the States, and I’m not going to be there, even though I’m completely thinking about this a lot at the moment… I really need to get the stuff I’m thinking about around this area out in the open as quickly as possible before my brain melts…
- McSweeney’s Internet Tendency: 7 Habits of Highly Successful People. Some good ones here that I really ought to try and incorporate into my daily work…
- Sub-$100 laptop design unveiled “Nicholas Negroponte, chairman and founder of the Massachusetts Institute of Technology Media Labs, has been outlining designs for a sub-$100 PC.”
- A timeline of what happens in the movie Primer Thoroughly enjoyed the movie although was a bit off my game and had absolutely no idea what the crap was going on. Need to get on DVD.
- Early Season Two **spoilers** for people fascinated by “Lost” Symbols on sharks! What the crap!?
- Salon gushes over Serenity, only pausing to lament that the only thing better than it would be to have Firefly back… “Joss Whedon’s unfairly canceled TV series “Firefly” comes to the big screen — and marvelously. But for all its wit, it’s just not the same.”
- There’s a weird article on Socialtext about how Wikipedia is eating Google and I’m pretty sure I think most of it’s rubbish. Apparently Wikipedia is linked to by lots of mediocre sites and has lots of internal links. This is why it has the world’s top pagerank. I checked. Wikipedia’s homepage has a pagerank of ‘8’. The BBC’s is ‘9’. I suspect the whole thing is rubbish.
- This is a bit random: An illustator vector mockup of the Safari toolbar for those webdesigners who build in Illustrator and may need to show off their designs in print or on a projector Basically it stops all the nasty pixellation things happening when you make things bigger. Except why do it with Safari? That’s got to be a bit random, surely?
- Mock-ups of the proposed MIT $100 laptop It’s a nice bit of design work, all things considered. I hope it’s true and achievable. At the moment it looks rather better than a lot of much more expensive machines…
Categories
4 replies on “Links for 2005-10-02”
Tom, the claim it’s #1 may or may not be perfectly true, but Wikipedia does benefit from the way PageRank worked–at least the original algorithm. Only Google knows how it works today, of course.
cf. http://www.webworkshop.net/pagerank.html
I curbed my enthusiasm though. I admit, I got a little carried away with myself.
Well if you mean by ‘not perfectly true’ that it is demonstrably false with little effort, then yeah. I guess my question with the ‘benefit’ stuff is whether or not it benefits unfairly from the algorithm in question. I can’t see that it does – in fact you could argue that the limited internal navigation and the large number of outgoing links (if you believe that page you linked to) might make it quite a poor candidate. Personally, it looks to me like a well-assembled, good informational website that people link to a lot and as such is likely to get more pagerank because it’s worth getting more pagerank.
Tom,
There is a lot more to the story behind PageRank than what the Google toolbar suggests. For instance, Google selects certain sites as authorities. It’s impossible that Wikipedia will be gain the highest PageRank, but it is increasingly gaining the highest relevance on the page results. You can try it for yourself:
http://ericgalloway.blogspot.com/2005/07/do-20-of-wikipedia-articles-have-1.html
As I have been told, I’m (unsurprisingly) not exactly the first person to make this obvious claim about the relationship between the PageRank algorithm and Wikipedia.
http://www.nivi.com/blog/article/wikipedia-defeats-google-sometimes/
http://ericgalloway.blogspot.com/2005/09/does-yahoo-search-rank-wikipedia.html
Eric Galloway in particular has done a lot of investigation on this topic.
Regarding fairness, there’s this argument
http://www.qwikly.com/2005/07/wikipedia-and-google-work-together.html
Google’s PageRank algorithm is objective in and of itself, and in that sense ‘fair’, but it is not ‘in and of itself’. Rather, it is in a feedback cycle with the entire World Wide Web, and as a result it is only as ‘fair’ as the world of human endeavours.
Now, I first noticed the correlation between wiki size and PageRank in the early days of wikidom, and later discovered the underlying mathematical reason why. These days I think about how this relates to my more pressing problem of WikiSpam, and why wikis are hugely disruptive to Google’s quality.
If you recall, not so long ago, Google was increasingly returning spam links as relevant links. The strategy was to build huge link farms. Google altered their search engine to counter. One technique they use is to identify ‘cookie cutter’ websites in the hopes this will encourage high quality content to be rated highly.
It’s a natural economic relationship, really, and one that will continue to be fought. As long as Google uses a statistical model to deduce quality rather than human intellect, people will look for ways to bias those statistics in their own favour. Google will look for a way to punish those people, and other people will look for a way to punish and extort Google.
The search results are never going to be fair. What is more important to me is “are they useful?”
Google is not in the business of fair. They are in the business of selling search services and advertising, and they succed by answering questions usefully enough that people will go to them first for quick answers. If you want fair, you should use a multiple search engine approach, plus manual classification systems, plus personal research to find a more balanced picture–which is a process that is implicitly incorporated in the way Wikipedia works, mind you. (Wikipedians frequently do actual non-Internet library research. Shocking!)
However, I think fair is overrated when it comes to search engines because it’s unsettleable. Because the scope of the Web is too large to understand, you’ll never be able to predict, know, or decide whether a given ordered result set is ‘fair’. In particular, there are no objective criteria to measure. Sure, searching for ‘IBM’ should return http://www.ibm.com, but what about ‘Tom Coates’? Why are you first and not http://www.math.harvard.edu/~tomc/? You benefit because the structure of the PageRank algorithm likes your big content-full website with lots of blog hits, but that doesn’t mean that you are more relevant to my search query if I was looking for more of Dr. Coates’ papers.
Similarly, why am I a more relevant “Sunir” than Sunir Kapoor? I’m not as powerful as he is, but I do write a lot on wikis.
I’m sure the PR algorithm is constantly being tweaked. It has nothing to do with being a public company. They owe it to their shareholders to continue providing the best search engine to deliver profit. PageRank algo’s hardly affect profit until they start skewing SERPs to the point that customers start searching elsewhere. Furthermore, PR does not equal SERP, but is just one of hundreds of factors in the SERP algo.
I agree – furthermore I see no reason to believe google use only one algorythm, IMO they have numerous algos which they shuffle in and out all of the time & tweak in between times… what better way to keep ahead of the seo community for starters…