ss_blog_claim=b1f73fce8095b4ae385f9842451cdc1c
Showing posts with label Page Ranking. Show all posts
Showing posts with label Page Ranking. Show all posts

Sunday, December 14, 2008

Google SEO Page Rank rigged

Google admitted that their Page Ranking System is rigged. That tells me that SEO’s attempting to sell expensive services to dominate master Google are mostly full of baloney.

Here is the original article:

Original URL: http://www.theregister.co.uk/2008/12/12/googlewashing_revisited/

Google cranks up the Consensus Engine

Manufacturing isn't dead - it just went to Mountain View

By Andrew Orlowski (andrew.orlowski@theregister.co.uk)

Posted in Music and Media, 12th December 2008 19:38 GMT

Free Download - Comparing Data Center Batteries, Flywheels and Ultracapacitors

Google this week admitted that its staff will pick and choose what appears in its search results. It's a historic statement - and nobody has yet grasped its significance.

Not so very long ago, Google disclaimed responsibility for its search results by explaining that these were chosen by a computer algorithm. The disclaimer lives on at Google News, where we are assured that:

The selection and placement of stories on this page were determined automatically by a computer program.

A few years ago, Google's apparently unimpeachable objectivity got some people very excited (http://www.internetisshit.org/2.php), and technology utopians began to herald Google as the conduit for a new form of democracy. Google was only too pleased to encourage this view. It explained that its algorithm "relies on the uniquely democratic nature of the web by using its vast link structure as an indicator of an individual page's value. "

That Google was impartial was one of the articles of faith. For if Google was ever to be found to be applying subjective human judgment directly on the process, it would be akin to the voting machines being rigged.

For these soothsayers of the Hive Mind, the years ahead looked prosperous. As blog-aware marketing and media consultants, they saw a lucrative future in explaining the New Emergent World Order to the uninitiated. (That part has come true - Web 2.0 "gurus" now advise large media companies).

It wasn't surprising, then, that when five years ago I described how a small, self-selected number of people could rig Google's search results (http://www.theregister.co.uk/2003/04/03/antiwar_slogan_coined_repurposed/), the reaction from the people doing the rigging was violently antagonistic. Who lifted that rock? they cried.

But what was once Googlewashing by a select few now has Google's active participation.

This week Marissa Meyer explained that editorial judgments will play a key role in Google searches. It was reported by Tech Crunch proprietor Michael Arrington - who Nick Carr called the "Madam of the Web 2.0 Brothel" - but its significance wasn't noted. The irony flew safely over his head at 30,000 feet. Arrington observed:

Mayer also talked about Google’s use of user data created by actions on Wiki search to improve search results on Google in general. For now that data is not being used to change overall search results, she said. But in the future it’s likely Google will use the data to at least make obvious changes. An example is if “thousands of people” were to knock a search result off a search page, they’d be likely to make a change.

Now what, you may be thinking, is an "obvious change"? Is it one that is frivolous? (Thereby introducing a Google Frivolitimeter™ [Beta]). Or is it one that goes against the grain of the consensus? If so, then who decides what the consensus must be? Make no mistake, Google is moving into new territory: not only making arbitrary, editorial choices - really no different to Fox News, say, or any other media organization. It's now in the business of validating and manufacturing consent: not only reporting what people say, but how you should think.

Who's hand is upon the wheel, here?

None of this would matter, if it wasn't for one other trend: a paralysing loss of confidence in media companies.



Old media is hooked on the drug that kills it

Today, the media organisations look to Google to explain what is really happening in the world. Convinced that they can't lead, the only option left is to follow. So they reflect ourselves - or more accurately, they reflect the unstinting efforts of small self-selecting pockets of activists - back at us. In the absence of editorial confidence, Google - the Monster that threatens to Eat The Media - now defines the purpose of the media. All media companies need do is "tap into the zeitgeist" - Google Zeitgeist™!

Take this example from a quality British broadsheet.

One journalist on the paper lamented that:

...it's becoming all too clear at The Telegraph, whose online business plan seems to be centred on chasing hits through Google by rehashing and rewriting stories that people are already interested in.

The digital director of the Telegraph recently suggested the newspaper could work even closer with Google... by subsuming its identity into the Ad Giant. Why couldn't The Telegraph run off a telegraph.google.com domain and allow Google to take care of all the technology? he mused (http://www.guardian.co.uk/media/greenslade/2008/sep/18/uk).

Not all companies have the same suicidal lack of foresight as The Telegraph's resident guru - but many share the same apocalyptic conclusion.

Today, Google's cute little explanation of being "uniquely democratic" is no longer present on that page. A subtly different explanation has taken its place - one which acknowledges that in the new democracy of Web 2.0, some votes are more equal than others.

PageRank also considers the importance of each page that casts a vote, as votes from some pages are considered to have greater value, thus giving the linked page greater value. We have always taken a pragmatic approach to help improve search quality and create useful products, and our technology uses the collective intelligence of the web to determine a page's importance.

Google's New Age motto

Picture culled from Google's 2006 analyst presentation

So you see, it's not rigged! How could Google "rig" a system that only reflects our finest and most noble sentiments back at us - mediated by a technocratic priesthood of unquestionable moral authority?

Google has taken Googlewashing in house. ®

Tuesday, July 22, 2008

Lets talk SEO

Thousands of servers ...billions of web pages.... the possibility of individually sifting through the WWW is null. The search engine gods cull the information you need from the Internet...from tracking down an elusive expert for communication to presenting the most unconventional views on the planet. Name it and click it. Beyond all the hype created about the web heavens they rule, let's attempt to keep the argument balanced. From Google to Voice of the Shuttle (for humanities research) these ubiquitous gods that enrich the net, can be unfair ...and do wear pitfalls. And considering the rate at which the Internet continues to grow, the problems of these gods are only exacerbated further.

Primarily, what you need to digest is the fact that search engines fall short of Mandrake's magic mechanism! They simply don't create URLs out of thin air but instead send their spiders crawling across those sites that have rendered prayers (and expensive offerings!) to them for consideration. Even when sites like Google claim to have a massive 3 billion web pages in its database, a large portion of the web nation is invisible to these spiders. To think they are simply ignorant of the Invisible Web. This invisible web holds that content, normal search engines can't index because the information on many web sites is in databases that are only searchable within that site. Sites like www.imdb.com - The Internet Movie Database , www.incywincy.com - IncyWincy, the invisible web search engine and www.completeplanet.com - The Complete Planet that cover this area are perhaps the only way you can access content from that portion of the Internet, invisible to the search gods. Here, you don't perform a direct content search but search for the resources that may access the content. (Meaning - be sure to set aside considerable time for digging.)

None of the search engines indexes everything on the Web (I mean none). Tried research literature on popular search engines? AltaVista to Yahoo, will list thousands of sources on education, human resource development, etc. etc. but mostly from magazines, newspapers, and various organizations' own Web pages, rather than from research journals and dissertations- the main sources of research literature. That's because most of the journals and dissertations are not yet available publicly on the Web. Thought they'll get you all that's hosted on the web? Think again.

The Web is huge and growing exponentially. Simple searches, using a single word or phrase, will often yield thousands of "hits", most of which will be irrelevant. A layman going in for a piece of info to the internet has to deal with a more severe issue - too much information! And if you don't learn how to control the information overload from these websites, returned by a search result, roll out the red carpet for some frustration. A very common problem results from sites that have a lot of pages with similar content. For e.g., if a discussion thread (in a forum) goes on for a hundred posts there will be a hundred pages all with similar titles, each containing a wee bit of information. Now instead of just one link, all hundred of those darn pages will crop up your search result, crowding out other relevant site. Regardless of all the sophistication technology has brought in, many well thought-out search phrases produce list after list of irrelevant web pages. The typical search still requires sifting through dirt to find the gold. If you are not specific enough, you may get too many irrelevant hits.

As said, these search engines do not actually search the web directly but their centralized server instead. And unless this database is updated continually to index modified, moved, deleted or renamed documents, you will land yourself amidst broken links and stale copies of web pages. So if they inadequately handle dynamic web pages whose content changes frequently, chances are for the information they reference to quickly go out-of-date. After they wage their never ending war with over-zealous promoters (spamdexers rather), where do they have time to keep their databases current and their search algorithms tuned? No surprise if a perfectly worthwhile site may go unlisted!

Similarly, many of the Web search engines are undergoing rapid development and are not well documented. You will have only an approximate idea of how they are working, and unknown shortcomings may cause them to miss desired information. Not to mention, amongst the first class information, the web also houses false, misleading, deceptive and dressed up information actually produced by charlatans. The Web itself is unstable and tomorrow they may not find you the site they found you today. Well if you could predict them, they would not be god!...would they?! The syntax (word order and punctuation) for various types of complex searches varies some from search engine to search engine, and small errors in the syntax can seriously compromise the search. For instance, try the same phrase search on different search engines and you'll know what I mean. Novices... read this line - using search engines does involve a learning curve. Many beginning Internet users, because of these disadvantages, become discouraged and frustrated. Like a journalist put it, "Not showing favoritism to its business clients is certainly a rare virtue in these times." Search engines have increasingly turned to two significant revenue streams. Paid placement: In addition to the main editorial-driven search results, the search engines display a second - and sometimes third - listing that's usually commercial in nature. The more you pay, the higher you'll appear in the search results. Paid inclusion: An advertiser or content partner pays the search engine to crawl its site and include the results in the main editorial listing. So?...more likely to be in the hit list but then again - no guarantees. Of course those refusing to favor certain devotees are industry leaders like that publishes paid listings, but clearly marks them as 'Sponsored Links.'

The possibility of these 'for-profit' search gods (which haven't yet made much profit) for taking fees to skew their searches, can't be ruled out. But as a searcher, the hit list you are provided with by the engine should obviously rank in the order of relevancy and interest. Search command languages can often be complex and confusing and the ranking algorithm is unique to each god based on the number of occurrences of the search phrase in a page, if it appears in the page title, or in a heading, or the URL itself, or the meta tag etc. or on a weighted average of a number of these relevance scores. E.g. Google (www.google.com) uses its patented PageRank TM and ranks the importance of search results by examining the links that lead to a specific site. The more links that lead to a site, the higher the site is ranked. Pop on popularity!
Alta Vista, HotBot, Lycos, Infoseek and MSN Search use keyword indexes ? fast access to millions of documents. The lack of an index structure and poor accuracy of the size of the WWW, will not make searching any easier. Large number of sites indexed. Keyword searching can be difficult to get right.In reality, however, the prevalence of a certain keyword is not always in proportion to the relevance of a page. Take this example. A search on sari - the national costume of India ?in a popular search engine, returned among it's top sites, the following links:?www.scri.sari.ac.uk/- of the Scottish Crop research Institute ?www.ubudsari.com/ -a health resort in Indonesia?www.sari-energy.org/ - The South Asia Regional Initiative for Energy Cooperation and DevelopmentPretty useful sites for someone very much interested in knowing how to drape or the tradition of the sari?! (Well, no prayer goes unanswered...whether you like the answer or not!) By using keywords to determine how each page will be ranked in search results and not simply counting the number of instances of a word on a page, search engines are attempting to make the rankings better by assigning more weight to things like titles, subheadings, and so on.Now, unless you have a clear idea of what you're looking for, it may be difficult or impossible to use a keyword search, especially if the vocabulary of the subject is unfamiliar. Similarly, the concept based search of Excite (instead of individual words, the words that you enter into a search are grouped and attempted to determine the meaning) is a difficult task and yields inconsistent results.Besides who reviews or evaluates these sites for quality or authority? They are simply compiled by a computer program. These active search engines rely on computerized retrieval mechanisms called "spiders", "crawlers", or "robots", to visit Web sites, on a regular basis and retrieve relevant keywords to index and store in a searchable database. And from this huge database yields often unmanageable and comprehensive results....results whose relevance is determined by their computers. The irrelevant sites (high percentage of noise, as it's called), questionable ranking mechanisms and poor quality control may be the result of less human involvement to weed out junk. Thought human intervention would solve all probes....read on.

From the very first search engine ? Yahoo to about.com, Snap.com, Magellan, NetGuide, Go Network, LookSmart, NBCi and Starting Point, all subject directories index and review documents under categories ? making them more manageable. Unlike active search engines, these passive or human-selected search engines like don't roam the web directly and are human controlled, relying on individual submissions. Perhaps the easiest to use in town, but the indexing structure these search engines cover only a small portion of the actual number of WWW sites and thus is certainly not your bet if you intend specific, narrow or complex topics. Subject designations may be arbitrary, confusing or wrong. A search looks for matches only in the descriptions submitted. Never contains full text of the web they link to - you can only search what you see titles, descriptions, subject categories, etc. Human-labor intensive process limits database currency, size, rate of growth and timeliness. You may have to branch through the categories repeatedly before arriving at the right page. They may be several months behind the times because of the need for human organization. Try looking for some obscure topic....chances for the people that maintain the directory to have excluded those pages. Obviously, machines can blindly count keywords but they can't make common-sense judgement as humans can. But then why does human-edited directories respond with all this junk?!
And here's about those meta search engines. A comprehensive search on the entire WWW using The Big Hub, Dogpile, Highway61, Internet Sleuth or Savvysearch , covering as many documents as possible may sound as good an idea as a one stop shopping.Meta search engines do not create their own databases. They rely on existing active and passive search engine indexes to retrieve search results. And the very fact that they access multiple keyword indexes reduces their response time. It sure does save your time by searching several search engines at once but at the expense of redundant, unwanted and overwhelming results....much more ? important misses. The default search mode differs from search site to search site, so the same search is not always appropriate in different search engine software. The quality and size of the databases vary widely.

Weighted Search Engines like Ask Jeeves and RagingSearch allows the user to type queries in plain English without advanced searching knowledge, again at the expense of inaccurate and undetailed searching. Review or Ranking Sources like Argus Clearinghouse (www.clearinghouse.net),eBlast (eblast.com) and Librarian's Index to the Internet (lii.org). They evaluate website quality from sources they find or accept submissions from but cover a minimal number of sites.

As a webmaster, your site registration with the biggest billboards in Times Square can get you closer to bingo! for the searcher. Those who didn't even know you existed before are in your living room in New York time!Your URL registration is a no-brainer, considering the generation of flocking traffic to your site. Certainly a quick and inexpensive method, yet is only a component of the overall marketing strategy that in itself offers no guarantees, no instant results and demands continued effort for the webmaster. Commerce rules the web. Like how a notable Internet caveman put it, "Web publishers also find dealing with search engines to be a frustrating pursuit. Everybody wants their pages to be easy for the world to find, but getting your site listed can be tough. Search sites may take a long time to list your site, may never list it at all, and may drop it after a few months for no reason. If you resubmit often, as it is very tempting to do, you may even be branded a spamdexer and barred from a search site. And as for trying to get a good ranking, forget it! You have to keep up with all the arcane and ever-changing rules of a dozen different search engines, and adjust the keywords on your pages just so...all the while fighting against the very plausible theory that in fact none of this stuff matters, and the search sites assign rankings at random or by whim.

"To make the best use of Web search engines--to find what you need and avoid an avalanche of irrelevant hits-- pick search engines that are well suited to your needs. And lest you'd want to cry "Ye immortal gods! where in the world are we?", spend a few hours becoming moderately proficient with each. Each works somewhat differently, most importantly in respect to how you broaden or narrow a search.

Finding the appropriate search engine for your particular information need, can be frustrating. To effectively use these search engines, it is important to understand what they are, how they work, and how they differ. For e.g. while using a meta search engine, remember that each engine has its own methods of displaying and ranking results. Remember, search strategies affect the results. If the user is unaware of basic search strategies, results may be spotty.
Quoting Charlie Morris (the former editor of The Web developer's journal) - "Search engines and directories survive, and indeed flourish, because they're all we've got. If you want to use the wealth of information that is the Web, you've got to be able to find what you want, and search engines and directories are the only way to do that. Getting good search results is a matter of chance. Depending on what you're searching for, you may get a meaty list of good resources, or you may get page after page of irrelevant drivel. By laboriously refining your search, and using several different search engines and directories (and especially by using appropriate specialty directories), you can usually find what you need in the end."

Search engines are very useful, no doubt. Right from getting a quick view of a topic to finding expert contact info...verily certain issues lie in their lap. Now the very reason we bother about these search engines so much is because they're all we've got! Though there sure is a lot of room for improvement, the hour's need is to not get caught in the middle of the road. By simply understanding what, how and where to seek, you'd spare yourself the fate of chanting that old Jewish proverb "If God lived on earth, people would break his windows."

Happy searching!