Search Engine Strategies: Looking at Links

A special report from the Search Engine Strategies 2001 Conference, August 12-14, San Jose CA.

Link analysis is one of the most important, and widely misunderstood techniques search engines use to determine relevance, and a key element of successful search engine optimization. Representatives from Google and Teoma explain how it’s done.

If you have spent any time over the past few years studying search engine marketing you are probably familiar with the linking craze going on in the industry. Everyone from experts to those new to the field toss about terms like “link popularity” and “page rank” and it seems that all related discussion forums and web sites have entire sections devoted to linking. As the foundation of the web, links have always been important, but links themselves haven’t changed much since the day they were created so why all the renewed interest?

The reason is that the major search engines are utilizing links more and more to improve the relevance of their search results. However, the world of links and their use by search engines can get confusing quickly. To help sort through the more important elements of linking the session “Looking at Links” was held at the Search Engine Strategies conference in San Jose, California. The search engines that utilize links the most, Google and Teoma, both sent representatives to explain why links are important to their engines, and how to best utilize them on a web site.

Daniel Dulitz, Director of Technology for Google, started things off by stating one of the more important points of the session — as search engine indexes grow larger it becomes almost impossible to determine a web page’s relevancy based solely upon on-the-page factors (page text, metas, titles, etc.). It’s this fact, combined with the reality that most on-the-page factors can’t be trusted due to abuse, that prompted Google to begin looking at the link structure of the web to help determine a page’s relevance to a query.

According to Dulitz, when determining the relevance of a web page to a search they use their PageRank system to attempt to “model the behavior of web surfers” by analyzing the manner in which pages are linked to one another. He explained that Google views the interlinking of web pages as a way of “leveraging the democratic structure of the web” with links equating to votes.

Google essentially treats each link from one site to another as a vote for the site receiving the link (link popularity), but each vote is not created equal. Dulitz used a simple diagram to show that each page of a site only has one vote to give, so the more links to different sites on the same page the less of a vote each one receives. He also stated that links from higher quality sites carry more weight than those of lesser quality sites (e.g. sites with hidden links, involved in link farms, no incoming links, etc.). In addition, Google not only analyzes who is linking to whom, but they also analyze the text in and around the links to help determine the relevance of the pages receiving the links.

Dulitz then went on to provide some helpful linking tips for webmasters. When building links to their sites they should:

* Get similar sites to link to their site
* Attract links naturally by creating high quality content
* Avoid link farms, and other deceptive tactics

When linking to other sites webmasters should:

* Link to sites similar or related to their own
* Link to the most relevant page for their visitors (e.g. don’t just link to the home page)
* Avoid linking to unrelated sites

He then provided some additional general linking tips:

* Make sure all links work across all major browsers (e.g. no JavaScript)
* Use descriptive text within links whenever possible (e.g. no “Click Here” links)
* Make sure all links are easy to find and help users easily navigate the site
* Avoid hidden links

Paul Gardi, Vice President of Search for Ask Jeeves/Teoma, began with similar comments to those made by Dulitz. Gardi stated that “due to statistical convergence” and the ease with which they can be abused, neither page text analysis nor standard link popularity can be relied upon when determining the relevancy of a web page. Specifically, he mentioned that standard link popularity is ineffective because it does not help determine the subject or the context of the site, and larger more popular sites tend to overwhelm smaller sites that may actually be more relevant to a search.

To combat these issues Teoma views the web as a global entity that contains many subject based web site communities. They study these subject communities and the manner in which they are interlinked within themselves and with each other to determine not only their link popularity, but also the subject and context of the involved sites. According to Gardi, Teoma is able to do this by using their unique method of ranking sites. He explained that rather than relying on general link popularity to determine results, their engine attempts to employ a “subject specific popularity” to locate the most popular sites within a specific subject community. This is done by first analyzing the web as a whole to identify subject communities.

Teoma then employs link popularity within those communities to determine which sites are the “authorities” on the subject of the query and it’s those sites that are returned as their results to a search. In addition, he mentioned that by analyzing the links of the authority sites their technology is also able to locate high-quality resource pages (links pages) that are related to the original query. Each of these components is then made available on their search results page as follows: “Results” are the authorities, “Refine” is the related subject communities, and “Resources” are the related links pages.

At the end of his presentation, Gardi offered some basic advice for improving a web page’s rank at Teoma. He said:

* Optimize your pages
* Obtain links from subject related sites
* Avoid cloaking
* Avoid automatic queries of Teoma (e.g. automated rank checking)
* Avoid linking to bad sites

Overall, the session was well received and very informative, especially for those new to the subject. Considering that most major search engines now utilize some method of link analysis, anyone that has a vested interest in being properly indexed by the search engines should consider attending in the future.

Craig Fifield is Product Designer for Microsoft bCentral’s Small Business Web site analysis and submission service, Submit It!

Related reading

Search engine results: The ten year evolution
Five ways PPC customer support can help SMBs
#GoogleDoBetter The latest on internal issues at Google and Alphabet
Google Sandbox Is it still affecting new sites in 2019