Total Pageviews

Search: This Blog, Linked From Here, The Web, My fav sites, My Blogroll

11 February 2011

Visible's Web search history





To understand the Web in the broadest and deepest sense, to
fully partake of the vision that I and my colleagues share, one
must understand how the Web came to be.
—Tim Berners-Lee, Weaving the Web


Internet Vs Web
“Internet” and “Web”are  not synonyms. The Internet is a networking protocol (set of rules) that allows computers of all types to connect to and communicate with other computers on the Internet.  Is the world’s largest computer network, made up of hundreds millions of computers. It’s really nothing more than the “plumbing” that allows information of various kinds to flow from computer to computer around the world.

The World Wide Web (Web), on the other hand, is a software protocol that runs on top of the Internet, allowing users to easily access files stored on Internet computers. Is one of many interfaces to the Internet, making it easy to retrieve text, pictures, and multimedia files from computers without having to know complicated commands. You just click a link and voila: You miraculously see a page displayed on your browser screen.

The Web is only about 20 years old, whereas the Internet is 50-something. Prior to the Web, omputers could communicate with one another, but the interfaces weren’t as slick or easy to use as the Web. Many of these older interfaces, or “protocols,” are still around and offer many different, unique ways of communicating with other computers (and other people).
Other Internet protocols and interfaces include:
As you see, the Internet is much more than the Web. In fact, databases accessed via Web interfaces, make up a significant portion of the Invisible Web.

The Internet’s origins trace back to a project sponsored by the U.S. Defense Advanced Research Agency (DARPA) in 1969 as a means for researchers and defense contractors to share information (Kahn, 2000).
The Web was created in 1990 by Tim Berners-Lee, a computer programmer working for the European Organization for Nuclear Research (CERN)

Prior to the Web, accessing files on the Internet was a challenging task, requiring specialized knowledge and skills. The Web made it easy to retrieve a wide variety of files, including text, images, audio, and video by the simple mechanism of clicking a hypertext link.
Hypertext
A system that allows computerized objects (text, images, sounds, etc.) to be linked together. A hypertext link points to a specific object, or a specific place with a text; clicking the link opens the file associated with the object.

How the Internet Came to Be
  1. Up until the mid-1960s, most computers were stand-alone machines that did not connect to or communicate with other computers.
  2.  In 1962 J.C.R. Licklider, a professor at MIT, wrote a paper envisioning a globally connected “Galactic Network” of computers (Leiner, 2000). The idea was far-out at the time, but it caught the attention of Larry Roberts, a project manager at the U.S. Defense Department’s Advanced Research Projects Agency (ARPA). 
  3. In 1966 Roberts submitted a proposal to ARPA that would allow the agency’s numerous and disparate computers to be connected in a network similar to Licklider’s Galactic Network. Roberts’ proposal was accepted, and work began on the “ARPANET,” which would in time become what we know as today’s Internet
  4. The first “node” on the ARPANET was installed at UCLA in 1969 and 
  5. gradually, throughout the 1970s, universities and defense contractors working on ARPA projects began to connect to the ARPANET.
  6. In 1973 the U.S. Defense Advanced Research Projects Agency (DARPA) initiated another research program to allow networked computers to communicate transparently across multiple linked networks. Whereas the ARPANET was just one network, the new project was designed to be a “network of networks.” According to Vint Cerf, widely regarded as one of the “fathers” of the Internet, “ This was called the Internetting project and the system of networks which emerged from the research was known as the ‘Internet’ ” (Cerf, 2000).
  7. It wasn’t until the mid 1980s, with the simultaneous explosion in use of personal computers, and the widespread adoption of a universal standard of Internet communication called Transmission Control Protocol/Internet Protocol (TCP/IP), that the Internet became widely available to anyone desiring to connect to it. Other government agencies fostered the growth of the Internet by contributing communications “backbones” that were specifically designed to carry Internet traffic. 
  8. By the late 1980s, the Internet had grown from its initial network of a few computers to a robust communications network supported by governments and commercial enterprises around the world.
  9. Despite this increased accessibility, the Internet was still primarily a tool for academics and government contractors well into the early 1990s
  10. As more and more computers connected to the Internet, users began to demand tools that would allow them to search for and locate text and other files on computers anywhere on the Net.

Early Net Search Tools
Although sophisticated search and information retrieval techniques date back to the late 1950s and early ‘60s, these techniques were used primarily in closed or proprietary systems.

Early Internet search and retrieval tools lacked even the most basic capabilities, primarily because it was thought that traditional information retrieval techniques would not work well on an open, unstructured information universe like the Internet. Accessing a file on the Internet was a two-part process. 
  1. First, you needed to establish direct connection to the remote computer where the file was located using a terminal emulation program called Telnet
  2. Then you needed to use another program, called a File Transfer Protocol (FTP) client, to fetch the file itself. 
For many years, to access a file it was necessary to know both
  1. the address of the computer and 
  2. the exact location and name of the file you were looking for
there were no search engines or other file-finding tools like the ones we’re familiar with today.
Telnet
A terminal emulation program that runs on your computer, allowing you to access a remote computer via a TCP/IP network and execute commands on that computer as if you were directly connected to it. Many libraries offered telnet access to their catalogs.
File Transfer Protocol (FTP)
A set of rules for sending and receiving files of all types between computers connected to the Internet.
Thus, “search” often meant:
  • sending a request for help to an e-mail message list or discussion forum and hoping some kind soul would respond with the details you needed to fetch the file you were looking for. 
  • The situation improved somewhat with the introduction of “anonymous” FTP servers, which were centralized file-servers specifically intended for enabling the easy sharing of files. The servers were anonymous because they were not password protected—anyone could simply log on and request any file on the system.
Files on FTP servers were organized in hierarchical directories, much like files are organized in hierarchical folders on personal computer systems today.
The hierarchical structure made it easy for the FTP server to display a directory listing of all the files stored on the server, but you still needed good knowledge of the contents of the FTP server. 
If the file you were looking for didn’t exist on the FTP server you were logged into, you were out of luck.
The first true search tool for files stored on FTP servers was called Archie, created in 1990 by a small team of systems administrators and graduate students at McGill University in Montreal
Archie was the prototype of today’s search engines, but it was primitive and extremely limited compared to what we have today. Archie roamed the Internet searching for files available on anonymous FTP servers, downloading directory listings of every anonymous FTP server it could find. These listings were stored in a central, searchable database called the Internet Archives Database at McGill University, and were updated monthly. Although it represented a major step forward, the Archie database was still extremely primitive, limiting searches to a specific file name, or for computer programs that performed specific functions. Nonetheless, it proved extremely popular—nearly 50 percent of Internet traffic to Montreal in the early ’90s was Archie related, according to Peter Deutsch, who headed up the McGill University Archie team. The team licensed Archie to
others, with the first shadow sites launched in Australia and Finland in 1992. The Archie network reached a peak of 63 installations around the world by 1995. A legacy Archie server is still maintained active for historic purposes at University of Warsaw.

“In the brief period following the release of Archie, there was an explosion of Internet-based research projects, including WWW, Gopher, WAIS, and others” (Deutsch, 2000). “Each explored a different area of the Internet information problem space, and each offered its own insights into how to build and deploy Internet-based services,” wrote Deutsch. 

Gopher, an alternative to Archie, was created by Mark McCahill and his team at the University of Minnesota in 1991 and was named for the university’s mascot, the Golden Gopher. 
Gopher essentially combined the Telnet and FTP protocols, allowing users to click hyperlinked menus to access information on demand without resorting to additional commands. 
Using a series of menus that allowed the user to drill down through successively more specific categories, users could ultimately access the full text of documents, graphics, and even music files, though not integrated in a single format. Gopher made it easy to browse for information on the Internet.
According to Gopher creator McCahill, “Before Gopher there wasn’t an easy way of having the sort of big distributed system where there were seamless pointers between stuff on one machine and another machine. You had to know the name of this machine and if you wanted to go over here you had to know its name. “Gopher takes care of all that stuff for you. So navigating around Gopher is easy. It’s point and click typically. So it’s something that anybody could use to find things. It’s also very easy to put information up so a lot of people started running servers themselves and it was the first of the easy-to-use, no muss, no fuss, you can just crawl around and look for information tools. It was the one that wasn’t written for techies.” Gopher’s “no muss, no fuss” interface was an early precursor of what later evolved into popular Web directories like Yahoo!. “Typically you set this up so that you can start out with a sort of overview or general structure of a bunch of information, choose the items that you’re interested in to move into a more specialized area and then either look at items by browsing around and finding some documents or submitting searches,” said McCahill.

A problem with Gopher was that it was designed to provide a listing of files available on computers in a specific location—the University of Minnesota, for example. While Gopher servers were searchable, there was no centralized directory for searching all other computers that were both using Gopher and connected to the Internet, or “Gopherspace” as it was called.

In November 1992, Fred Barrie and Steven Foster of the University of Nevada System Computing Services group solved this problem, creating a program called Veronica, a centralized Archie-like search tool for Gopher files.
In 1993 another program called Jughead added keyword search and Boolean operator capabilities to Gopher search.
Keyword
A word or phrase entered in a query form that a search system attempts to match in text documents in its database.
Boolean
A system of logical operators (AND, OR, NOT) that allows true-false operations to be performed on search queries, potentially narrowing or expanding results when used with keywords
Popular legend has it that Archie, Veronica and Jughead were named after cartoon characters. Archie in fact is shorthand for “Archives.” Veronica was likely named after the cartoon character (she was Archie’s girlfriend), though it’s officially an acronym for “Very Easy Rodent-Oriented Net-Wide Index to Computerized Archives.” And Jughead (Archie and Veronica’s cartoon pal) is an acronym for “Jonzy’s Universal Gopher Hierarchy Excavation and Display,” after its creator, Rhett “Jonzy” Jones, who developed the program while at the University of Utah Computer Center.

A third major search protocol developed around this time was Wide Area Information Servers (WAIS). Developed by Brewster Kahle and his colleagues at Thinking Machines, WAIS worked much like today’s metasearch engines. The WAIS client resided on your local machine, and allowed you to search for information on other Internet servers using natural language, rather than using computer commands. The servers themselves were responsible for interpreting the query and returning appropriate results, freeing the user from the necessity of learning the specific query language of each server. WAIS used an extension to a standard protocol called Z39.50 that was in wide use at the time. In essence, WAIS provided a single computer-to-computer protocol for searching for  information. This information could be text, pictures, voice, or formatted documents. The quality of the search results was a direct result of how effectively each server interpreted the WAIS query.
All of the early Internet search protocols represented a giant leap over the awkward access tools rovided by Telnet and FTP. Nonetheless, they still dealt with information as discrete data objects. And these protocols lacked the ability to make connections between disparate types of information—text, sounds, images, and so on—to form the conceptual links that transformed raw data into useful information. 

Although search was becoming more sophisticated, information on the Internet lacked popular appeal. In the late 1980s, the Internet was still primarily a playground for scientists, academics, government agencies, and their contractors.

Fortunately, at about the same time, a software engineer in Switzerland was tinkering with a program that eventually gave rise to the World Wide Web. He called his program Enquire Within Upon Everything, borrowing the title from a book of Victorian advice that provided helpful information on everything from removing stains to investing money.


Enquire Within Upon Everything

“Suppose all the information stored on computers everywhere were linked, I thought. Suppose I could program my computer to create a space in which anything could be linked to anything. All the bits of information in every computer at CERN, and on the planet, would be available to me and to anyone else. There would be a single, global information space.

“Once a bit of information in that space was labeled with an address, I could tell my computer to get it. By being able to reference anything with equal ease, a computer could represent associations between things that might seem
unrelated but somehow did, in fact, share a relationship. A Web of information would form.”
— Tim Berners-Lee, Weaving the Web


The Web was created in 1990 by Tim Berners-Lee, who at the time was a contract programmer at the Organization for Nuclear Research (CERN) high-energy physics laboratory in Geneva, Switzerland. The Web was a side project Berners-Lee took on to help him keep track of the mind-boggling diversity of people, computers, research equipment, and other resources that are de rigueur at a massive research institution like CERN. One of the primary challenges faced by CERN scientists was the very diversity that gave it strength. The lab hosted thousands of researchers every year, arriving from countries all over the world, each speaking different languages and working with unique computing systems. And since high-energy physics research projects tend to spawn huge amounts of experimental data, a program that could simplify access to information and foster collaboration was something of a Holy Grail.

Berners-Lee had been tinkering with programs that allowed relatively easy, decentralized linking capabilities for nearly a decade before he created the Web.

  1. He had been influenced by the work of Vannevar Bush, who served as Director of the Office of Scientific Research and Development during World War II. In a landmark paper called “As We May Think,” Bush proposed a system he called MEMEX, “a device in which an individual stores all his books, records, and communications, and which is mechanized so that it may be consulted with exceeding speed and flexibility” (Bush, 1945). The materials stored in the MEMEX would be indexed, of course, but Bush aspired to go beyond simple search and retrieval. The MEMEX would allow the user to build conceptual “trails” as he moved from document to document, creating lasting associations between different components of the MEMEX that could be recalled at a later time. Bush called this “associative indexing ... the basic idea of which is a provision whereby any item may be caused at will to select immediately and automatically another. This is the essential feature of the MEMEX. The process of tying two items together is the important thing.” 
  2. In Bush’s visionary writings, it’s easy for us to see the seeds of what we now call hypertext. But it wasn’t until 1965 that Ted Nelson actually described a computerized system that would operate in a manner similar to what Bush envisioned. Nelson called his system “hypertext” and described the next-generation MEMEX in a system he called Xanadu. Nelson’s project never achieved enough momentum to have a significant impact on the world. 
  3. Another twenty years would pass before Xerox implemented the first mainstream hypertext program, called NoteCards, in 1985
  4. A year later, Owl Ltd. created a program called Guide, which functioned in many respects like a contemporary Web browser, but lacked Internet connectivity. Bill Atkinson, an Apple Computer programmer best known for creating MacPaint, the first bitmap painting program, created the first truly popular hypertext program in 1987. His HyperCard program was specifically for the Macintosh, and it also lacked Net connectivity. Nonetheless, the program proved popular, and the basic functionality and concepts of hypertext were assimilated by Microsoft, appearing first in standard help systems for Windows software.

Weaving the Web
The foundations and pieces necessary to build a system like the World Wide Web were in place well before Tim Berners-Lee began his tinkering. 
But unlike others before him, Berners-Lee’s brilliant insight was that a simple form of hypertext, integrated with the universal communication protocols offered by the Internet, would create a platform-independent system with a uniform interface for any computer connected to the Internet. 
He tried to persuade many of the key players in the hypertext industry to adopt his ideas for connecting to the Net, but none were able to grasp his vision of simple, universal connectivity.
So Berners-Lee set out to do the job himself, creating a set of tools that collectively became the prototype for the World Wide Web (Connolly, 2000). In a remarkable burst of energy, Berners-Lee began work in October 1990 on the first Web client—the program that allowed the creation, editing, and browsing of hypertext pages. He called the client WorldWideWeb, after the mathematical term used to describe a collection of nodes and links in which any node can be linked to any other. “Friends at CERN gave me a hard time, saying it would never take off—especially since it yielded an acronym that was nine syllables long when spoken,” he wrote in Weaving the Web.
  1. To make the client simple and platform independent, Berners-Lee created HTML, or HyperText Markup Language, which was a dramatically simplified version of a text formatting language called SGML(Standard Generalized Markup Language). All Web documents formatted with HTML tags would display identically on any computer in the world.
  2. Next, he created the HyperText Transfer Protocol (HTTP), the set of rules that computers would use to communicate over the Internet and allow hypertext links to automatically retrieve documents regardless of their location. He also devised the Universal Resource Identifier, a standard way of giving documents on the Internet a unique address (what we call URLs today). 
  3. Finally, he brought all of the pieces together in the form of a Web server, which stored HTML documents and served them to other computers making HTTP requests for documents with URLs. 
  4. Berners-Lee completed his work on the initial Web tools by Christmas 1990. 
In little more than two months he had created a system that had been envisioned for decades. Building the tools was the easy part. The hard part was to get the world to embrace the Web. Berners-Lee wrote the original programs for the Web on the NeXT system, but thanks to his tireless efforts to persuade others to use the system, there were soon Web clients and servers available in a variety of different operating systems. As more and more people within CERN began to use the Web, the initial skepticism began to wear away.

Berners-Lee began actively to promote the Web outside of the lab, attending conferences and articipating in Internet mailing and discussion lists. Slowly, the Web began to grow as more and more people implemented clients and servers around the world. There were really two seminal events that sparked the explosion in popular use of the Web. 
  1. The first was the development of graphical Web browsers, including ViolaWWWMosaic, and others that integrated text and images into a single browser window. For the first time, Internet information could be displayed in a visually appealing format previously limited to CD-ROM-based multimedia systems. This set off a wave of creativity among Web users, establishing a new publishing medium that was freely available to anyone with Internet access and the basic skills required to design a Web page.
  2. Then in 1995, the U.S. National Science Foundation ceased being the central manager of the core Internet communications backbone, and transferred both funds and control to the private sector. Companies were free to register “dot-com” domain names and establish an online presence. It didn’t take long for business and industry to realize that the Web was a powerful new avenue for online commerce, triggering the dot-com gold rush of the late 1990s.

Early Web Navigation
The Web succeeded where other early systems failed to catch on largely because of its decentralized nature.

Despite the fact that the first servers were at CERN, neither Berners-Lee nor the lab exercised control over who put up a new server anywhere on the Internet.
Anyone could establish his or her own Web server. The only requirement was to link to other servers, and inform other Web users about the new server so they could in turn create links back to it.
But this decentralized nature also created a problem. Despite the ease with which users could navigate from server to server on the Web simply by clicking links, navigation was ironically becoming more difficult as the Web grew. No one was “in charge” of the Web; there was no central authority to create and maintain an index of the growing number of available documents.

To facilitate communication and cross-linkage between early adopters of the Web, Berners-Lee established a list Web of servers that could be accessed via hyperlinks. This was the first Web directory. This early Web guide is still online
Beyond the list of servers at CERN, there were few centralized directories, and no global Web search services. People notified the world about new Web pages in much the same way they had previously
announced new Net resources, via e-mail lists or online discussions. 
Eventually, some enterprising observers of the Web began creating lists of links to their favorite sites. John Makulowich, Joel Jones, Justin Hall, and the people at O’Reilly & Associates publishing company were among the most noted authors maintaining popular link lists.
Eventually, many of these link lists started “What’s New” or “What’s Cool” pages, serving as de facto announcement services for new Web pages. But they relied on Web page authors to submit information, and the Web’s relentless growth rate ultimately made it impossible to keep the lists either current or comprehensive.

What was needed was an automated approach to Web page discovery and indexing. The Web had now grown large enough that information scientists became interested in creating search services specifically for the Web. Sophisticated information retrieval techniques had been available since the early 1960s, but they were only effective when searching closed, relatively structured databases. The open, laissez-faire nature of the Web made it too messy to easily adapt traditional information retrieval techniques. New, Web-centric approaches were needed. But how best to approach the problem? Web search would clearly have to be more sophisticated than a simple Archie-type service. But should these new “search engines” attempt to index the full text of Web documents, much as earlier Gopher tools had done, or simply broker requests to local Web search services on individual computers, following the WAIS model?


The First Search Engines
Tim Berners-Lee’s vision of the Web was of an information space where data of all types could be freely accessed. But in the early days of the Web, the reality was that most of the Web consisted of simple HTML text documents. Since few servers offered local site search services, developers of the first Web search engines opted for the model of indexing the full text of pages stored on Web servers. To adapt traditional information retrieval techniques to Web search, they built huge databases that attempted to replicate the Web, searching over these relatively controlled, closed archives of pages rather than trying to search the Web itself in real time.
With this fateful architectural decision, limiting search engines to HTML text documents and essentially ignoring all other types of data available via the Web, the Invisible Web was born.
The biggest challenge search engines faced was simply locating all of the pages on the Web. Since the Web lacked a centralized structure, the only way for a search engine to find Web pages to index was by following links to pages and gathering new links from those pages to add to the queue to visit for indexing. This was a task that required computer assistance, simply to keep up with all of the new pages being added to the Web each day.

But there was a subtler problem that needed solving. Search engines wanted to fetch and index all pages on the Web, but the search engines frequently revisited popular pages at the expense of new or obscure pages, because popular pages had the most links pointing to them—which the crawlers naturally followed. What was needed was an automated program that had a certain amount of intelligence, able to recognize when a link pointed to a previously indexed page and ignoring it in favor of finding new pages.These programs became known as Web robots—“autonomous agents” that could find their way around the Web discovering new Web pages. Autonomous is simply a fancy way of saying that the agent programs can do things on their own without a person directly controlling them, and that they have some degree of intelligence, meaning they can make decisions and take action based on these decisions.
In June 1993 Mathew Gray, a physics student at MIT, created the first widely recognized Web robot, dubbed the “World Wide Web Wanderer.”
Gray’s interest was limited to determining the size of the Web and tracking its continuing growth. The Wanderer simply visited Web pages and reported on their existence, but didn’t actually fetch or store pages in a database. Nonetheless, Gray’s robot led the way for more sophisticated programs that would both visit and fetch Web pages for storage and indexing in search engine databases.
The year 1994 was a watershed one for Web search engines. Brian Pinkerton, a graduate student in Computer Sciences at the University of Washington, created a robot called WebCrawler in January 1994.
Pinkerton created his robot because his school friends were always sending him e-mails about the cool sites they had found on the Web, and Pinkerton didn’t have time to surf to find sites on his own—he wanted to “cut to the chase” by searching for them directly. WebCrawler went beyond Gray’s Wanderer by actually retrieving the full text of Web documents and storing them in a keyword-searchable database. Pinkerton made WebCrawler public in April 1994 via a Web interface. The database contained entries from about 6,000 different servers, and after a week was handling 100+ queries per day. The first Web search engine was born. The image evoked by Pinkerton’s robot “crawling” the Web caught the imagination of programmers working on automatic indexing of the Web. Specialized search engine robots soon became known generically as “crawlers” or “spiders,” and their page-gathering activity was called “crawling” or “spidering” the Web.

Crawler-based search engines proliferated in 1994. Many of the early search engines were the result of academic or corporate research projects. Two popular engines were the World Wide Web Worm, created by Oliver McBryan at the University of Colorado, and WWW JumpStation, by Jonathon Fletcher at the University of Stirling in the U.K.

Neither lasted long: Idealab purchased WWWWorm and transformed it into the first version of the GoTo search engine. JumpStation simply faded out of favor as two other search services launched in 1994 gained popularity: Lycos and Yahoo!.

Michael Mauldin and his team at the Center for Machine Translation at Carnegie Mellon University created Lycos (named for the wolf spider, Lycosidae lycosa, which catches its prey by pursuit, rather than in a web). Lycos quickly gained acclaim and prominence in the Web community, for the sheer number of pages it included in its index (1.5 million documents by January 1995) and the quality of its search results. Lycos also pioneered the use of automated abstracts of documents in search results, something not offered by WWW Worm or JumpStation.

Also in 1994, two graduate students at Stanford University created “Jerry’s Guide to the Internet,” built with the help of search spiders, but consisting of editorially selected links compiled by hand into a hierarchically organized directory. In a whimsical acknowledgment of this structure, Jerry Wang and David Filo renamed their service “Yet Another Hierarchical Officious Oracle,” commonly known today as Yahoo!.

In 1995 Infoseek, AltaVista, and Excite made their debuts, each offering different capabilities for the searcher.
Metasearch engines—programs that searched several search engines simultaneously—also made an appearance this year. 
SavvySearch, created by Daniel Dreilinger at Colorado State University, was the first metasearch engine, and MetaCrawler, from the University of Washington, soon followed. From this point on, search engines began appearing almost every day.
As useful and necessary as they were for finding documents, Web search engines all shared a common weakness: They were designed for one specific task—to find and index Web documents, and to point users to the most relevant documents in response to keyword queries.
During the Web’s early years, when most of its content consisted of simple HTML pages, search engines performed their tasks admirably. But the Internet continued to evolve, with information being made available in many formats other than simple text documents. For a wide variety of reasons, Web search services began to fall behind in keeping up with both the growth of the Web and in their ability to recognize and index non-text information—what we refer to as the Invisible Web.
To become an expert searcher, you need to have a thorough understanding of the tools at your disposal and, even more importantly, when to use them. 
Now that you have a sense of the history of the Web and the design philosophy that led to its universal adoption, let’s take a closer look at contemporary search services, focusing on their strengths but also illuminating their weaknesses.


Information Seeking on the Visible Web

The creators of the Internet were fundamentally interested in solving a single problem: how to connect isolated computers in a universal network, allowing any machine to communicate with any other regardless of type or location. 

The network protocol they developed proved to be an elegant and robust solution to this problem, so much so that a myth emerged that the Internet was designed to survive a nuclear attack
(Hafner and Lyon, 1998). In solving the connectivity problem, however, the Net’s pioneers largely ignored three other major problems—problems that made using the Internet significantly challenging to all but the most skilled computer users. 
  1. The first problem was one of incompatible hardware. Although the TCP/IP network protocol allows hardware of virtually any type to establish basic communications, once a system is connected to the network it may not be able to meaningfully interact with other systems. Programs called emulators” that mimic other types of hardware are often required for successful communication.
  2. The second problem was one of incompatible software. A computer running the UNIX operating system is completely different from one running a Windows or Macintosh operating system. Again, translation programs were often necessary to establish communication. 
  3. Finally, even if computers with compatible hardware and software connected over the Internet, they still often encountered the third major problem: incompatible data structures. Information can be stored in a wide variety of data structures on a computer, ranging from a simple text file to a complex “relational” database consisting of a wide range of data types.
In creating the Web, Tim Berners-Lee sought to solve all three of these problems. To a large extent, he succeeded in solving the problems of hardware and software incompatibilities. Like the TCP/IP network protocol, the Web’s HTML language was designed to function almost identically on computers using any type of hardware or software. The goal was to allow users to access information with a simple point and click interface that required no other knowledge of how a system functioned. 
HTML would also display documents in virtually identical format regardless of the type of hardware or software running on either the computer serving the page or the client computer viewing the document. To meet this goal, HTML was engineered as a very simple, barebones language. Although Berners-Lee foresaw Web documents linking to a wide range of disparate data types, the primary focus was on text documents. Thus the third problem—the ability to access a wide range of data types—was only partially solved.
The simplicity of the point and click interface also has an Achilles’ heel. It’s an excellent method for browsing through a collection of related documents—simply click, click, click, and documents are displayed with little other effort. Browsing is completely unsuitable and inefficient, however, for searching through a large information space.

To understand the essential differences between browsing and searching, think of how you use a library.
If you’re familiar with a subject, it’s often most useful to browse in the section where books about the subject are shelved. Because of the way he library is organized, often using either the Dewey Decimal or Library of Congress Classification system, you know that all of the titles in the section are related, and often leads to unexpected discoveries that prove quite valuable.
If you’re unfamiliar with a subject, however, browsing is both inefficient and potentially futile if you fail to locate the section of the library where the material you’re interested in is shelved. Searching, using the specialized tools offered by a library’s catalog, is far more likely to provide satisfactory results.
Using the Web to find information has much in common with using the library. 
Sometimes browsing provides the best results, while other information needs require nothing less than sophisticated, powerful searching to achieve the best results.We look at browsing and searching—the two predominant methods for finding information on the Web. In examining the strengths and weaknesses of each approach, you’ll understand how general-purpose information-seeking tools work—an essential foundation for later understanding why they cannot fully access the riches of the Invisible Web.


Browsing vs. Searching
There are two fundamental methods for finding information on the Web: browsing and searching. 
  1. Browsing is the process of following a hypertext trail of links created by other Web users. A hypertext link is a pointer to another document, image, or other object on the Web. The words making up the link are the title or description of the document that you will retrieve by clicking the link. By its very nature, browsing the Web is both easy and intuitive.
  2. Searching, on the other hand, relies on powerful software that seeks to match the keywords you specify with the most relevant documents on the Web. Effective searching, unlike browsing, requires learning how to use the search software as well as lots of practice to develop skills to achieve satisfactory results.
When the Web was still relatively new and small, browsing was an adequate method for locating relevant information. Just as books in a particular section of a library are related, links from one document tend to point to other documents that are related in some way. However, as the Web grew in size and diversity, the manual, time-intensive nature of browsing from page to page made locating relevant information quickly and efficiently all but impossible. Web users were crying out for tools that could help satisfy their information needs.

Search tools using two very different methods emerged to help users locate information on the Web.
  1. One method, called a Web directory, was modeled on early Internet search tools like Archie and Gopher
  2. The other method, called a search engine, drew on classic information retrieval techniques that had been widely used in closed, proprietary databases but hardly at all in the open universe of the Internet. 
In general, Web directories provide a context-based framework for structured browsing, while search engines, as their name implies, provide no context but allow searching for specific keywords or phrases. 
Web directories are similar to a table of contents in a book; search engines are more akin to an index.
Like a table of contents, a Web directory uses a hierarchical structure to provide a high level overview of major topics. Browsing the table of contents allows the reader to quickly turn to interesting sections of a book by examining the titles of chapters and subchapters. Browsing the subject-oriented categories and subcategories in a Web directory likewise leads to categories pointing to relevant Web documents.
In both cases, however, since the information is limited to brief descriptions of what you’ll find, you have no assurance that what you’re looking for is contained in either the section of the book or on a specific Web page. You must ultimately go directly to the source and read the information yourself to decide.

A book’s index offers a much finer level of granularity, providing explicit pointers to specific keywords or phrases regardless of where they appear in the book. Search engines are essentially full-text indexes of Web pages, and will locate keywords or phrases in any matching documents regardless of where they are physically located on the Web.

From a structural standpoint, a directory has the form of a hierarchical graph, with generic top-level categories leading to increasingly more specific subcategories as the user drills down the hierarchy by clicking hypertext links. Ultimately, at the bottom-most node for a branch of the hierarchical graph, the user is presented with a list of document titles hand-selected by humans as the most appropriate for the subcategory.

Search engines have no such hierarchical structure. Rather, they are fulltext indexes of millions of Web pages, organized in an inverted index structure. Whenever a searcher enters a query, the entire index is searched, and a variety of algorithms are used to find relationships and compute statistical correlations between keywords and documents. Documents judged to have the most “relevance” are presented first in a result list.
Relevance
The degree to which a retrieved Web document matches a user’s query or information need. Relevance is often a complex calculation that weighs many factors, ultimately resulting in a score that’s expressed as a percentage value.
Browsing and searching both play important roles for the searcher. By design, browsing is the most efficient way to use a directory, and searching with keywords or phrases is the best way to use a search engine. Unfortunately, the boundaries between what might be considered a “pure” directory or search engine are often blurred. For example,
  • most directories offer a search form that allows you to quickly locate relevant categories or links without browsing. 
  • Similarly, most search engines offer access to directory data provided by a business partner that’s independent of the actual search engine database. 
  • Muddying the waters further, when you enter keywords into a search engine, results are presented in a browsable list.
These ambiguities notwithstanding, it’s reasonably easy to tell the difference between a Web directory and a search engine—and it’s an important distinction to make when deciding what kind of search tool to use for a particular information need. Let’s take a closer look at Web directories and search engines, examining how they’re compiled and how they work to provide results to a searcher.


Web Directories
Web directories, such as Yahoo!, LookSmart, and the Open Directory Project (ODP) are collections of links to Web pages and sites that are arranged by subject. They are typically hierarchical in nature, organized into a structure that classifies human knowledge by topic, technically known as ontology. These ontologies often resemble the structure used by traditional library catalog systems, with major subject areas divided into smaller, more specific subcategories.
Directories take advantage of the power of hypertext, creating a clickable link for each topic, subtopic, and ultimate end document, allowing the user to successively drill down from a broad subject category to a narrow subcategory or specific document. This extensively linked structure makes Web directories ideal tools for browsing for information. To use another real-world metaphor,
directories are similar to telephone yellow pages, because they are organized by category or topic, and often contain more information than bare-bones white pages listings.

How Web Directories Work
The context and structure provided by a directory’s ontology allows its builders to be very precise in how they categorize pages. Many directories annotate their links with descriptions or comments, so you can get an idea of what a Web site or page is about before clicking through and examining the actual page a hyperlink points to.

There are two general approaches to building directories. 
  1. The closed model used by Yahoo!, LookSmart, and NBCi relies on a small group of employees, usually called editors, to select and annotate links for each category in the directory’s ontology. The expertise of directory editors varies, but they are usually subject to some sort of quality-control mechanism that assures consistency throughout the directory.
  2. The open model, as epitomized by the Open Directory Project, Wherewithal, and other less well known services, relies on a cadre of volunteer editors to compile the directory. Open projects tend to have more problems with quality control over time, but since they cost so little to compile and are often made freely available to anyone wishing to use their data, they have proven popular with the Web community at large.

Most Web directories offer some type of internal search that allows the user to bypass browsing and get results from deep within the directory by using a keyword query. Remember, however, that directories consist only of links and annotations. Using a directory’s search function searches the words making up these links and annotations, not the full-text of Web documents they point to, so it’s possible that your results will be incomplete or omit potentially good matches.

Since most directories tend to be small, rarely larger than one to two million links, internal results are often supplemented with additional results from a general-purpose search engine. These supplemental results are called “fall-through” or “fall-over” results. Usually, supplemental results are differentiated from directory results in some way. For example, a search on Yahoo! typically returns results labeled “Web Sites” since Yahoo! entries almost always point to the main page of a Web site. Fall-through results are labeled “Web Pages” and are provided by Yahoo!’s search partner.
Different search engines power fall-through results for other popular Web directories. For example, MSN and About.com use fall-through results provided by Inktomi, while LookSmart uses AltaVista results.
When a directory search fails to return any results, fall-through results from a search engine partner are often presented as primary results.
In general, Web directories are excellent resources for getting a highlevel overview of a subject, since all entries in a particular category are “exposed” on the page (rather than limited to an arbitrary number, such as 10 to 20, per page).

Issues with Web Directories
Because their scope is limited and links have been hand-selected, directories are very powerful tools for certain types of searches.
However, no directory can provide the comprehensive coverage of the Web that a search engine can. Here are some of the important issues to consider when deciding whether to use a Web directory for a particular type of search.

Directories are inherently small. It takes time for an editor to find appropriate resources to include in a directory, and to create meaningful annotations. Because they are hand compiled, directories are by nature much smaller than most search engines. This size limitation has both positives and negatives for the searcher.

  1. On the positive side, directories typically have a narrow or selective focus. A narrow or selective focus makes the searcher’s job easier by limiting the number of possible options. Most directories have a policy that requires editors to perform at least a cursory evaluation of a Web page or site before including a link to it in the directory, in theory assuring that only hyperlinks to the highest quality content are included. Directories also generally feature detailed, objective, succinct annotations, written by a person who has taken the time to evaluate and understand a resource, so they often convey a lot of detail. Search engine annotations, on the other hand, are often arbitrarily assembled by software from parts of the page that may convey little information about the page.
  2. On the negative side, Web directories are often arbitrarily limited, either by design or by extraneous factors such as a lack of time, knowledge, or skill on the part of the editorial staff.
Unseen Editorial Policies. Although many Web directories publish editorial standards and selection criteria, other factors may influence the contents of a directory. Ordinary “office politics” may favor one editor’s work over another’s, for example. With open model directories, which rely on the efforts of volunteers, a number of cases have been reported where some editors took advantage of their position to remove or block competitors’ listings.
Though most directories strive to provide objective coverage of the Web, it’s important to be vigilant for any signs of bias and the quality of resources covered.
Timeliness. Directory links are typically maintained by hand, so upkeep and maintenance is a very large issue. The constantly changing, dynamic nature of the Web means that sites are removed, URLs are changed, companies merge—all of these events effectively “break” links in a directory. Link checking is an important part of keeping a directory up to date, but not all directories do a good job of frequently verifying the links in the collection.
Directories are also vulnerable to “bait and switch” tactics by Webmasters. Such tactics are generally used only when a site stands no chance of being included in a directory because its content violates the directory’s editorial policy. Adult sites often use this tactic, for example, by submitting a bogus “family friendly” site, which is evaluated by editors. Once the URL for the site is included in the directory, the “bait” site is taken down and replaced with risqué content that wouldn’t have passed muster in the first place. The link to the site (and its underlying URL) remains the same, but the site’s content is entirely different from what was originally submitted by the Webmaster and approved by the directory editor. It’s a risky tactic, since many directories permanently ban sites that they catch trying to fool them in this manner.Unfortunately, it can be effective, as the chicanery is unlikely to be discovered unless users complain.

Lopsided coverage. Directory ontologies may not accurately reflect a balanced view of what’s available on the Web. For the specialized directories, this isn’t necessarily a bad thing. However, lopsided coverage in a general-purpose directory is a serious disservice to the searcher. Some directories have editorial policies that mandate a site being listed in a particular category even if hundreds or thousands of other sites are included in the category. This can make it difficult to find
resources in these “overloaded” categories.
Examples of huge, lopsided categories include Yahoo!’s “Business” category and the Open Directory Project’s “Society” category.
The opposite side of this problem occurs when some categories receive little or no coverage, whether due to editorial neglect or simply a lack of resources to assure comprehensive coverage of the Web.

Charging for listings. There has been a notable trend toward charging Webmasters a listing fee to be included in many directories. While most businesses consider this a reasonable expense, other organizations or individuals may find the cost to be too great. This “pay to play” trend almost certainly excludes countless valuable sites from the directories that abide by this policy.


Search Engines
Search engines are databases containing full-text indexes of Web pages. When you use a search engine, you are actually searching this database of retrieved Web pages, not the Web itself. Search engine databases are finely tuned to provide rapid results, which is impossible if the engines were to attempt to search the billions of pages on the Web in real time.

Search engines are similar to telephone white pages, which contain simple listings of names and addresses. Unlike yellow pages, which are organized by category and often include a lot of descriptive information about businesses, white pages provide minimal, bare bones information. However, they’re organized in a way that makes it very easy to look up an address simply by using a name like “Smith” or “Veerhoven.”
Search engines are compiled by software “robots” that voraciously suck millions of pages into their indices every day. When you search an index, you’re trying to coax it to find a good match between the keywords you type in and all of the words contained in the search engine’s database. In essence, you’re relying on a computer to essentially do simple pattern-matching between your search terms and the words in the index. AltaVista, HotBot, and Google are examples of search engines.


How Search Engines Work
Search engines are complex programs. In a nutshell, they consist of several distinct parts:

  1.  The Web crawler (or spider), which finds and fetches Web pages
  2.  The indexer,  indexes every word on every page and stores the resulting index of words in a huge database
  3. The query processor, which compares your search query to the index and recommends the best possible matching documents
Let’s take a closer look at each part.
Myth: All Search Engines Are Alike
Search engines vary widely in their comprehensiveness, currency, and coverage of the Web. Even beyond those factors, search engines, like people, have “personalities,” with strengths and weaknesses, admirable traits, and irritating flaws. By turns search engines can be stolidly reliable and exasperatingly flaky. And just like people, if you ask for something outside their area of expertise, you’ll get the electronic equivalent of a blank stare and a resounding “huh?”
A huge issue is the lack of commonality between interfaces, syntax, and capabilities. Although they share superficial similarities, all search engines are unique both in terms of the part of the Web they have indexed, and how they process search queries and rank results. Many searchers make the mistake of continually using their “favorite” engine for all searches. Even if they fail to get useful results, they’ll keep banging away, trying different keywords or otherwise trying to coax an answer from a system that simply may not be able to provide it. A far better approach is to spend time with all of the major search engines and get to know how they work, what types of queries they handle better than others, and generally what kind of “personality” they have. If one search engine doesn’t provide the results you’re looking for, switch to another. And most important of all, if none of the engines seem to provide reasonable results, you’ve just got a good clue that what you’re seeking is likely to be located on the Invisible Web—if, in fact, it’s available online at all.

WEB CRAWLERS
Web crawlers have the sole mission of finding and retrieving pages on the Web and handing them off to the search engine’s indexers. It’s easy to imagine a Web crawler as a little sprite scuttling across the luminous strands of cyberspace, but in reality Web crawlers do not traverse the Web at all. In fact, crawlers function much like your Web browser, by sending a request to a Web server for a Web page, downloading the entire page, then handing it off to the search engine’s indexer.
Crawlers, of course, request and fetch pages much more quickly than you can with a Web browser. In fact most Web crawlers can request hundreds or even thousands of unique pages simultaneously. Given this power, most crawlers are programmed to spread out their requests for pages from individual servers over a period of time to avoid over-whelming the server or consuming so much bandwidth that human users are crowded out. Crawlers find pages in two ways. 
  1. Most search engines have an “add URL” form, which allows Web authors to notify the search engine of a Web page’s address. In the early days of the Web, this method for alerting a search engine to the existence of a new Web page worked well—the crawler simply took the list of all URLs submitted and retrieved the underlying pages.Unfortunately, spammers figured out how to create automated bots that bombarded the add URL form with millions of URLs pointing to spam pages. Most search engines now reject almost 95 percent of all URLs submitted through their add URL forms. It’s likely that, over time, most search engines will phase-out their add URL forms in favor of the second method that crawlers can use to discover pages—one that’s more easy to control.
  2. This second method of Web page discovery takes advantage of the hypertext links embedded in most Web pages. When a crawler fetches a page, it culls all of the links appearing on the page and adds them to a queue for subsequent crawling. As the crawler works its way through the queue, links found on each new page are also added to the queue. Harvesting links from actual Web pages dramatically decreases the amount of spam a crawler encounters, because most Web authors only link to what they believe are high-quality pages.
By harvesting links from every page it encounters, a crawler can quickly build a list of links that can cover broad reaches of the Web. This technique also allows crawlers to probe deep within individual sites, following internal navigation links. In theory, a crawler can discover and index virtually every page on a site starting from a single URL, if the site is well designed with extensive internal navigation links.
Although their function is simple, crawlers must be programmed to handle several challenges. First, since most crawlers send out simultaneous requests for thousands of pages, the queue of “visit  soon” URLs must be constantly examined and compared with URLs already existing in the search engine’s index. Duplicates in the queue must be eliminated to prevent the crawler from fetching the same page more than once. If a Web page has already been crawled and indexed, the crawler must determine if enough time has passed to justify revisiting the page, to assure that the most up-to-date copy is in the index. And because crawling is a resource-intensive operation that costs money, most search engines limit the number of pages that will be crawled and indexed from any one Web site. 
This is a crucial point—you can’t assume that just because a search engine indexes some pages from a site that it indexes all of the site’s pages.
Because much of the Web is highly connected via hypertext links, crawling can be surprisingly efficient. A May 2000 study published by researchers at AltaVista, Compaq, and IBM drew several interesting conclusions that demonstrate that crawling can, in theory, discover most pages on the visible Web (Broder et al., 2000). The study found that:
  • For any randomly chosen source and destination page, the probability that a direct hyperlink path exists from the source to the destination is only 24 percent.
  • If a direct hypertext path does exist between randomly chosen pages, its average length is 16 links. In other words, a Web browser would have to click links on 16 pages to get from random page A to random page B. This finding is less than the 19 degrees of separation postulated in a previous study, but also excludes the 76 percent of pages lacking direct paths.
  •  If an undirected path exists (meaning that links can be followed forward or backward, a technique available to search engine spiders but not to a person using a Web browser), its average length is about six degrees.
  •  More than 90 percent of all pages on the Web are reachable from one another by following either forward or backward links. This is good news for search engines attempting to create comprehensive indexes of the Web.
These findings suggest that efficient crawling can uncover much of the visible Web.

SEARCH ENGINE INDEXERS
When a crawler fetches a page, it hands it off to an indexer, which stores the full text of the page in the search engine’s database, typically in an inverted index data structure. An inverted index is sorted alphabetically, with each index entry storing the word, a list of the documents in which the word appears, and in some cases the actual locations within the text where the word occurs. This structure is  ideally suited to keyword-based queries, providing rapid access to documents containing the desired keywords.

As an example, an inverted index for the phrases “life is good,” “bad or good,” “good love,” and “love of life” would contain identifiers for each phrase (numbered one through four), and the position of the word
within the phrase.
Table 2.2 shows the structure of this index.
To improve search performance, some search engines eliminate common words called stop words (such as is, or, and of in the above example). Stop words are so common they provide little or no benefit in narrowing a search so they can safely be discarded. The indexer may also take other performance-enhancing steps like eliminating punctuation and multiple spaces, and may convert all letters to lowercase.

Some search engines save space in their indexes by truncating words to their root form, relying on the query processor to expand queries by adding suffixes to the root forms of search terms. Indexing the full text of Web pages allows a search engine to go beyond simply matching single keywords. If the location of each word is recorded, proximity operators such as NEAR can be used to limit searches. The engine can also match multi-word phrases, sentences, or even larger chunks of text. If a search engine indexes HTML code in addition to the text on the page, searches can also be limited to specific fields on a page, such as the title, URL, body, and so on.


THE QUERY PROCESSOR
The query processor is arguably the most complex part of a search engine. The query processor has several parts, including
  1. the primary user interface (the search form), 
  2. the actual “engine” that evaluates a query and matches it with the most relevant documents in the search engine database of indexed Web pages, and 
  3. the results-output formatter
The search form and the results format vary little from search engine to search engine. All have both basic and advanced search forms, each offering slightly varying limiting, control, and other user-specified functions. And most result formats are equally similar, typically displaying search results and a few additional extras like suggested related searches, most popular searches, and so on.

The major differentiator of one search engine from another lies in the way relevance is calculated. Each engine is unique, emphasizing certain variables and downplaying others to calculate the relevance of a document as it pertains to a query.
  • Some engines rely heavily on statistical analysis of text, performing sophisticated pattern-matching comparisons to find the most relevant documents for a query. 
  • Others use link analysis, attempting to capture the collective wisdom of the Web by finding the documents most cited by other Web authors for a particular query.
How an engine calculates relevance is ultimately what forms its “personality” and determines its suitability for handling a particular type of query.
Search engine companies closely guard the formulas used to calculate relevance, and change them constantly as algorithms are updated for improved quality or tweaked to outwit the latest technique used by spammers. 
Nonetheless, over time, a searcher can generally get to know how well a particular engine will perform for a
query, and select an engine appropriately.
Myth: Search Engine Indexes Are Current
Crawling the Web is an expensive, resource-intensive operation. It costs a search engine a certain amount of money each time it retrieves and indexes a page. Once a page is included in a search engine’s index, the engine may not recrawl the page for an extended period of time. This may be because the crawler is smart enough to know that the page doesn’t change frequently, but more often it’s because limited resources require that the crawler focus on newer or more popular areas of the Web.
A lack of freshness in a search engine index can lead to bizarre or irrelevant results if a Web page has changed since the last time a search spider indexed the page. Let’s look at an example.
Say that on June 1, a Web author creates a Web page that has 2,500 words in 10 paragraphs. The author may or may not submit the URL of the page to search engines requesting that they spider the page.
Nonetheless, some people on the Web create links to the page, so
Web crawlers can find it anyway.
On July 20, a spider from search engine A crawls the page and
includes the page in its index.
On August 20, a spider from search engine B crawls the same page.
The page is now included in two separate indexes, and a search for
keywords on the page will return an accurate result from both engines.
But then on September 20, the Web author decides to make some
changes, and eliminates paragraphs 3, 4, and 5 from the document.
The author does not notify the engines that he has changed the page (by resubmitting the URL to the “submit URL” form provided by each
engine). The copy of the page in the search engines’ indexes no longer
matches the actual page on the Web.
On October 5, a user of search engine A submits a query that causes
the Web page to be displayed in the result list. However, the keywords
in the query that caused the engine to include the page in the result list
were in the paragraphs removed by the page author on September 20th.
In this case, the search engine has returned an accurate result based on the copy of the page in its index, but based its relevance ranking on an old version of the page that no longer contains what the searcher is looking for. If the searcher clicks through and reads the updated page he won’t find what he is looking for, leaving the searcher puzzled and confused.
The searcher will likely blame the “poor” result on the “dumb” search
engine, when in fact the problem was simply that the engine was calculating relevance using a copy of a document that was out of date.
Most of the major search engines have stepped up efforts to recrawl Web pages on a more frequent basis. Many are employing “smart” crawlers that learn to revisit frequently changing pages more often. Others are simply throwing more resources and money at the problem, attempting to keep up with the rapidly changing Web by brute force. Nonetheless, the currency of all search indexes is a problem every
searcher should consider.
Issues with Search Engines
Just as Web directories have a set of issues of concern to a searcher, so do search engines. Some of these issues are technical; others have to do with choices made by the architects and engineers who create and maintain the engines.

Cost of crawling. Crawling the Web is a resource-intensive operation. The search engine provider must maintain computers with sufficient power and processing capability to keep up with the explosive growth of the Web, as well as a high-speed connection to the Internet backbone. It costs money every time a page is fetched and stored in the search engine’s database. There are also costs associated with query processing, but in general crawling is by far the most expensive part of maintaining a search engine.

Since no search engine’s resources are unlimited, decisions must be made to keep the cost of crawling within an acceptable budgetary range.
  1. Some engines limit the total number of pages in their index, dumping older pages when newer ones are found. 
  2. Others limit the frequency of recrawl, so pages in the index may be stale or out of date. 
  3. Still others limit their crawling to certain portions or domains believed to contain reliable, non-duplicated material.
Whenever an engine decides to limit its crawling, it means pages that potentially could be included in its index are not. It’s tempting to think that these unretrieved pages are part of the Invisible Web, but they aren’t. They are visible and indexable, but the search engines have made a conscious decision not to index them. Competent searchers must take this into account when planning a research strategy.

Much has been made of these overlooked pages, and many of the major engines are making serious efforts to include them and make their indexes more comprehensive. Unfortunately, the engines have also discovered through their “deep crawls” that there’s a tremendous amount of duplication and spam on the Web.
The trade-off between excluding bogus material and assuring that all truly relevant material will be found is a difficult one, and virtually assures that no engine will ever have a totally comprehensive index of the Web.
“Dumb” crawlers. At their most basic level, crawlers are uncomplicated programs. Crawlers are designed simply to find and fetch Web pages.
To discover unindexed pages, crawlers rely on links they’ve discovered on other pages. If a Web page has no links pointing to it, a search engine spider cannot retrieve it unless it has been submitted to the search engine’s “add URL” form.
Another problem with search engines compiled by crawlers is simply that it takes a lot of time to crawl the entire Web, even when the crawler is hitting millions of pages a day. Crawler lag time is a two-fold issue:
  1. First, there’s typically a gap between when a page is published on the Web and when a crawler discovers it. 
  2. Second, there’s a time lag between when a crawler first finds a page and when it recrawls the page looking for fresh content.
Both of these time issues can contribute to incomplete or inaccurate search results.

Current generation crawlers also have little ability to determine the quality or appropriateness of a Web page, or whether it is a page that changes frequently and should be recrawled on a timely basis.

User expectations and skills. Users often have unrealistic expectations of what search engines can do and the data that they contain. Trying to determine the best handful of documents from a corpus of millions or billions of pages, using just a few keywords in a query is almost an impossible task. Yet most searchers do little more than enter simple two- or three-word queries, and rarely take advantage of the advanced limiting and control functions all search engines offer.

Search engines go to great lengths to successfully cope with these woefully basic queries.
  1. One way they cope is to create a set of preprogrammed results for popular queries. For example, if a pop star releases a hit song that generates lots of interest, the engine may be preprogrammed to respond with results pointing to biographical information about the star, discographies, and links to other related music resources. 
  2. Another method is to adjust results so that the most popular, commonly selected pages rise to the top of a result lists.
These are just two techniques search engines apply to help users who can’t or won’t take advantage of the powerful tools that are available to them.
Taking the time to learn how a search engine works, and taking advantage of the full capabilities it offers can improve search results dramatically.
Speedy Response vs. Thorough Results. Now that we live on “Internet time,” everyone expects nearly instantaneous results from search engines. To accommodate this demand for speed, search engines rarely do as thorough an analysis as they might if time were not an issue. Shortcuts are taken, total results are truncated, and invariably important documents will be omitted from a result list.

Fortunately, increases in both processing power and bandwidth are providing search engines with the capability to use more computationally intensive techniques without sacrificing the need for speed. Unfortunately, the relentless growth of the Web works against improvements in computing power and bandwidth simply because as the Web grows the search space required to fully evaluate it also increases.

Bias toward text. Most current-generation search engines are highly optimized to index text. If there is no text on a page—say it’s nothing but a graphic image, or a sound or audio file—there is nothing for the engine to index. For non-text objects such as images, audio, video, or other streaming media files, a search engine can record, in an Archie-like manner, filename and location details but not much more. While researchers are working on techniques for indexing non-text objects, for the time being non-text objects make up a considerable part of the
Invisible Web.

Myth: Search Engines Overlap in Coverage
The Web is commonly described as a huge haystack of information, with searchers looking for the elusive needle. In the words of Dr. Matthew Koll, an early pioneer in recognizing and writing about the Invisible Web, there are many ways that a search engine can address the needle in the haystack problem (Koll,1998).
It can consider:
A known needle in a known haystack
A known needle in an unknown haystack
An unknown needle in an unknown haystack
Any needle in a haystack
The sharpest needle in a haystack
Most of the sharpest needles in a haystack
All the needles in a haystack
Affirmation of no needles in the haystack
Things like needles in any haystack
Let me know whenever a new needle shows up
Where are the haystacks?
Needles, haystacks—whatever.
We show that the search engines don’t index the entire Web. It’s tempting to think that if you simply combine results from all of the major search engine indexes you’ll get a much more comprehensive map of the Web, but this isn’t true. Greg Notess, owner of Search Engine Showdown, regularly tests the major search engines for overlap. His work demonstrates that there is surprisingly little overlap between the major search engines and directories (Notess, 2000).
In one recent analysis, Notess compared the results of five small searches run on 14 different search engines. The five searches found a total of 795 hits, 298 of which represented unique pages. Of those 298 hits, 110 were found by only one of the 14 search engines, while another 79 were found by only two.
Given the increasing use of measures such as clickthrough analysis
and link popularity, the major engines will have much more overlap of popular, well-linked pages than on more obscure, less popular pages. If you’re searching for something popular or timely, you need be less concerned about the lack of overlap than if you’re looking for something unusual or rare. In this case, it’s vital to use more than one search engine to assure comprehensive results.

Search Engines vs. Directories
To summarize, search engines and Web directories both have sets of features that can be useful, depending on the searcher’s information need. Table 2.3 compares these primary features.

Early search engines did a competent job of keeping pace with the
growth of simple text documents on the Web. But as the Web grew, it
increasingly shifted from a collection of simple HTML documents to a
rich information space filled with a wide variety of data types. More
importantly, owners of content stored in databases began offering Web
gateways that allowed user access via the Web. This new richness of
data was beyond the capability of search engine spiders, designed only
to handle simple HTML text pages.

Specialized and Hybrid Search Tools

Directories and search engines are the two fundamental tools available to the Web searcher. However, there are also a variety of specialized and hybrid search tools that should be part of any searcher’s arsenal.
These tools take many forms (targeted directories and focused crawlers, vertical portals (vortals), metasearch engines, value-added search services, “alternative” search tools, and fee-based Web services.), and sometimes they are confused with Invisible Web resources. They are not—they simply extend or enhance the capabilities of search engines and directories that focus on material that’s part of the visible Web.
These alternative tools have both positives and negatives for the searcher. On the plus side, these tools tend to be smaller, and are focused on specific categories or sources of information.
  1. Because they are small relative to the general-purpose search engines and directories, they are often more reliable and stable. 
  2. These tools tend also to be updated more frequently than the general-purpose tools. 
  3. Their interfaces are often optimized for maximum utility for the underlying information they search. 
  4. Because they are limited to specific categories, they often offer greater depth of coverage for a category than the general-purpose tools.

This flexibility also can be perceived as a minus.
  1. The searcher must learn how to use multiple interfaces, the specialized commands and other unique capabilities offered by each tool. 
  2. These hybrid resources also require an investment in time to understand the scope of the materials they cover. In many ways, they are close kin of the books found in a ready reference section of a library, as they are often focused to answer specific questions, quickly and easily.


Targeted Directories and Focused Crawlers
Focus on a specific subject area or domain of knowledge. Just as you have always used a subject-specific reference book to answer a specific question, targeted search tools can help you save time and pinpoint specific information quickly on the Web. Why attempt to pull a needle from a large haystack with material from all branches of knowledge when a specialized tool allows you to limit your search in
specific ways as it relates to the type of information being searched?
Searching a smaller, targeted universe of data can increase your precision because there is potentially less unrelated or irrelevant information to get in the way.
As an example, consider the problem of looking up a friend’s phone number. Intuitively, most of us would instantly reach for a phone book rather than an encyclopedia. Why? Because phone books are specialized resources that are designed solely for the purpose of looking up phone numbers. Encyclopedias, on the other hand, are general-purpose reference works containing a wide range of information on a huge variety of subjects—and most people would rightly think it’s ridiculous to use an encyclopedia to look up a phone number.

Yet this is exactly what many of those same people do when they use a general-purpose search engine or directory such as AltaVista or Yahoo! when there are far more appropriate search tools available for a specialized task.
The problem is partly that many people think search engines and Web directories are somehow all encompassing, able to process any request for information. They may not realize that there are specialized resources that will do a much better job of satisfying their information need.
Targeted directories and crawlers are functionally similar to general-purpose Web directories and search engines, but will often yield much higher quality results for certain types of searches, for a number of reasons. Let’s take a closer look at both of these types of specialized Web search tools.

Targeted Directories
Targeted directories are Web guides compiled by humans that focus on a particular specialized topic or area of information. They are generally compiled by subject matter experts whose principal concerns are quality, authority, and reliability of the resources included in the directory.
Standards for inclusion in a targeted directory tend to be more rigorous than for general-purpose Web directories. And, quite often, annotations are extensive, almost functioning as an abstract of a Web resource rather than the more common brief descriptions offered by general-purpose directories.

EXAMPLES OF TARGETED DIRECTORIES
  1. EEVL (Edinburgh Engineering Virtual Library) EEVL is a guide to engineering information on the Internet. It is a free service, created and run by a team of information specialists from Heriot-Watt University, with input from a number of other universities in the U.K. The site features a catalogue of quality engineering resources (selected by subject consultants), targeted engineering search engines, bibliographic and events databases, including the Recent Advances in Manufacturing bibliographic database, a directory of science and technology librarians, an Engineering on the Internet bibliography, and Hot Links to useful sites that offer many specialized ways of limiting your search for very specific types of information. 
  2. Classics Resources on the Web An extensive directory of literary, linguistic, historical, and philosophical resources developed and maintained by the Humanities faculty at Cambridge University. 
  3. The Internet Intelligence Index The Internet Intelligence Index is designed to help you gather competitive intelligence information. It contains links to more than 600 intelligence-related Internet sites, covering everything from macro-economic data to individual patent and stock quote information. The
    Internet Intelligence Index is maintained by Fuld & Company, a leading firm in the field of Competitor Intelligence.
  4. Odden’s Bookmarks Subtitled “The Fascinating World of Maps and Mapping,” this “oddly” named directory (after its compiler, Roelof P. Oddens) features more than 11,000 cartographic links in areas including maps and
    atlases, sellers of cartographic material, map collections, cartographic
    and geoservers, cartographic and geographical societies, departments
    of cartography, government cartography, and many more.
Targeted directories offer several advantages over general-purpose Web directories such as Yahoo! or LookSmart. These advantages include:

Comprehensive subject coverage. Many of the pages included in targeted directories are available via general-purpose directories, but targeted directories are more likely to have comprehensive coverage of a particular subject area. In essence, they are “concentrated haystacks” where needles are more likely to be the size of railroad spikes, and much easier to find.

Up-to-date listings. People who compile and maintain targeted directories have a vested interest in keeping the directory up to date, as they are usually closely and personally associated with the directory by colleagues or other professionals in the field. Keeping the directory up to date is often a matter of honor and professional reputation for the compiler.

Listings of new resources. Since the compilers of targeted directories are often subject-matter experts in their field, they closely follow new developments in their fields. They also can rely on a network of other people in the field to monitor new developments and alert them when a new Web resource is available. Site developers are also motivated to inform targeted directories of new resources for the prestige factor of securing a listing. Finally, many focused directories are compiled by non-profit organizations, which are not subject to the occasional subtle pressure of bias from advertisers.


Focused Crawlers
Like targeted directories, focused crawlers center on specific subjects or topics.
But just as general-purpose search engines are far more comprehensive than Web directories, focused crawlers attempt to find and index as many pages as possible within a particular subject area by spidering a limited domain of Web sites known to cover a particular topic.
Focused crawlers will often spider sites more thoroughly and frequently than a general-purpose search engine. Because focused crawlers cover a much smaller portion of the Web than general-purpose search engines, the cost of crawling is much less significant. Most focused crawlers are limited to specific sites by editors who supply a list of URLs to be crawled, rather than randomly crawling the Web. Editorial selection of URLs to crawl reduces the “signal to noise” ratio in the resulting index, by limiting the crawler to sites that have been determined to be appropriate for the specialized index. This doesn’t mean links to new resources are ignored, however. In fact, focused crawlers can often discover new resources and add them to their indexes more quickly than general-purpose search engines by following newly created links from a trusted relevant Web site.

EXAMPLES OF FOCUSED CRAWLERS
  1. LawCrawler LawCrawler is a legal search tool and database that provide precision by enabling searches to be focused on sites and within specific domains known to feature high-quality legal information. LawCrawler provides the ability to limit your search to certain types of law related resources. These resources include a legal dictionary, legal NewsLaw reviews, U.S. government sites, the U.S. Constitution, the U.S. Legal Code, U.S. Supreme Court opinions, and information published by all federal circuit courts—even worldwide sites with legal information. Because LawCrawler is powered by the AltaVista search engine software, the searcher can also employ any of the advanced search capabilities provided by AltaVista, but the search is usefully restricted to the specific legal information domains indexed by LawCrawler. LawCrawler is part of the FedLaw service, which has numerous other legal resources, including a targeted directory.
  2.  PsychCrawler This site is sponsored by an organization with knowledge of the topic, the American Psychological Association, which has a vested interest in making sure that high-quality material is crawled.
  3. PoliticalInformation.Com This focused search engine contains an index of the contents of more than 5,000 Web sites chosen for the quality of their content. Targeted sites are recrawled every 2-3 weeks to keep the results current. The site features links to over 270,000 documents solely related to politics and the political process. 
Some focused crawlers do nothing more than limit searches to specific top-level domains. Most sites on the Web are in the dot-com top-level domain. But other top-level domains are restricted to certain types of organizations or individuals. Dot-mil (.mil) for example, is for U.S. military Web sites. Dot-edu (.edu) is for educational institutions, dot-gov (.gov) for U.S. government, and so on. Some examples follow.

  1.  SearchMil.Com SearchMil.com specializes in military focused searches, combining exhaustive coverage of the dot-mil domain with powerful search engine technology that ranks results in order of popularity. SearchMil’s index contains over 1 million military-specific pages.
  2. SearchEdu.Com SearchEdu.com is a focused search engine devoted to university and education-related Web sites. SearchEdu’s index contains more than 20 million pages of academic materials from universities, schools, and libraries, and results are ranked in order of popularity.
  3. UsaGov.Gov  is a focused crawler designed to be a “one stop” portal for government information of all kinds in the United States. The site has more than 20 million pages indexed, with more added every day. UsaGov also features a high-quality directory of government sites, so it’s actually a hybrid consisting of both a targeted directory and focused crawler.


Vertical Portals (Vortals)
Vertical Portals (also known as “Vortals”) are mini-versions of general-purpose search engines and directories that focus on a specific topic or subject area. They are often made up of both a targeted directory and listings compiled by a focused crawler. Vortals are most often associated with “Business to Business” (B2B) sites, catering to both the information and commerce needs of particular industries or services. While many Vortals provide great information resources for searchers, their primary focus is usually on providing a virtual marketplace for the trade of goods and services. As such, Vortals can be excellent resources for researchers looking for current information on business concerns ranging from manufacturing to global trade.
The searcher should beware, however, that Vortals often have an unabashed profit motive that may introduce a degree of bias into the information they provide. We’re not suggesting you avoid Vortals; rather, just keep a more vigilant eye out for quality or bias than you normally might with “pure” Web guides.

How to Find Vortals
There are thousands of Vortals on all conceivable subjects. Rather than providing specific examples, here are two focused directories that specialize in Vortals.
  1. VerticalPortal.com  A directory of vertical portals, defined as sites that are big enough to be considered a “destination,” and generally have community interaction, topical news, and various other portal features. 
  2. Yahoo! B2B Marketplace is a product directory of Vortal sites on the Web. Search by keyword or browse through many commerce Web sites for products organized by category.

Metasearch Engines
Metasearch engines submit queries to multiple search engines and Web directories simultaneously. Rather than crawling the Web and building its own index, a metasearch engine relies on the indices created by other search engines. This allows you to quickly get results from more than one general-purpose search engine.
While some searchers swear by metasearch engines, they are somewhat limited because they can only pass through a small set of the advanced search commands to any given engine. Searchers typically take a “least common denominator” approach that favors quantity of results over refined queries that might return higher quality results.

EXAMPLES OF METASEARCH ENGINES
  1. MetaCrawler  sends your query to most of the major search engines and directories, and then combines the results in a single list, eliminating duplicates. Unlike the other metasearch engines reviewed here, MetaCrawler sums the scores given by each service it uses, and presents them in a “voted” ordering, with the score (from 1 to 1,000) presented in bold type next to each result. You can customize MetaCrawler to use your favorite search engines, increase or decrease the results per page and per source, set the timeout duration you’re willing to wait for results to be returned, and limit the domains that are searched.
  2. Dogpile The rather unusually named Dogpile searches more than two dozen major search sites, including Usenet newsgroups, FTP sites for file downloads, newswires for current headlines, and business news from several sources. Unlike MetaCrawler, which post-processes the results using its own relevance algorithms, Dogpile returns the exact results from each engine with the utmost haste. Results are presented as each engine replies to the query. Dogpile searches three sites at a time.
    Results from each search engine are grouped together, with the descriptions provided by each site. At the bottom of the results list is a button that continues your search with the next three search engines.


Issues with Metasearch Engines
Metasearch engines attempt to solve the haystack problem by searching several major search engines simultaneously. The idea is to combine haystacks to increase the probability of finding relevant results for a search. The trap many people fall into is thinking that metasearch engines give you much broader coverage of the Web. Searchers reason that since each engine indexes only a portion of the Web, by using a metasearch engine, the probability of finding documents that an individual engine might have missed is increased. In other words, combining partial haystacks should create a much more complete, larger haystack.

In theory, this is true. In practice, however, you don’t get significantly broader coverage. Why? Because the metasearch engines run up against the same limitations on total results returned that you encounter searching a single engine. Even if search engines report millions of potential matches on your keywords, for practical reasons, all of the engines limit the total number of results you can see—usually between 200 and 1,000 total results. Results found beyond these arbitrary cutoff points are effectively inaccessible without further query refinement. Quite the opposite of searching whole haystacks with great precision, you’re actually searching only portions of partial haystacks with less precision!
While metasearch engines aren’t necessarily the best tools for improving the precision of a search, they do provide several other useful functions by “combining haystacks.”
  • Use a metasearch engine when you’re just beginning a search on an unfamiliar topic, to get a sense of what keywords might be the most effective for specific engines. 
  • Use a metasearch engine when you want extremely quick and dirty results for a popular one- or two-word query. 
  • Finally, use these engines when you want to quickly compare and contrast how the major services process a particular query.
Remember that each metasearch engine searches different sources. Be sure to examine the list of engines searched to assure that your results from your favorite engines are included.

Metasearch engines generally do nothing more than submit a simple search to the various engines they query. They aren’t able to pass on advanced search queries that use Boolean or other operators to limit or refine results.
Put another way, metasearch engines are increasing the potential relevance of search results at the expense of precision.
Quite the opposite of searching whole haystacks, you’re actually searching only portions of partial haystacks.


Value-Added Search Services
Some Web search services combine a general-purpose search engine or directory with a proprietary collection of material that’s only available for a fee. These value-added search services try to combine the best of both worlds: free resources available on the Web, and high-quality information offered by reputable information publishers.

Many of these services evolved from models used by early consumer-oriented services like the Source, CompuServe, Prodigy, and, of course, today’s most popular closed information system, America Online (AOL). Unlike the consumer services, which require a subscription to the complete service to gain access to any of its parts, value-added search services on the Web tend to offer varying levels of access based on need and willingness to pay. Most offer at least some free information (sometimes requiring filling out a simple registration form first). The premium information offered by these services is generally first rate, and well worth the modest fees if the searcher is looking for authoritative, high-quality information.

EXAMPLES OF VALUE-ADDED SEARCH SERVICES
  1. Northern Light  The Northern Light Special Collection is an online business library comprising more than 7,000 trusted, full-text journals, books, magazines, newswires, and reference sources. The breadth of information available in the Special Collection is unique to Northern Light, and includes a wide range of diverse sources such as American Banker, ENR: Engineering News Record, The Lancet, PR Newswire, and ABC News Transcripts. This content is fully searchable and results are seamlessly integrated with Web search results.
    Most Special Collection documents range in price from $1.00 to $4.00 per article, with a few higher value sources, such as WEFA and Investext reports costing more. Unlike expensive subscription services that require you to commit to pay up front before you can tell how much you will use, Northern Light will only ask you to pay for what you actually read. They also have an automated “return” policy where your money is refunded for documents that don’t meet your needs.
  2. The Electric Library The Electric Library offers free searching and provides brief abstracts of results on its Web site, but charges a subscription fee for full access.
    The service is often compared to Northern Light, though the Electric Library uses a subscription model with no transactional fees. It is icensed to more than 15,000 schools and libraries, and has more than 80,000 individual subscribers. Subscriptions are available on a monthly ($9.95) or annual ($59.95) basis.
    The Electric Library Personal Edition is also unique because its database contains only copyrighted content. Licensed content includes material from 400 publishers, with more than 1,000 titles. Segregated into six categories, the Electric Library contains more than 5.5 million newspaper articles, nearly 750,000 magazine articles, 450,000 book chapters, 1,500 maps, 145,000 television and radio transcripts, and 115,000 photos and images. Fully 95 percent of the content in Electric Library isn’t available elsewhere on the Web. 
  3. EDGAR Online Throughout the year, every U.S. public company is required to disclose the critical business, financial, and competitive details of its activ ities to the Securities and Exchange Commission. EDGAR Onlineprovides fast and easy access to this SEC information. Visitors to the site get limited access to information on individuals mentioned in proxy statements, historical filings back to 1994, today’s filings, and full search by any of ten criteria, including company name, ticker symbol, filing type, industries, and sector. For professionals or others who require more advanced options, EDGAR Online Premium offers subscribers real-time e-mail notification and advanced management tools that can pinpoint the critical business, financial, and competitive information contained in every SEC filing.

Alternative Search Tools
Most people think of searching the Web as firing up a browser and pointing it to a search engine or directory home page, typing keyword queries, and reading the results on the browser screen. While most traditional search engines and directories do indeed operate this way, there are a number of alternative search tools that both transcend limitations and offer additional features not generally available otherwise.

Browser Agents
Browser agents are programs that work in conjunction with a Web browser to enhance and extend the browser’s capabilities. While these tools focus on “search,” they work differently from traditional keyword-based search services. The tools attempt to automate the search process by analyzing and offering selected resources proactively. Others go beyond simple keyword matching schemes to analyze the context of the entire query, from words, sentences, and even entire paragraphs.

 Browser agents can be very handy add-on tools for the Web searcher.
The downside to all of these tools is that they are stand-alone pieces of
software that must be downloaded, installed, and run in conjunction
with other programs. This can quickly lead to desktop clutter, and
potential system conflicts because the programs are not designed to
work together. The best strategy for the searcher is to sample a number
of these programs and select one or two that prove the most useful for
day-to-day searching needs.

Client-Based Search Tools
Client-based Web search utilities (sometimes called “agents” or “bots”) reside on your own computer or network, not on the Internet. Like other software, they are customizable to meet your personal needs. Because the software resides on your own computer, it can “afford” to be more computationally intensive in processing your search queries.
Client-based Web search utilities have a number of key features generally not found on Web-based search engines and directories.
  • They search multiple sites simultaneously, and you can choose which ones are used.
  • They can eliminate dead links from results.
  • They can download pages to your computer for more extensive analysis and faster browsing.
  • Their search refinement tools are often very sophisticated.
  • They generate reports that can be saved, repeated, or automatically updated.
  • They can, to a limited degree, access Invisible Web sites that general-purpose search engines cannot penetrate.
Copernic is a search agent that can access more than 600 search engines and specialized information sources grouped in 55 categories. Categories include the Web, newsgroups, e-mail addresses, books, business, games, movies, music, newspapers, software, sports news, tech news, and others.
Copernic allows you to schedule search updates and get e-mail alerts when pages you are tracking have changed. Copernic also does a good job at presenting results, removing advertisement banners, and filtering out useless information and dead links. Copernic supports both Boolean searching and natural language queries. Like browser agents, client-based Web search tools have both pros
and cons for the searcher. If you find yourself running the same search repeatedly and would like to track how results change over time, these utilities can be a godsend. Similarly, their ability to access some Invisible Web resources makes them a useful addition to (but not substitute for) your Invisible Web search arsenal.


Web Rings
Simply put, a Web ring is a group of Web pages that are linked to one another, ostensibly because they all share a similar topic or interest. Uniform navigation buttons let you travel from page to page in the ring until you end up back at your starting point. In effect, a Web ring is a “closed circuit” of Web sites, linked together by a common theme.

These themes are what make Web rings interesting from a searching perspective. As part of a Web ring, a Web site is voluntarily linking with other Web sites that have similar characteristics. Unlike the random categorizations generated by search engines, Web rings allow Webmasters to make direct and specific affiliations with other sites that they think are related to their own. And unlike a list of links, which can become outdated or broken over time, Web ring URLs are kept in a central database that keeps the membership—and links to participating
sites—fresh and up to date.
For more information on Web Rings, see the Web Ring home page.


Fee-Based Web-Accessible Services
Finally, there is an entire class of search tools accessible via the Web, but which do not search the Web themselves. Often, these search tools are proprietary systems that have existed for many years prior to the Web, but in recent years have created Web interfaces for easier access by clients.
The most significant of these services are provided by the traditional online information companies, which over the years have compiled huge proprietary databases of information. These companies include familiar names such as Thomson’s Dialog, LexisNexis, and Dow-Jones. Many of these companies got their start in the very early days of the information-processing industry during the 1970s.
Information professionals love these services, because they aggregate data from numerous sources and providers, serving as huge marts of information—many times larger than what is available on the Web. The twin hallmarks of these services are extremely sophisticated querying and reporting tools, and “massaged” databases where information quality is first rate. As you might expect, gaining access to this sort of premium content is quite expensive.
 Myth: Search Indexes Are Comprehensive
Perhaps the most pervasive myth about search engines is that they provide comprehensive coverage of the Web. In fact, exactly the opposite is true: no search engine knows about every page on the Web, and most include only 20-50 percent of even visible Web pages, let alone Invisible Web content. Why?
Part of the problem, of course, is that search crawlers simply can’t keep up with the explosive pace of growth on the Web. With millions of new pages added daily, it’s too expensive and resource-intensive to attempt to find every new page added to the Web.
A second problem is that crawlers simply can’t find all of the pages on the Web. This can be either because there are no links pointing to some pages (and crawlers can only discover pages by following links), or because the crawler simply can’t or won’t access some types of content. This type of content, is genuinely part of the Invisible Web.
A third reason has to do with the individual characteristics of each search engine, and the rules it follows for gathering and indexing Web pages. Each engine is unique, and they all follow different rules in gathering material. For example:
• Each engine re-crawls Web pages on its own unique schedule.
• Some learn about pages via user submissions, while others do not.
• Some types of content, while technically crawlable, pose chal-
lenges or difficulties for spiders. These pages may use frames, or
consist of code or objects that do not make spiders happy.
• Webmasters may choose to voluntarily exclude pages by using
the Robots Exclusion Protocol, which prevents a crawler from
accessing and indexing a page.
• Some engines arbitrarily drop pages from their indexes, either
deliberately to make room for new pages, or inadvertently, creating a nettlesome situation for both searchers and Webmasters looking for pages that simply have been “booted out” of the index for no apparent reason.
Webmasters themselves are often to blame for broken links by moving or removing pages on a server. Material on the Web can be changed, added, or removed in just a few seconds. URLs are very explicit addresses—they point to specific filenames in specific directories on Web servers. If a file is renamed, or moved to another directory, its indexed URL automatically becomes broken, and the search engine may not be able to find the new location of the page on its own.
The lack of comprehensive coverage may or may not be a problem for a searcher. Popular, heavily trafficked sites are generally well represented in most general-purpose search engines. More obscure or less-frequently visited sites may not be. However, simply being omitted from a search engine index does not make a page or a site invisible. If a Web site consists of simple HTML pages and has no roadblocks in place to block crawlers, it’s firmly in the realm of the visible Web, because at any minute a crawler may discover and index the site. Nonetheless, most Web content that falls into this category will be very difficult to find unless the searcher knows exactly where to look. Even though technically visible, content that’s not included in search engine indexes is all but invisible to most searchers. We call this the “Opaque” Web.


Resources

search engine history

The Invisible Web: Uncovering Information Sources Search Engines Can’t See
by Chris Sherman and Gary Price (2001)
ISBN 0-910965-51-X




No comments:

Post a Comment