What Are Search Engines?
Most of us often face the problem of searching the web. Nowadays, the global network is one of the most important sources of information there is, its main goal being to make information easily accessible. That's where the main problem arises: how to find what you need among all those innumerable terabytes of data. The World Wide Web is overloaded with various stuff related to diverse interests and activities of human beings who inhabit the globe. How can you tell what a site is devoted to without visiting it? Besides, the number of resources grew as quickly as the Internet’s own development, and many of them closely resembled each other (and still do). This situation necessitated finding a reliable (and at the same time fast) way to simplify the search process, otherwise there would be absolutely no point to the World Wide Web. So, development and deployment of the first search engines closely followed the birth of the World Wide Web.
How It All Began
At the start, search engines developed quite rapidly. The "grandfather" of all modern search engines was Archie, launched in 1990, the creation of Alan Emtage, a student at McGill University, Montreal. Three years later, the University of Nevada System Computing Services deployed Veronica. These search engines created databases and collected information on the files existing in the global network. But they were soon overwhelmed by the fast growth of the net, and others stepped forward.
World Wide Web Wanderer was the first automated Internet robot, whereas ALIWEB, launched in Autumn of 1993, was the first rough model of a modern web directory that is filled up by site owners or editors. At about the same time, the first 'spiders' appeared. These were: JumpStation, World Wide Web Worm, and Repository-Based Software Engineering starting the new era of World Wide Web search. Google and Yahoo are two of their better-known descendants.
http://galaxy.com/info/history2.html.
Search Engines Today
Modern web searchers are divided into two main groups:
• search engines and
• directories.
Search engines automatically 'crawl' web pages (by following hyperlinks) and store copies of them in an index, so that they can generate a list of resources according to users' requests (see ‘How Search Engines Work’, below). Directories are compiled by site owners or directory editors (in other words, humans) according to categories. In truth, most modern web search combine the two systems to produce their results.
How Search Engines Work
All search engines consist of three main parts:
• the spider (or worm);
• the index; and
• the search algorithm.
The first of these, the spider (or worm), continuously ‘crawls’ web space, following links that lead both to within the limits of a website and to completely different websites. A spider ‘reads’ all pages’ content and passes the data to the index.
The Index is the second part of a search engine. It is a storage area for spidered web pages and can be of a huge magnitude (Google’s index, for example is said to consist of three billion pages).
The third part of a search engine system is the most sophisticated. It is the search algorithm, a very complicated mechanism that sorts an immense database within a few seconds and produces the results list. Looking like a web page (or, most often, lots of pages), it contains links to resources that match users' queries (i.e., relevant resources). The most relevant ones (as the search engine sees it) are nearer the top of the list. They are the ones most likely to be clicked by the user of the search engine. A site owner should therefore take heed of the site's relevancy to the keywords it is expected will be used to find it.
http://www.searchenginewatch.com/webmasters/article.php/2168031
A Relevancy calculation algorithm is unique for every search engine, and is a trade secret, kept hidden from the public. However, there are some common principles, which will be discussed in the following paragraph.
http://www.searchenginewatch.com/webmasters/article.php/2167961
What to Do to Have Your Web Site Found through Search Engines
There are some simple rules to make your resource relevant enough to be ranked in the top 10 by the majority of search engines.
Rule 1: Work on the body copy
A search engine determines the topic of your site judging by the textual information (or content) of every page. Of course, it cannot comprehend the content the way humans do, but this is not critical. It is much more important to include keywords, which are found and compared with users' queries by the programme. The more often you use targeted keywords, the better your page will be ranked when a search on those keywords is made.
You can increase the relevancy of your targeted keywords still more if you include them in the HTML title of your page (
tag), in subheaders ( -