Latest "Search Engines" Articles
Page# 1 (last added articles shown first)
How a search engine like Google assembles its index
(...) The result is every link on every page until all the links happen to be followed. This is the way new pages are added to the site's database, by using those links the spider hasn't seen before.
The pages discovered by the spider are copied verbatim to the search site's database and copied over each time they're updated. (...)
Google searches take place all over the Internet
(...) Keep in mind, Google also owns DoubleClick, the largest Internet ad serving technology. So even if you don't see a Google search box anywhere, the ads on that page could be served by Google. Part of the reason the Google habit is so hard to break is that you can't escape it. (...)
How Google keeps its users satisfied with the best search results
(...) So how can you identify those apertures? Oftentimes, they don't exist, so you have to create them.
Image recognition is an area Google has struggled with for some time, so its investment in Pixazza makes sense. Recognizing the limitations of its algorithm - and propensity for Web site publishers to stuff keywords into image ALT tags to fool it - Google created a "game" called Google Image Labeler. (...)
Search engine ad campaign tips and infos
(...) All of the major search engines want to deliver the best possible user experience to searchers. Each search engine wants to make the information that appears on its search engine results pages (SERPs) as useful, relevant, and informative as possible, no matter what the query. This is why the search engines are depending on users like you to help sustain their success and provide the best user experience possible. (...)
Discover the features of AltaVista and what it does best
(...) It digs down into the page and finds actual words and phrases from within the full-text of the website. As a result, if you can pin your search down to one word or a unique phrase, you can find it - quickly. It offers a broad range of search functionality and when it returns results, matching search term are highlighted - a time saver. (...)
Gigablast and Lycos are promising search engines
(...) It reads your results, then makes categorizes the common groupings it finds in its results - a sort of clustering.
Unlike all the other search engines, the default search setting on Gigablast was an OR operator. Most search tools use an AND operator. (...)
Take a glimpse at these great meta search tools
(...) You can also build results by grading the sites you see into useful or garbage. This allows you to customize and filter future results. What remains unclear is if they have built their own database, or are they buying it from one of the other search firms? Regardless, this is a useful new search tool. (...)
Why search engines do not index all web pages
Remember: it is costly to search, index, and store the results. So, many search engine crawlers limit the number of pages on a website they search, or limit the number of pages indexed, dumping older ones and replacing them with new ones, or restrict the kinds of pages they crawl by cataloging only certain types of domain names. Another problem is simple economics. (...)
Tips and hints on how AlltheWeb works and what features it has
(...) AlltheWeb has an outstanding news search feature, which allows you to search all the news categories, or pinpoint which news resources you want it to search. This feature will only get stronger as it is integrated with other Yahoo! properties.
AlltheWeb ties together its clustered results with the latest news stories on your query, and huge, frequently updated indexed database to pull from - these are strengths. (...)
Google search tips and why it is the most popular search engine
(...) What separates Google from the others is that it is simple to use and its top few results tend to be remarkably accurate - its deceptively simple basic interface give you little control, but great results.
Google was the first general search engine that provided access to pages at the time they were indexed, which they call "cached" pages. This feature allows you to find pages that are no longer available on the Internet - a very useful research tool, which other search engines like AlltheWeb have duplicated. (...)
Teoma and WiseNut strengths and weaknesses
Teoma's database remains considerably smaller than Google and Yahoo!, though it has grown to over 500 million documents. The newest version of the search engine adds an advanced search page, greater coverage of the Web, and the ability to set search preferences, among other options.
What Teoma does best
What Teoma does, it does very well. (...)
Short list of the best search tools online today
(...) MSN Search is already a worthwhile tool, but much more is expected from Microsoft when its new search engine is ready.
Netscape, another division of Time Warner, also uses Google for its main listings. About the only difference here is in appearance. (...)
Taking a closer look at three meta search engines
(...) Now you have a choice: you may view the results by relevance or by search engine, which makes it easier to assess your results. Above the query box are buttons for choosing to search web pages, images, audio, multimedia, news, and shopping. Dogpile has also added some phrase searching, Boolean operators, adult and domain filtering, and language selection, but only from the advanced search menu. (...)
What are search engine crawlers and what do they do
(...) This cuts down on the bogus URLs and helps the crawler be thorough. The search engine crawls and indexes information, but search engines do further refinement before the information is available to the public. The companies perform spam detection and removal, duplication detection and removal, and also do some database quality testing. (...)
Search engine algorithms and how websites rank
Nobody does it the same, which is why you get different results when you do a search, even the exact same search on two different search engines. Some search engines like Google put extra emphasis on what sites are linked to the primary site ("link analysis"), while others, like Teoma, rank searches for "hubs" and "authorities" related to your query terms that form a "community" about the topic.
Search engines have found some creative ways to deal with bad queries. (...)
Natural language search engines and when to use them
(...) The pages are isolated. Then, the engine summarizes and analyzes potentially relevant documents, finding commonality among relevant documents.
Inktomi and many other search tools like Oingo, Lexiquest, Iphrase and Manning and Napier Information System's Map-It offer phrase detection and more advanced natural language products including concept mapping to find what you mean, not what you say. (...)
Comparison between search engines and subject directories
As a rule, the more specific you are, the less likely you will be to find what you are looking for in a subject directory. Some subject directories do not have search capabilities. In those cases, your only choice is to browse through the categories and subcategories. (...)
What are meta search tools and what purpose do they have
Most meta-search tools can conduct only initial, simple searches, meaning you cannot refine or limit your search. This is the big problem with meta-search tools: to get wide access to many search tools, you give up several things, including the ability to do any kind of sophisticated or advanced searching. Ultimately, the results from meta-search tools are less precise. (...)
Search engine positioning and the so called sponsored links
To the credit of the major meta-search tools, when I conducted a similar search in March 2004, on "breast cancer" using Dogpile, thirty-one of the first sixty results were "sponsored" - in other words, paid for - but all of those results were clearly marked as sponsored results. A similar search on Overture, the paid-placement company that Yahoo! owns, all of the top forty results were sponsored but also clearly marked.
Many of the meta-search engines still do not clearly mark the results. (...)
How does the indexer of a search engine actually work
(...) As an example, when Google announced in February, 2004, that it had increased its total number of pages to 4.28 billion, it did not mention that a portion of those results were un-indexed pages. Yes, you still had access to billions of Google's pages, just not all 4. (...)
Interesting search engine facts every webmaster should know about
(...) Frequent updates cost a lot of money, so more and more, how good a search tool is can depend on how much money its owners are putting into it, and how good its supporting partners are. Most of the major search engines claim to index 500 million documents. AlltheWeb indexes 3. (...)
Special portals of search engines are known as vortals
(...) It is the starting place for information about that industry and for everyone who works in that industry. A legal vortal is FindLaw a central clearinghouse for legal information. Energy Central is a key site for nuclear power and electric energy issues. (...)
Paid placement is just one of the ways a search engine makes money
(...) Now paid placement in search tools is a common practice, although it is one you may not know about. It is so widespread that, at a mid-2001 conference run by Search Engine Watch editor Danny Sullivan, the vast majority of the companies presenting their technologies were firms specializing in getting listings put into search engines, a field they dub "search engine optimization." Those people who viewed placement as a violation of the spirit of search engines were a tiny minority amid a wave of placement proponents. (...)
Enter page# 1 (last added articles shown first)