What is web spider ?

Web spider which is generally known as a crawler. This are the search engine bots or robots used to crawl the website to index site into its database.
Following are the well know crawlers.
1. Google - Google bot
2. Yahoo - Yahoo Slurp bot
3. MSN - MSN bot

Thanks
 
Googlebot is the search bot computer software used by Google, which builds up documents from the web to build some sort of searchable index for the The search engines engine.

A problem which web owners have often noted while using Googlebot is that it takes up a significant amount of bandwidth. This can cause internet sites to exceed their bandwidth limit and be taken down in the short term. This is especially troublesome for reflect sites which host a lot of gigabytes of data. Google supplies "Webmaster Tools" that let website owners to throttle typically the crawl rate.
 
A spider is a program that visits Web sites and reads their pages and other information in order to create entries for a search engine index. The major search engines on the Web all have such a program, which is also known as a "crawler" or a "bot." Spiders are typically programmed to visit sites that have been submitted by their owners as new or updated. Entire sites or specific pages can be selectively visited and indexed. Spiders are called spiders because they usually visit many sites in parallel at the same time, their "legs" spanning a large area of the "web." Spiders can crawl through a site's pages in several ways. One way is to follow all the hypertext links in each page until all the pages have been read.
 
A spider is a program that visits Web sites and reads their pages and other information in order to create entries for a search engine index. The major search engines on the Web all have such a program, which is also known as a "crawler" or a "bot." Spiders are typically programmed to visit sites that have been submitted by their owners as new or updated. Entire sites or specific pages can be selectively visited and indexed. Spiders are called spiders because they usually visit many sites in parallel at the same time, their "legs" spanning a large area of the "web." Spiders can crawl through a site's pages in several ways. One way is to follow all the hypertext links in each page until all the pages have been read.
 
Last edited:
A web crawler (also known as a web spider or web robot) is a program or automated script which browses the World Wide Web in a methodical, automated manner. This process is called Web crawling or spidering. Many legitimate sites, in particular search engines, use spidering as a means of providing up-to-date data.
 
Back
Top