PDA

View Full Version : Why is a Robots.txt File Used?



Arunsingh
06-12-2013, 09:37 AM
Why should I have a robots.txt file?

Thanks...

Rajdeep Bose
06-13-2013, 12:49 AM
This section has a few handy examples.

To prevent FreeFind from indexing your site at all:

user-agent: FreeFind
disallow: /

To prevent FreeFind from indexing common Front Page image map junk:

user-agent: FreeFind
disallow: /_vti_bin/shtml.exe/

To prevent FreeFind from indexing a test directory and a private file:

user-agent: FreeFind
disallow: /test/
disallow: private.html

blessy_smith
06-13-2013, 02:46 AM
Robots.txt are also helpful in blocking the bots from indexing directories that contain scripts. If you have a very plain site that you would be ok with having the engines crawl everything than you do not need a robots.txt.

brandappz
06-13-2013, 06:17 AM
Robots.txt is a simple text (not html) file you put on your website root directory to tell search robots which pages you would like them not to visit. By defining a few rules in this text file, you can instruct robots to not crawl certain files, directories within your site, or at all.

madtteam28
06-13-2013, 07:09 AM
Robots.txt is a text file that consist of half of the URLs of some pages that you don't want search engines to be crawled.

Johnpaul
06-13-2013, 07:36 AM
If you dont want to crawl the thing in search engine, just create robot.txt and add them in it. Search engine will nevrer crawl them.

Mariajoseph
06-13-2013, 09:19 AM
To tell which search engine and which pages to crawl and index.....

godwin
05-14-2020, 08:05 AM
The robots exclusion standard, also known as the robots exclusion protocol or simply robots.txt, is a standard used by websites to communicate with web crawlers and other web robots. The standard specifies how to inform the web robot about which areas of the website should not be processed or scanned.

RH-Calvin
05-14-2020, 01:42 PM
Robots.txt is a text file that lists webpages which contain instructions for search engines robots. The file lists webpages that are allowed and disallowed from search engine crawling.

dombowkett
05-14-2020, 08:07 PM
Robots.txt is used to check which portion of website we have to index OR not.

Maria Jonas
05-15-2020, 07:24 AM
Why should I have a robots.txt file?

Thanks...

Have you Googled this? If not, do it now!

GeethaN
05-16-2020, 05:45 AM
A robots. txt file tells search engine crawlers which pages or files the crawler can or can't request from your site. This is used mainly to avoid overloading your site with requests; it is not a mechanism for keeping a web page out of Google.