Page 1 of 2 12 LastLast
Results 1 to 15 of 28
  1. #1

    What is Robots.txt?

    Hello Friends,

    Please tell me what is robots.txt.

  2. #2
    Senior Member
    Join Date
    Jul 2017
    Posts
    227
    Web site owners use the /robots.txt file to give instructions about their site to web robots; this is called The Robots Exclusion Protocol. The "User-agent: *" means this section applies to all robots. The "Disallow: /" tells the robot that it should not visit any pages on the site.

  3. #3
    Registered User
    Join Date
    Feb 2017
    Posts
    754
    Robots Txt is an HTML attribute that is used to inform the search engines not to crawl and index the web pages in the website.

  4. #4
    Registered User
    Join Date
    Mar 2016
    Posts
    294
    Robots.txt is a file to give instructions to web robots about the website crawling; this is called The Robots Exclusion Protocol. The "User-agent: *" means this section applies to all robots. The "Disallow: /" tells the robot that it should not visit any pages on the site.

  5. #5
    Registered User
    Join Date
    Dec 2015
    Location
    USA
    Posts
    1,060
    It is Robot txt and this means that there are certain places on the website where there is personal information and in these places customers are not allowed to go so in this case robot txt is used

  6. #6
    Registered User 24x7servermanag's Avatar
    Join Date
    Jul 2017
    Location
    India
    Posts
    1,020
    Robot.txt is used to crawl the pages website. It tells that which part of the area should not be accessed. We can define the pages which should not be accessed by putting the disallow tag in robot.txt. Those disallow pages are restricted to visit. It also help to index the web content.
    Server Management Company
    India's Leading Managed Service Provider | Skype: techs24x7
    Cpanel Technical Discussions - Lets talk !

  7. #7
    Member
    Join Date
    Jul 2017
    Location
    India
    Posts
    72
    robots.txt file is used to give instruction to the bots that crawl the website.

  8. #8
    Senior Member
    Join Date
    May 2017
    Posts
    254
    The robots exclusion standard, also known as the robots exclusion protocol or simply robots.txt, is a standard used by websites to communicate with web crawlers and other web robots. The standard specifies how to inform the web robot about which areas of the website should not be processed or scanned.

  9. #9
    Junior Member
    Join Date
    Jul 2017
    Posts
    26
    robot.txt is a file which let you communicate between the web crawler and website...in the way of telling which web pages to be crawled and which not to be

  10. #10
    Registered User
    Join Date
    Apr 2015
    Location
    New Delhi
    Posts
    679
    Thank For Sharing Valuable Information...

  11. #11
    Registered User
    Join Date
    Jan 2013
    Location
    India
    Posts
    54
    Robots.txt is a test file uploaded at the root of any website. It is used to allow or block URLs of the website to be crawled by different search engine. We need to follow a set protocol prescribed by robots.txt community

  12. #12
    Registered User
    Join Date
    Aug 2010
    Posts
    73
    A txt file that gives instructions to web crawlers what to do.
    HostOX - Quality web hosting with super fast UK support.
    oXnames - Simple Domain registration with no BS

  13. #13

  14. #14
    Registered User
    Join Date
    Aug 2018
    Location
    Atlanta, Georgia
    Posts
    62
    Robot.txt is utilized to crawl the webpage internet site. It informs. We can specify the pages that must not be obtained by putting the disallow label in robot.txt. Those disallow pages have been limited to see. Additionally, it help index the internet content.

  15. #15
    Registered User
    Join Date
    Nov 2018
    Posts
    38
    It’s a file that instructs search engines how to crawl a website.

    It’s not necessary for all sites - search engines will still crawl your site without it
    Using it to block a search engine, or all search engines, is only an instruction - it can be easily ignored. Don’t use it to hide sensitive data
    You can use it to tell search engines where your XML sitemap is located (or sitemaps, if you have more than one)
    You can use it to prevent search engines crawling particular files or entire folders. You can specify to allow some, but not all search engines.
    It doesn’t remove a page from Google’s search index if it’s already in there - though the page will no longer be crawled. It will show in the search results with: ‘This page was blocked with robots.txt” or similar line of text.
    It can also be used to set a crawl delay to prevent some search engines from tying up your server resources by crawling your site too aggressively or using up bandwidth.

Page 1 of 2 12 LastLast

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •  

  Find Web Hosting      
  Shared Web Hosting UNIX & Linux Web Hosting Windows Web Hosting Adult Web Hosting
  ASP ASP.NET Web Hosting Reseller Web Hosting VPS Web Hosting Managed Web Hosting
  Cloud Web Hosting Dedicated Server E-commerce Web Hosting Cheap Web Hosting


Premium Partners:


Visit forums.thewebhostbiz.com: to discuss the web hosting business, buy and sell websites and domain names, and discuss current web hosting tools and software.