Subscribe Us

Header Ads

Closure of Robot .Txt Files




    Basic Definition-
    • From 1 September 2019 ,Google announced that Google Bot will no longer obey  a robot.txt directive related to indexing.
    Google Official Tweet
    • The reason  behind the no-index robots.txt directive won’t be supported is because it’s not an official directive.
    • The no-index will stop the page showing in search results, and the disallow will stop it being crawled.
    • Read my previous blog post in order to in depth knowledge about robot.txt files.
    • Below syntax will be no longer used in google.
    • That’s no longer the case. The no-index robots.txt directive is no longer supported.
    Syntax

    user-agent:*
    Disallow/AdminPanel/


      How to control crawling:-

      • No-index in robots meta tags is the most effective way to remove URLs from the index when crawling is allowed.
      <meta name="googlebot" content="noindex">
      • 404 and 410 HTTP status codes which are related to "Page not found".
      • Password protection.
      • Disallow in robot.txt.
      • Search console remove URL tool. 



        Official Google Announcement:-



        Official Google Tweet Announcement:-


        Post a Comment

        0 Comments