# robots.txt for http://www.Hisways.org/ # NO AI bots 2023.12.05: https://neil-clarke.com/block-the-bots-that-feed-ai-models-by-scraping-your-website/ User-agent: CCBot Disallow: / User-agent: ChatGPT-User Disallow: / User-agent: GPTBot Disallow: / User-agent: Google-Extended Disallow: / User-agent: anthropic-ai Disallow: / User-agent: Omgilibot Disallow: / User-agent: Omgili Disallow: / User-agent: FacebookBot Disallow: / User-agent: SBIder Disallow: / User-agent: * psbot Disallow: /ads.txt Disallow: /cig-bin/ Disallow: /scripts/ Disallow: /wp-admin/ Disallow: /wp-content/ Disallow: /wp-json/ Disallow: /ws_ftp.log Disallow: /xhis-bkm.htm Disallow: /zhis-bkm.htm Disallow: /zpages/ Disallow: /zseminar/ #Disallow: /images/ #Disallow: /SignUp/ *Disallow: /SignUp-male.txt # 2/13/2019 added "/ads.txt b/c google searching causes 404 error every 24 hours. # 2/4/11 `SBIder/Nutch-1.0-dev (http://www.sitesell.com/sbider.html)` spider agent was purposly # causing 404 errors so I would look them up and read their self-advert. # Must have 2 carrage returns to properly terminate file otherwise robot can # ignore "robots.txt" file. # # Everything you wanted to know about robots.txt http://www.robotstxt.org/wc/norobots.html or # http://www.robotstxt.org/orig.html ## "Any number of agent id(s) can be placed on the User-Agent line so long as ## they are separated by white space (WS), but the User-Agent line must have at ## least one agent id." or see: http://www.seoconsultants.com/robots-text-file/ ## Added ``psbot`` cause it was messing up Urchin v3.0 stats. 1/20/04. # # Disallow: /stats/ removed 2/19/02 I think this caused password pbms. # Disallow: /scripts/ 2/20/02 ; re-installed 3/4/02 # Disallow: /1webdir.htm removed all these 3/1/01 # Disallow: /zseminar/ removed this 6/21/01 having pbms w/slide-show. ; re-installed 3/02