Set robots.txt defaults to the following
-
User-agent:ia_archiver
Disallow: /User-agent: archive.org_bot
Disallow: /User-agent: *
Disallow: /User-agent: Rogerbot
User-agent: Exabot
User-agent: MJ12bot
User-agent: Dotbot
User-agent: Gigabot
User-agent: Baiduspider
User-agent: Ezooms
User-agent: Nutch
User-agent: archive.org_bot
User-agent: MJ12bot
User-agent: YandexBot
User-agent: AhrefsBot
User-agent: HTTrack
User-agent: Wget
User-agent: Zeus
User-agent: BLEXBot
User-agent: burroboot
User-agent: DOC
User-agent: MJ12Bot
User-agent: SemrushBot
User-agent: spbot
User-agent: UbiCrawler
User-agent: Zao
User-agent: Netsparker
User-agent: sitecheck.internetseer.com
User-agent: Zealbot
User-agent: MSIECrawler
User-agent: SiteSnagger
User-agent: WebStripper
User-agent: WebCopier
User-agent: Fetch
User-agent: Offline Explorer
User-agent: Teleport
User-agent: TeleportPro
User-agent: WebZIP
User-agent: linko
User-agent: Microsoft.URL.Control
User-agent: Xenu
User-agent: larbin
User-agent: libwww
User-agent: ZyBORG
User-agent: Download Ninja
User-agent: grub-client
User-agent: k2spider
User-agent: NPBot
User-agent: WebReaper
User-agent: CyotekWebCrawler
User-agent: Whizbang
User-agent: UniverseBot
User-agent: SlySearch
Disallow: / -
User-agent:ia_archiver
Disallow: /User-agent: archive.org_bot
Disallow: /User-agent: *
Disallow: /User-agent: Rogerbot
User-agent: Exabot
User-agent: MJ12bot
User-agent: Dotbot
User-agent: Gigabot
User-agent: Baiduspider
User-agent: Ezooms
User-agent: Nutch
User-agent: archive.org_bot
User-agent: MJ12bot
User-agent: YandexBot
User-agent: AhrefsBot
User-agent: HTTrack
User-agent: Wget
User-agent: Zeus
User-agent: BLEXBot
User-agent: burroboot
User-agent: DOC
User-agent: MJ12Bot
User-agent: SemrushBot
User-agent: spbot
User-agent: UbiCrawler
User-agent: Zao
User-agent: Netsparker
User-agent: sitecheck.internetseer.com
User-agent: Zealbot
User-agent: MSIECrawler
User-agent: SiteSnagger
User-agent: WebStripper
User-agent: WebCopier
User-agent: Fetch
User-agent: Offline Explorer
User-agent: Teleport
User-agent: TeleportPro
User-agent: WebZIP
User-agent: linko
User-agent: Microsoft.URL.Control
User-agent: Xenu
User-agent: larbin
User-agent: libwww
User-agent: ZyBORG
User-agent: Download Ninja
User-agent: grub-client
User-agent: k2spider
User-agent: NPBot
User-agent: WebReaper
User-agent: CyotekWebCrawler
User-agent: Whizbang
User-agent: UniverseBot
User-agent: SlySearch
Disallow: /Having this as "the default" would probably collide with some users using Cloudron to host public facing websites (myself included).
Besides:
@subtlecourage said in Set robots.txt defaults to the following:
User-agent: *
Disallow: /is already added to robots.txt of the app if you go into the "security" tab of a website and select "Disable indexing". What makes you think that bots that ignore
*(meaning all user agents) will not index your website if you name their user agent individually? -
Having this as "the default" would probably collide with some users using Cloudron to host public facing websites (myself included).
Besides:
@subtlecourage said in Set robots.txt defaults to the following:
User-agent: *
Disallow: /is already added to robots.txt of the app if you go into the "security" tab of a website and select "Disable indexing". What makes you think that bots that ignore
*(meaning all user agents) will not index your website if you name their user agent individually?@fbartels Optimism and hope? /s
Great points.
Hello! It looks like you're interested in this conversation, but you don't have an account yet.
Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.
With your input, this post could be even better 💗
Register Login