feat: add various robots.txt and YAML configurations for user agent handling and crawl delays

This commit is contained in:
Jason Cameron
2025-06-13 13:31:47 -04:00
parent 5a85cd41a2
commit 714496e317
14 changed files with 231 additions and 356 deletions

View File

@@ -0,0 +1,36 @@
- name: robots-txt-policy-crawl-delay-1
action: WEIGH
expression:
single: "true"
weight:
adjust: 3
- name: robots-txt-policy-disallow-2
action: CHALLENGE
expression:
single: path.startsWith("/admin")
- name: robots-txt-policy-blacklist-3
action: DENY
expression:
single: userAgent.contains("BadBot")
- name: robots-txt-policy-crawl-delay-4
action: WEIGH
expression:
single: userAgent.contains("SpamBot")
weight:
adjust: 3
- name: robots-txt-policy-blacklist-5
action: DENY
expression:
single: userAgent.contains("SpamBot")
- name: robots-txt-policy-crawl-delay-6
action: WEIGH
expression:
single: userAgent.contains("Googlebot")
weight:
adjust: 3
- name: robots-txt-policy-disallow-7
action: CHALLENGE
expression:
all:
- userAgent.contains("Googlebot")
- path.startsWith("/search")