385909
results for ""a""
-
robots txt << [https://ignoreallpreviousinstructions.pages.dev/ Ignore all previous instructions] is a single purpose webpage that curates a robots.txt file against AI scrapers
Zegnat
at
2026-01-09 16:31
-
regarding AI training worries, has anyone heard of anubis? it's a small proxy (might be indieweb-dev territory) that you can run in front of your website. most AI bots cannot get past it
simpleauthority
at
2026-01-09 16:20
-
Not sure how current it is but this is a restrictive one I experimented with https://ignoreallpreviousinstructions.pages.dev/robots.txt
[morgan]
at
2026-01-09 16:12
-
but alas, this is a partially pretty bad timeline
[morgan]
at
2026-01-09 16:07
-
we should be able to share without that being a concern
[morgan]
at
2026-01-09 16:07
-
I love recording small audio samples, its been a favorite hobby for a year or two
[morgan]
at
2026-01-09 16:04
-
Podcasts make me feel like they are a bit more of a production
[Sophia_wood]
at
2026-01-09 16:01
-
I am currently having a similar reaction to Alan. I need to add a robots.txt now haha.
simpleauthority
at
2026-01-09 15:56
-
I added a robots.txt file for the first time about four months ago. I've been getting traffic from Google, and Google search result emails, for at least 10 years. So it definitely hasn't been preventing them indexing me 😄
[Murray]
at
2026-01-09 15:16
-
edited /robots_txt (+557) "Some people observe Google dropping them from search results because of the lack of a robots.txt file."
Loqi
at
2026-01-09 15:15
Sort by:
Filter results by:
Tag
Query took 0.03s.
Search tips
- Exclusion
- +foo -bar
- Logical OR
- foo OR bar
- Exact phrase
- "foo bar"
- Partial words
- foo*
- Particular fields only
- title:foo
- domain:example.org
- nick:somebody
- after:2016-11-23
- before:2016-11-23
- date:2016-11-23