This repository is a collection of resources that might help with avoiding gen AI. It was written by coders for coders, as an attempt to help others as an unpaid volunteer effort. You'll likely find mistakes, please fix them.
The primary purpose is to provide information, fast and reliably. We link to other projects and try to always provide the data sources. Even though this repository is primarily meant for developers, it may still provide useful information for "casual" people.
- Software
- Avoid AI Citations
- Avoid AI Crawling
- UBlock Filters
- Poison AI Scrapers
- Clear AI Guidelines
- Footnote
This is an incomplete list of software to use when trying to avoid AI.
The practical choices with good driver support tend to be Windows, the Apple ecosystem, or Linux. If you want to avoid gen AI, avoid Windows. Apple with their macOS seem to also be quite pro-AI and seem to be restrictive and anti free open software too.
In general, Linux is your best choice for avoiding AI and to have user freedom. (Note: We don't recommend ChromeOS as it seems to be folding into Android and Android seems anti user freedom too).
We don't want to recommend specific Linux variants here, but the following links may help you:
This section has strategies to avoid AI chat bots from interacting with your website or your code project.
-
Adding a CLAUDE.md with the token to stop Claude access. This token apparently stops Claude Code from interacting with your code project if put into a
CLAUDE.mdfile in your code tree. It may also stop it from interacting with your website if you put the token into the web page code, e.g. as a HTML comment. (Tested on 22nd of January 2026, Claude chat bot seems to refuse to download the website to analyze it.) -
Adding a AGENTS.md. This could stop AI agents from interacting with the repository, by giving it names it can legally not respond to.
-
Adding a copilot-instructions.md. This gives copilot instructions for interacting with the repository. This could force it to stop interactions with it completely.
-
For websites only:
<meta name="robots" content="nosnippet" />Source: https://developers.google.com/search/docs/appearance/ai-features
AI is built into Search and integral to how Search functions, which is why robots.txt directives for Googlebot is the control for site owners to manage access to how their sites are crawled for Search. To limit the information shown from your pages in Search, use nosnippet,
Currently hard to do, but here are some known options:
- Anubis. Sadly, using this will likely unlist
all protected pages from web searches, too. But the landing page could be
excluded, for example.
- Anubis can be configured to allow well-behaved bots and legitimate automation scenarios if you so desire
- Blocking crawlers using robots.txt. It's important to understand that AI companies are not forced to follow this file, they might, but could also ignore it. Most big companies will respect it but you never know.
- Try to avoid code hosters that run on one of the big cloud providers that are known for gathering data for AI training, or that integrate gen AI directly into the UI. As of today, Codeberg.org appears to be fairly safe, while Github.com appears to integrate Co-Pilot AI deeply.
These are some filters for blocking AI buttons / content. When using some of these filters you may experience a degraded experience because certain other buttons might not show up. The easiest solution most of the times is reloading the page:
This section has ideas on how to poison the data of AI scrapers once they snuck past blocking mechanisms, to disincentivize scraping.
-
Glaze is supposedly able to protect an artists artstyle from AI scraping, by slightly altering the image. They claim that these changes are not visible to the human eye, but can project the artstyle from being scraped.
-
Nightshade is supposedly able to poison AI data sets, by altering the image slightly, not noticeable to the human eye.
Sources:
- Iocaine. A lightweight garbage generator that aims to keep AI crawlers away from heavy operations, and instead traps them in a maze made to poison the dataset.
- The default handler does a lot, but the author also provides their own handler in the form of Nam-Shub of Enki.
Although it's unlikely that the AI enthusiast are going to follow a guideline / coc, it's still a viable option to drive away the few people who respect rules set up by a maintainer. Please note these are just examples, use any at your own risk.
We have some examples for how such a Code Of Conduct might look like:
There are also snippets that might inspire an anti AI README.md note:
This work is licensed under CC BY-NC-SA 4.0
The licence of this project was chosen to prevent commercial usage, which we hope will possibly provide plenty of headaches for AI scrapers. This statement isn't legal advice, and we have no idea if it's true or not, but it still makes us happy. 😊