Take control of how Large Language Models (LLMs) and AI bots interact with your website.
Configuration
Toggle which common LLM crawlers are allowed to access your site. The default is to allow.
Add any other user-agents you want to explicitly allow, one per line.
Add any other user-agents you want to explicitly block, one per line.
Provide a link to your XML sitemap to help AI models better understand your site structure.
Generated llm.txt
# Your llm.txt file will be generated here...
What is an llm.txt File?
An llm.txt
file is a proposed standard, similar to robots.txt
, that gives website owners control over how Large Language Models (LLMs) and AI crawlers access and use their content. This is one of the most common approached today to manage your site's data in the age of generative AI.
After all, llm.txt
file can help you specify which AI user-agents are allowed or disallowed from crawling your site for training purposes. Eventually, it can help you protect your intellectual property and manage server resources.
Why You Need to Control AI Access
- Protect Intellectual Property: Prevent your unique content from being used to train commercial AI models without your permission.
- Manage Server Load: Aggressive AI crawlers can put a strain on your server resources. And here a
llm.txt
file helps manage this traffic properly. - Ensure Data Privacy: Disallow bots from accessing sensitive (or user-generated) content sections of your site.
- Future-Proof Your Site: As AI becomes more integrated with search, having clear directives will be beneficial and even crucial for maintaining control over your website content.