Call Anytime 24/7
Mail Us For Support
Office Address
Lahore, Punjab, Pakistan

As AI models like GPT, Gemini, and Claude increasingly crawl websites for training data, site owners now have a way to control how their content interacts with these systems—through a file called llms.txt. Just like robots.txt helps you manage search engine crawlers, llms.txt serves as your site’s policy file for Large Language Models (LLMs).
Adding an llms.txt file to your website gives you control over whether AI systems can read, use, or ignore your content. This step has become essential for privacy, data management, and brand safety in the age of generative AI.
llms.txt (short for Large Language Model System text) is a plain text file placed in your site’s root directory (e.g., https://yourdomain.com/llms.txt). It communicates your preferences to AI crawlers, indicating whether they can use your site’s data for training or indexing.
For example:
User-agent: GPTBot
Disallow: /
This means OpenAI’s GPT crawler is not allowed to scrape your site.
This file ensures:
| Feature | robots.txt | llms.txt |
|---|---|---|
| Purpose | Controls traditional web crawlers (e.g., Googlebot) | Controls AI/LLM crawlers (e.g., GPTBot) |
| Placement | Root directory (/robots.txt) | Root directory (/llms.txt) |
| Impact | Affects indexing & SEO | Affects AI data training & content use |
A standard llms.txt follows a similar format to robots.txt:
User-agent: GPTBot
Disallow: /
User-agent: ClaudeBot
Allow: /
Each section specifies an AI crawler and your permission for it.
Before proceeding, ensure you have:
Open a text editor and create a simple file named llms.txt. Add your preferred crawler rules, such as:
User-agent: GPTBot
Disallow: /
/public_html/ or your site’s root folder.llms.txt directly into that directory.Go to https://yourdomain.com/llms.txt. If it loads properly, the file has been added successfully.
Use tools like https://llmstxt.org/tester to verify that your rules are correctly formatted and accessible.
Login to your hosting account and open File Manager.
llms.txt here.Ensure file permissions are set to 644 (readable by everyone).
Access https://yourdomain.com/llms.txt in your browser.
If you can see the file, your configuration is correct.
You can use tools like:
curl https://yourdomain.com/llms.txt)Regularly reviewing your file ensures you maintain the right data-sharing balance.
Ensure the file is in the root directory, not inside a subfolder like /wp-content/.
File permissions should be 644. Anything stricter (like 600) may block crawlers from reading the file.
1. Do I need both robots.txt and llms.txt?
Yes. robots.txt controls search engines, while llms.txt controls AI crawlers.
2. Can I block all AI bots?
Yes, use:
User-agent: *
Disallow: /
3. Where should llms.txt be placed?
Always in the root directory (e.g., /public_html/).
4. Will adding llms.txt affect SEO?
No, it only affects AI crawlers—not search engines.
5. How often should I update llms.txt?
Every few months or when new AI bots emerge.
6. What happens if I don’t add llms.txt?
AI crawlers may interpret it as permission to use your content for training.
Adding llms.txt to WordPress or cPanel is a simple but powerful way to manage how AI interacts with your website. By defining clear policies for AI crawlers, you protect your content, maintain data control, and ensure ethical AI engagement.
For further reading, check out https://llmstxt.org — a reliable resource for managing AI crawler settings.
Comments are closed