llms.txt: Purpose, Limits, and When to Use It

Learn what llms.txt is, see a working example, understand which AI systems read it, and how it differs from robots.txt.

Author: Alex Sky5 min read
Conceptual image of two digital gatekeepers, robots txt and llms.txt, managing access to data

llms.txt is an emerging convention for helping large language models find the most useful, human-curated content on a site. It is not the same thing as robots.txt, and it is not a reliable legal or technical mechanism for blocking model training.

That distinction matters. Many teams treat llms.txt like a defensive access-control file, but in practice it is better understood as a discovery and guidance layer for AI systems that choose to read it. If you want to restrict crawler access, robots.txt, bot-specific policies, authentication, and server controls still matter more.

This guide explains what llms.txt is for, where it helps, and where teams often overestimate it.

Quick takeaways

  • llms.txt is best treated as an AI-readable guide to your most useful content.
  • It does not replace robots.txt, authentication, or server-side protections.
  • Use llms.txt to improve discoverability and context, not as your only defense policy.

What llms.txt Is

The proposed llms.txt format gives site owners a place to point AI systems toward important pages, docs, summaries, and structured resources. Think of it as a curated map for LLM-oriented browsing rather than a crawl-permission file.

That makes it useful for:

  • highlighting canonical docs or knowledge-base sections
  • surfacing the pages you want models to read first
  • providing concise summaries or navigation hints for AI agents

What llms.txt Is Not

llms.txt is not a guaranteed enforcement mechanism. If your goal is to block training crawlers or restrict sensitive paths, you still need to rely on bot-specific robots.txt rules, login walls, rate limits, and server-level controls.

It is also not a replacement for good technical SEO. If your site is hard to crawl, poorly structured, or missing clear canonicals, llms.txt alone will not fix those problems. Google also does not require any special AI-only text files or markup to appear in AI features.

robots.txt vs. llms.txt

Use the two files for different jobs:

FileMain purposeBest use
robots.txtCrawl policyAllow, limit, or block bots from specific paths
llms.txtAI-readable guidanceHighlight the most useful pages or sections for LLM consumption

If you want AI visibility, these files often work together: robots.txt controls access, while llms.txt helps with navigation and prioritization.

When to Publish llms.txt

Publishing llms.txt makes sense when you have:

  • product docs or help centers with clear canonical pages
  • long documentation trees where models may benefit from a curated index
  • editorial content that is already well structured and source-backed

It is less useful if your site still has weak internal structure, duplicate pages, or unclear ownership signals. Fixing those foundations usually comes first.

Example llms.txt File

Here is a working example for a SaaS documentation site:

# Acme Project Management
> Acme is a project management platform for remote teams.
> It offers task tracking, time logging, and team dashboards.
 
## Documentation
- [Getting Started](https://docs.acme.com/getting-started): Setup guide for new users
- [API Reference](https://docs.acme.com/api): REST API endpoints and authentication
- [Integrations](https://docs.acme.com/integrations): Slack, GitHub, and Jira connectors
 
## Guides
- [Workflow Automation](https://docs.acme.com/guides/automation): How to set up automated task rules
- [Team Permissions](https://docs.acme.com/guides/permissions): Role-based access control
 
## Optional
- [Changelog](https://docs.acme.com/changelog): Recent updates and version history
- [Pricing](https://acme.com/pricing): Current plans and pricing

Key formatting rules from the llms.txt specification:

  • The file is Markdown, not a config file
  • Start with # Site Name and a > description blockquote
  • Use ## sections to group links by type (docs, guides, optional)
  • Each link includes a short description after a colon
  • Place the file at your root URL: https://yoursite.com/llms.txt

Some sites also publish a companion llms-full.txt file containing the full documentation text in a single Markdown file. This gives AI crawlers a single high-signal ingestion point instead of requiring them to stitch together many separate pages.

For SaaS-specific patterns and more examples, see llms.txt Examples for SaaS.

Which AI Systems Read llms.txt

As of April 2026, adoption by major AI platforms is limited:

AI Systemllms.txt SupportNotes
ChatGPT / GPTBotNot confirmedOpenAI has not committed to reading llms.txt
PerplexityNot confirmedUses its own crawling and retrieval pipeline
Google Gemini / AIONot confirmedGoogle does not require AI-specific text files
Claude / ClaudeBotNot confirmedAnthropic has not announced support
Cursor / Windsurf / IDEsPartialSome AI coding tools read llms.txt for documentation context
Documentation platformsYesGitBook, Mintlify, ReadMe, and Fern generate llms.txt automatically

Only about 2% of the most-cited domains in AI search currently have an llms.txt file. The specification is still a proposal, not a ratified standard, and no major AI company has formally committed to honoring it.

This is why llms.txt should be treated as a low-cost, low-risk discovery layer — not as a primary strategy for AI visibility. The real visibility drivers remain content quality, crawl access via robots.txt, structured data, and domain authority.

How to Implement llms.txt

Keep the file lightweight and editorially intentional:

  1. Identify your highest-value pages. Prioritize docs, guides, and reference pages that contain clear, factual information. Skip marketing landing pages.
  2. Write the file in Markdown following the format above. Use canonical URLs, not campaign or parameterized URLs.
  3. Host at your root URL. The file should be accessible at https://yoursite.com/llms.txt.
  4. Keep descriptions concise. Each link description should be one sentence explaining what the page covers.
  5. Update when content changes. Review the file quarterly or when major docs or navigation change.
  6. Consider llms-full.txt. If your documentation is under ~50 pages, a full-text companion file gives AI systems everything in one request.

Limitations

The biggest limitation is adoption. llms.txt is still a convention, not a universal standard with guaranteed support. Some crawlers may ignore it entirely. Others may use it only as a discovery hint.

There is also no enforcement mechanism. Unlike robots.txt, which has decades of established crawler behavior, llms.txt has no legal or technical weight behind it. Publishing the file costs very little effort, but expecting it to meaningfully change AI citation outcomes would be overestimating its current reach.

References

VibeMarketing: AI Marketing Platform That Actually Understands Your Business

Stop guessing and start growing. Our AI-powered platform provides tools and insights to help you grow your business.

No credit card required • 2-minute setup • Free SEO audit included