The llms.txt File: Complete Guide for 2026

GuidesBlake Folgado
The llms.txt File: Complete Guide for 2026

llms.txt is a plain text file you place at the root of your website (yourdomain.com/llms.txt) that gives AI systems a clean, structured map of your most important content — similar to how robots.txt tells search engines where to crawl. Instead of making ChatGPT, Claude, or Perplexity sift through your cluttered HTML, navigation menus, and ad scripts, llms.txt hands them a curated list of the pages that actually matter. In 2026, that matters because a growing share of your traffic now comes from AI answers instead of blue links — and the sites that make life easy for AI get cited more often.

This guide covers what llms.txt is, where it came from, what to put in it, and exactly how to create and upload one in under ten minutes. llms.txt is one of the core signals covered in our generative engine optimization guide, alongside schema, citations, and content structure.

What Is llms.txt?

llms.txt is a markdown file that lives at the root of your domain — so if your site is acme.com, the file sits at acme.com/llms.txt. It is written in plain text with a simple structure: a site name, a short description, and a list of links to your most important pages grouped into sections.

The goal is simple. AI models have limited context windows (the amount of text they can read at once). When an AI system visits your website to answer a question, it cannot — and should not — read every page. llms.txt gives it a pre-selected, human-curated list of the content that best represents your business, so the model can quickly find what it needs and cite you accurately.

Think of it as a restaurant menu for AI. Your full website is the kitchen, with everything you have ever cooked. llms.txt is the menu — the curated selection you want visitors to actually order from.

Where Did llms.txt Come From?

The llms.txt standard was proposed in September 2024 by Jeremy Howard, co-founder of fast.ai and Answer.AI, and one of the most respected figures in applied AI research. Howard noticed that AI systems were increasingly being used to read websites on behalf of users, but most websites were not designed for machine consumption.

The full specification lives at llmstxt.org, which is the official home of the standard. The proposal borrows the naming convention from robots.txt — the file websites have used since 1994 to tell search engine crawlers which pages they can access — but solves a different problem. robots.txt controls access; llms.txt controls comprehension.

According to llmstxt.org, the file is designed to "provide information to help LLMs use a website at inference time" — meaning at the exact moment an AI is trying to answer a user's question about your content, not months earlier during training.

Since the proposal, adoption has grown quickly. According to Writesonic's llms.txt adoption tracker, thousands of sites including Anthropic, Stripe, Zapier, Cloudflare, and Perplexity have published llms.txt files — well before any AI platform officially required them.

llms.txt vs robots.txt vs sitemap.xml

These three files are often confused. They all sit at the root of your domain and they all control how machines interact with your site. But they do very different jobs.

llms.txtrobots.txtsitemap.xml
PurposeCurate content for AI to read and citeBlock or allow crawlers from accessing pagesList every URL for search engines to index
AudienceLLMs (ChatGPT, Claude, Perplexity, Gemini)Search engine bots, scrapersSearch engine bots
Required?No — proposed standard, voluntaryNo — but universally supported since 1994No — but standard SEO practice
FormatMarkdown with links and short descriptionsPlain text rules (User-agent, Disallow)XML with <url> entries
Location/llms.txt/robots.txt/sitemap.xml
Typical length20–200 curated links5–30 lines of rulesEvery URL on the site (can be thousands)
GoalMake content easy to understandControl what gets crawledHelp search engines discover everything

The simplest way to remember it: robots.txt is a bouncer, sitemap.xml is a phonebook, and llms.txt is a menu.

Why llms.txt Matters for AI Visibility in 2026

The shift from search engines to AI answer engines is no longer a prediction — it is the default for a growing share of users. According to Semrush, ChatGPT referral traffic to websites grew by more than 400% between 2024 and 2026, and a meaningful portion of queries that used to go to Google now end inside an AI chat window.

When AI systems cannot parse your site cleanly, three things happen:

  1. You get misquoted. The model grabs outdated or irrelevant content and puts the wrong words in your mouth.
  2. You get skipped. If another site's content is easier to read, the AI cites them instead.
  3. You lose the traffic. AI answers with citations drive clicks; answers without them do not.

According to SearchEngineLand, well-structured content that is easy for LLMs to parse is significantly more likely to appear in AI citations — and llms.txt is one of the fastest, cheapest signals you can add.

Unlike SEO tactics that take months to show results, llms.txt is a one-file change. You can publish one in an afternoon. For more on the broader strategy, see GEO vs SEO: what's actually different and how to get cited by ChatGPT.

What Goes in an llms.txt File

The format is deliberately minimal. A valid llms.txt file has:

  1. An H1 heading with your site or business name — one line, no fancy formatting.
  2. A blockquote description — a 1–3 sentence summary of what your business does and who it serves.
  3. Optional context paragraphs — short explanations of anything an AI should know before reading your links.
  4. H2 sections with grouped links — each link is a markdown list item with the URL and a short description.
  5. Optional ## Optional section — secondary links you want the AI to know about but that are not critical.

Here is what a complete, real-world example looks like:

markdown
# ACME Coffee Roasters

> ACME Coffee Roasters is a speciality coffee company based in Portland, Oregon. We source single-origin beans directly from farms in Ethiopia, Colombia, and Guatemala and ship freshly roasted coffee to customers in the United States and Canada.

We have been roasting since 2012. Our team includes three Q-graders and a head roaster with 15 years of experience. We publish tasting notes, brewing guides, and farm profiles on our blog.

## Core pages

- [Homepage](https://acme-coffee.com): Main landing page with current featured beans
- [About us](https://acme-coffee.com/about): Our story, team, and sourcing philosophy
- [Shop all coffee](https://acme-coffee.com/shop): Complete catalogue of single-origin and blend coffees
- [Subscriptions](https://acme-coffee.com/subscribe): Monthly and bi-weekly delivery options
- [Wholesale](https://acme-coffee.com/wholesale): Coffee for cafes, restaurants, and offices

## Brewing guides

- [Pour over guide](https://acme-coffee.com/guides/pour-over): Step-by-step V60 method
- [French press guide](https://acme-coffee.com/guides/french-press): Ratios, grind size, and timing
- [Espresso at home](https://acme-coffee.com/guides/espresso): Equipment, dosing, and dialling in

Give your AI superpowers — connect once and access every tool.

Get started for free

Origin profiles

Policies

Optional

  • Blog: News, events, and industry commentary
  • Events: Cuppings and tasting workshops in Portland

Notice what is included and what is left out. No marketing fluff. No "welcome" pages. No navigation elements. Just the pages an AI would actually need to answer questions about the business.

## How to Create Your llms.txt in 10 Minutes

The fastest way to build one from scratch:

1. **Open a plain text editor.** Notepad, TextEdit (in plain text mode), or any code editor works. Do not use Word — it adds invisible formatting.
2. **Start with your business name as an H1.** Type `#` followed by your site or business name.
3. **Write a 1–3 sentence description inside a blockquote.** Start the line with `>` and describe what you do, who you serve, and where you operate.
4. **List your 5–15 most important pages as markdown links.** For each link, add a colon and a short description of what is on the page.
5. **Group related links under H2 headings.** Use categories like "Products", "Guides", "Pricing", "About", "Policies".
6. **Put anything optional or secondary under a final `## Optional` heading.**
7. **Save the file as `llms.txt`** — all lowercase, no extension other than `.txt` (make sure your editor does not add `.rtf` or similar).
8. **Review it.** Read it out loud as if you were a stranger. Does it explain what your business is? Does it point to the right pages?

If writing this by hand sounds tedious, you can generate a first draft using the [GEO tool](/tools/geo) — it will audit your site and suggest which pages to include.

## Where to Put the File

llms.txt must live at the **root of your domain**, accessible at `https://yourdomain.com/llms.txt`. If it is at `yourdomain.com/files/llms.txt` or `blog.yourdomain.com/llms.txt`, it will not be found.

Here is how to upload it on the most common platforms:

- **WordPress:** Install a plugin like "WP File Manager" (or use FTP). Upload `llms.txt` to the root directory — the same folder that contains `wp-config.php` and `index.php`. Some SEO plugins like Yoast and Rank Math are adding native llms.txt support — check your SEO plugin settings first.
- **Webflow:** Go to **Project Settings → SEO → Indexing** and look for the llms.txt field. If your plan does not expose it, add the file through a custom code injection or publish it as a static asset.
- **Vercel / Next.js:** Drop the file into your `public/` folder as `public/llms.txt`. It will be served at the root automatically on next deploy.
- **Netlify:** Drop the file into your `public/` or `static/` folder (depending on your framework). It deploys with your next build.
- **Shopify:** Shopify does not allow direct root file uploads, but you can create a page with the llms.txt content and use a custom template. Some SEO apps in the Shopify store now add llms.txt support.
- **Squarespace:** Go to **Settings → Advanced → Code Injection** — llms.txt support is being rolled into the platform, but for now, use a developer account to upload the file.

After uploading, wait 1–2 minutes and then visit `https://yourdomain.com/llms.txt` in your browser. If you see the raw text, you are live.

## How to Test Your llms.txt

Publishing the file is not the same as knowing it works. Here is how to verify it.

**Step 1: Check it loads.** Open `https://yourdomain.com/llms.txt` in an incognito browser window. You should see plain text. If you get a 404 or an HTML page instead, the file is in the wrong location.

**Step 2: Validate the markdown.** Paste the file contents into any markdown previewer (Dillinger.io, StackEdit, or a local editor with preview). Check that headings, bullets, and links render correctly.

**Step 3: Check every link.** Click through each URL in the file. Broken links in your llms.txt are worse than no file at all — they tell the AI your site is untrustworthy.

**Step 4: Ask Claude to read it.** This is the real test. Connect ToolRouter to Claude and ask: *"Read https://yourdomain.com/llms.txt and summarise what this business does."* If Claude returns a clean, accurate summary, your file is working.

To connect ToolRouter, in Claude go to **Settings → Connectors → Add custom connector** and enter:

- **Name:** ToolRouter
- **URL:** `https://api.toolrouter.com/mcp`

Works in Claude chat, Claude Desktop, and Cowork. No download required. Once connected, you can use the [SEO tool](/tools/seo) and [GEO tool](/tools/geo) to run full audits on your pages, including llms.txt checks. Get started at [toolrouter.com/connect](/connect).

## Common llms.txt Mistakes

The five mistakes that show up most often in llms.txt audits:

1. **Putting the file in the wrong place.** It must be at the root. `yourdomain.com/llms.txt` works; `yourdomain.com/seo/llms.txt` does not.
2. **Writing marketing fluff instead of facts.** The description should tell an AI what you sell and who buys it — not "we are the world's leading provider of innovative solutions". AI systems ignore adjectives.
3. **Including every page on the site.** llms.txt is a curated menu, not a sitemap. If you include 500 links, none of them get the attention they need. Stick to your 10–30 most important pages.
4. **Linking to pages that are blocked by robots.txt.** If a page is blocked for crawlers, an AI cannot read it either. Either unblock the page or remove it from llms.txt.
5. **Forgetting to update it.** When you launch a new product, rewrite your pricing page, or retire a guide, the llms.txt file needs to reflect that. A stale llms.txt is worse than no llms.txt.

Frequently Asked Questions

Is llms.txt required?

**No. llms.txt is a voluntary, community-proposed standard — no AI platform currently requires it.** But voluntary does not mean useless. robots.txt was also voluntary when it launched in 1994, and today every website on the internet has one. llms.txt is following the same trajectory, and the sites that adopt it early are the ones showing up in AI citations today.

Do ChatGPT and Claude actually read llms.txt?

**Yes, when they visit your site in response to a user question, both Claude and ChatGPT-family models will look for an llms.txt file and use it to understand your content.** They do not use llms.txt during training — it is used at "inference time", meaning when a user asks a question that requires the AI to check your website right now. Anthropic has publicly stated that Claude's web reading capabilities look for structured signals like llms.txt when available.

What's the difference between llms.txt and robots.txt?

**robots.txt tells search engine crawlers which pages they are allowed to access; llms.txt tells AI systems which pages are most important to read.** robots.txt is about permission and blocking. llms.txt is about curation and comprehension. You should have both files — they do not overlap or conflict. See the full comparison table earlier in this guide.

How often should I update my llms.txt?

**Update your llms.txt whenever the structure or focus of your site changes meaningfully.** That usually means reviewing it every 1–3 months and rewriting it every time you launch a new product, publish a major new content section, rewrite your pricing page, or retire old content. If your business is stable, one review per quarter is enough. If you ship weekly, check it weekly.

Does llms.txt help with Google SEO?

No — llms.txt is specifically for AI language models, not search engine crawlers. Google still relies on sitemap.xml and robots.txt for traditional search indexing. However, as Google's AI Overviews and Gemini-powered search features roll out, llms.txt may start influencing those results indirectly. Publishing one now is a low-cost bet on where search is going.

Can I have more than one llms.txt file?

The official spec also mentions an optional `llms-full.txt` file, which is an expanded version that includes the full content of your key pages (not just links to them). This is useful for technical documentation sites where AI systems need the actual content, not just URLs. For most businesses, the standard `llms.txt` is enough.

B
Founder at ToolRouter
Share this article

Related Posts