๐Ÿ“– User Guide

How to use the AI Bot Access toolkit

Three tools, one goal: make sure AI models can read your site and describe it accurately. Here's what each tool does, when to use it, and how to get the most out of it.

Contents
  1. Why this matters (the short version)
  2. Tool 1 โ€” robots.txt Scanner
  3. Tool 2 โ€” Free llms.txt Generator
  4. Tool 3 โ€” AI llms.txt Builder (BYOK)
  5. Which tool should I use?
  6. The recommended workflow
  7. Troubleshooting & FAQ

Why this matters (the short version)

AI-generated answers are replacing the top of search. When someone asks ChatGPT, Claude, Perplexity or Google's AI Overviews a question, the models decide which sources to cite based on two things: whether they're allowed to read your site, and whether they understand what your site is about.

The AI Bot Access toolkit covers both sides of that equation:

"llms" stands for Large Language Models โ€” the AI systems behind ChatGPT, Claude, Perplexity, and Google's Gemini.

Tool 1 โ€” robots.txt Scanner

Free ยท Public ยท No signup
robots.txt Scanner

What it does: Fetches your site's robots.txt, parses it using the same logic Google's robots parser uses, and resolves each major AI crawler against the ruleset. For each bot you see Allowed, Partially allowed, or Blocked, plus an explanation of why.

When to use it

Benefits

How to use it

  1. Go to aibotaccess.com and enter any domain (e.g. example.com or https://example.com).
  2. Click Check Site. The scanner fetches /robots.txt via a server-side proxy (no CORS issues) and analyzes it.
  3. Review the summary: Allowed, Partial, and Blocked counts tell you at a glance where you stand.
  4. Scroll the per-bot list to see exactly which rule (if any) is blocking each crawler.
  5. If you have blocks to fix, click ๐Ÿ“‹ Copy to Clipboard under the recommended robots.txt section and paste the result into your site's robots.txt file at the root of your domain.
  6. Re-scan to confirm the new rules behave as expected.

Reading the results

Tip. If a block shows up that you don't recognize, check for: (1) a bare User-agent: * with Disallow: /, (2) a WAF/firewall rule at your CDN (Cloudflare, etc.) that blocks bots before they even see robots.txt, or (3) a stale cached robots.txt served by your CDN.

Tool 2 โ€” Free llms.txt Generator

Free ยท Public ยท No signup
Free llms.txt Generator

What it does: Reads your site's sitemap.xml and homepage, ranks the most likely important pages, and drops them into a valid llms.txt template you can edit, copy, and publish. Purely template-based โ€” no AI call, no cost.

When to use it

Benefits

How to use it

  1. Go to the generator and enter your site URL.
  2. Click Generate llms.txt. The tool fetches your homepage (for title and description) and your sitemap (for URLs), then builds the template.
  3. Review the draft in the editable textarea. It will look like:
    # Your Site
    > One-line summary from your meta description
    
    ## Key pages
    - [Home](https://yoursite.com): replace this with a one-line description...
    - [About](https://yoursite.com/about): replace this...
  4. Replace the "replace this" placeholders with real one-line descriptions of each page.
  5. Click ๐Ÿ“‹ Copy or โฌ‡ Download llms.txt.
  6. Upload the file to the root of your domain so it's reachable at https://yoursite.com/llms.txt.
Where to put it. llms.txt belongs at your domain root, same place as robots.txt. In WordPress, a plugin or a rule in your theme's functions. In Next.js/Astro/static sites, drop it in the public/ folder. In Nginx/Apache, it's a file in your docroot.

Tool 3 โ€” AI llms.txt Builder (BYOK)

Premium ยท Access code ยท BYOK
AI llms.txt Builder

What it does: Crawls up to 200 of your pages, reads their actual content, and hands everything to Claude or GPT. The AI writes a full llms.txt with real, tailored, brand-aware descriptions โ€” no templates, no placeholders. You bring your own Anthropic or OpenAI API key, so you pay only the provider's per-token cost (pennies per site, typically).

When to use it

Benefits

Before you start

How to use it

  1. Visit the AI Builder landing page, enter your access code, and click Launch Builder. Your code is remembered on the device.
  2. On the builder page, enter your site URL (e.g. https://example.com).
  3. Pick your AI provider (Anthropic or OpenAI) and model. Claude Sonnet 4.6 is the recommended default; Haiku 4.5 is faster and cheaper for simple sites; Opus 4.6 is overkill for most cases.
  4. Paste your API key. Leave "Remember key on this device" checked unless you're on a shared machine.
  5. Use the Pages to scan slider โ€” 15 is the sensible default; go higher for large content sites. Watch the Estimated cost box update in real time.
  6. (Optional) Add brand notes or instructions in the text area โ€” e.g., "we're B2B enterprise, avoid marketing fluff" or "emphasize our developer audience."
  7. Click Generate llms.txt with AI โ†’.
  8. Watch the progress bar: homepage โ†’ sitemap โ†’ crawl N/Total โ†’ Claude/OpenAI call โ†’ done.
  9. If you need to abort, click โ–  Stop & save. Any pages already fetched are saved as a draft template so you don't lose work.
  10. Review the generated llms.txt. Edit freely, then ๐Ÿ“‹ Copy or โฌ‡ Download.
  11. Upload the file to your domain root at /llms.txt.
Cost estimate. A typical 15-page site run costs $0.05โ€“$0.15 on Haiku, $0.30โ€“$0.80 on Sonnet. A large 100-page site on Sonnet might run $2โ€“$5. These are your API costs paid directly to Anthropic or OpenAI โ€” we don't mark anything up.

Tips for better output

Which tool should I use?

Goal Best tool
Check if AI can read my siterobots.txt Scanner
Diagnose why a specific bot is blockedrobots.txt Scanner
Get a quick llms.txt template for a small siteFree Generator
Write llms.txt descriptions by hand with a good starting pointFree Generator
Produce a polished llms.txt for a large content siteAI Builder
Run llms.txt as an agency deliverableAI Builder
Generate multiple client llms.txt files fastAI Builder

The recommended workflow

Here's the end-to-end sequence we suggest for any site:

  1. Scan first. Run the robots.txt Scanner. If AI bots are blocked, paste the recommended snippet into your robots.txt and verify with a re-scan. No point writing an llms.txt if bots can't read the site.
  2. Generate a starter llms.txt. Run the Free Generator to get a structurally-correct baseline. For small sites this may be all you need after a round of manual edits.
  3. Upgrade to AI when it matters. For client work, marketing-heavy sites, or sites where description quality has real business impact, run the AI Builder with Sonnet 4.6 and 15โ€“30 pages.
  4. Publish. Upload the final llms.txt to your domain root so it's live at /llms.txt.
  5. Verify. Run the scanner one more time โ€” it'll confirm your new llms.txt is discoverable.
  6. Re-check quarterly. AI crawlers come and go, and your pages evolve. A 5-minute re-scan every 3 months is a small insurance premium on your AI visibility.

Troubleshooting & FAQ

The scanner says my robots.txt is blocking GPTBot, but I didn't write that rule.

Most often this is a bare User-agent: * + Disallow: / combination, or a security plugin/firewall that injects rules. Start by viewing your robots.txt in a browser (yoursite.com/robots.txt) and checking the raw content. If it's empty but bots still fail, suspect a CDN/WAF rule blocking by user-agent at a layer above robots.txt.

Will allowing AI crawlers hurt my Google rankings?

No. AI crawlers (GPTBot, ClaudeBot, etc.) are separate from traditional search crawlers (Googlebot, Bingbot). Allowing one does not affect the other.

The free generator only uses my sitemap โ€” what if I don't have one?

If there's no sitemap at /sitemap.xml or /sitemap_index.xml, the generator builds a template from your homepage alone. You can still edit the result to add key URLs manually.

How do I get an AI Builder access code?

Reach out via the waitlist on the landing page. Codes are limited during early access.

Is my API key safe in the AI Builder?

Your key is sent only to your chosen provider (Anthropic or OpenAI). It's stored in your browser's localStorage if you tick "Remember key on this device" so you don't re-enter it. Our serverless proxy relays the one request to the provider and does not log request bodies or keys. If you prefer, untick "Remember" after each run, or click Sign out to clear everything.

What if the AI-generated llms.txt isn't quite right?

The output is always shown in an editable textarea. Tighten voice, reorder sections, remove pages you don't want featured, or re-run with different instructions. The ๐Ÿ”„ Regenerate button reuses your domain and settings for a fresh pass.

Can I use these tools for client sites?

Yes โ€” especially the AI Builder. Nothing in the workflow identifies you or your clients to us, and the output is yours to ship under your own brand.

How often should I re-generate my llms.txt?

When you add or remove major pages, when your site's messaging changes materially, or every 6โ€“12 months as a refresh. Unlike robots.txt, llms.txt isn't fetched on every crawl โ€” models pull it occasionally as context, so minor page changes don't warrant re-generation.