Linux 软件免费装
Banner图

LLMs.txt Curator

开发者 seanmullins
更新时间 2026年4月6日 04:32
PHP版本: 7.4 及以上
WordPress版本: 6.9
版权: GPLv2 or later
版权网址: 版权信息

标签

chatgpt ai seo llm llms.txt ai visibility

下载

1.4.6

详情介绍:

LLMs.txt Curator helps site owners generate and curate a high-quality llms.txt file for AI assistants and retrieval systems, ensuring only relevant, well-described content is exposed to large language models. It generates and maintains llms.txt and llms-full.txt — the emerging standard for telling AI systems (ChatGPT, Claude, Perplexity, Gemini, and others) which pages on your site matter most and what they contain. Unlike auto-generators that dump every URL into a flat file, LLMs.txt Curator takes a curation-first approach. You choose the pages, organise them into sections, fill descriptions, override titles for AI, validate quality, and see exactly which AI bots are reading your file — all from a single interface. What makes this different Most llms.txt plugins treat the file as a static output. LLMs.txt Curator treats it as a living asset: Core features Description Suggestions in detail When pages lack descriptions, AI systems get less context. The suggestion engine fills the gap automatically:
  1. Schema markup (Rank Math, custom _schema_json)
  2. SEO plugin meta description
  3. WordPress excerpt
  4. Open Graph description (_og_description / og_description)
  5. First 160 characters of post content
Already-set descriptions are never touched. Pages that can't be filled automatically are listed for manual review. Quality Score Every generated llms.txt ends with a coverage report: Quality Score: 94% Pages included: 48 Pages with descriptions: 45 Pages missing descriptions: 3 This is visible to the AI systems reading your file, and to you in the Preview tab. AI Crawler Analytics Track 12 known bots: GPTBot, ChatGPT-User, ClaudeBot, PerplexityBot, Google-Extended, Applebot-Extended, Meta-ExternalAgent, Bytespider, CCBot, Cohere, DeepSeek, Amazonbot. The 7-day analytics card shows a visual bar chart of recent activity. All-time totals are kept separately. IP addresses are anonymised before storage — last octet zeroed for IPv4, last 80 bits for IPv6. No data leaves your server. Safety Mode Before generation, the validator checks: When Safety Mode is on, errors block generation and the results are shown immediately. Warnings are surfaced but don't block. Everything is explained in plain English. Scheduled regeneration How to get started
  1. Install and activate. The plugin auto-scans your content and creates initial sections.
  2. Curate: drag, drop, add, remove. Aim for 20-60 pages that best represent your site.
  3. Click Generate Missing Descriptions to fill gaps automatically.
  4. Fix any warnings shown in the Safety Mode card.
  5. Set the update mode that suits your workflow.
  6. Save & Generate. Your files are live at /llms.txt and /llms-full.txt.
  7. Enable the AI Crawler Log to see which bots start visiting.
WooCommerce integration When WooCommerce is active, the plugin automatically includes SKU, price, and stock status in llms.txt descriptions, and full product details in llms-full.txt. Products with "hidden" visibility are excluded, and you can optionally exclude out-of-stock products. Developer hooks

安装:

  1. Upload the llms-txt-curator folder to /wp-content/plugins/.
  2. Activate through the 'Plugins' menu.
  3. Go to Settings -> LLMs.txt Curator to begin curating.
Or install directly from the WordPress plugin repository by searching for "LLMs.txt Curator".

屏幕截图:

  • Description Suggestions -- one-click gap-filling with filled/review summary
  • Settings -- Update Mode selector and Safety Mode validation card
  • AI Crawler Analytics -- 7-day bot activity bar chart
  • Preview & Validate -- live llms.txt output with Quality Score footer
  • Change Detection -- banner alerting when curated pages have been updated

升级注意事项:

1.4.6 Adds explicit PHPCS ignore annotations for JSON blob input lines and updates Tested up to to 6.9. No functional changes. 1.4.5 Security fix: type validation hardening in sanitize_all_settings() and helper methods. No breaking changes. 1.4.4 Security fix: sanitizes preview AJAX input through the same sanitize_all_settings() pipeline used by save. No breaking changes. 1.4.0 Adds 30-day crawler analytics chart, last-crawled banner, first/last seen per bot, CSV export, and two new REST API endpoints. No breaking changes — existing data is preserved and the new daily aggregate store builds automatically from future visits. 1.3.5 Internal prefix change (llms_txt_ → llmscu_) for WordPress guidelines compliance. If you installed from the plugin repository, existing settings are preserved. If upgrading from a direct download, existing stored data will be re-initialised on first save. 1.3.3 Critical fix: resolves llms.txt 404 introduced in v1.3.2. Upgrade immediately if on v1.3.2. 1.3.2 Fixes UTF-8 encoding bug (em dashes rendering as â€"). Adds X-Robots-Tag: noindex option (on by default). No breaking changes — existing config preserved. 1.3.1 Security and robustness hardening: nosniff header, import size/schema validation, file size guard, query performance flags, diagnostics panel, health banner. Translation-ready. No breaking changes. 1.3.0 Adds full WordPress Multisite support. Existing single-site installs are unaffected. No breaking changes. 1.2.2 Adds 7-day AI crawler analytics. No breaking changes. 1.2.1 Adds Scheduled Regeneration and Safety Mode. No breaking changes. 1.2.0 Major quality-of-life release. Description Suggestions, Quality Score, Change Detection, Title Overrides. No breaking changes.

常见问题:

What is llms.txt?

llms.txt is a proposed standard (llmstxt.org) that provides AI systems with a curated, Markdown-formatted overview of a website's most important content. A strategic selection of the pages you want AI to know about -- not a sitemap.

What is llms-full.txt?

The companion file defined in the same spec. While llms.txt contains links and short descriptions, llms-full.txt contains the full Markdown content of each page. Optional -- enable it in Settings when ready.

Do I need both files?

No. llms.txt alone is sufficient. llms-full.txt is useful if you want AI systems to have immediate access to your full content without additional crawling.

What does Description Suggestions do?

Scans every curated page and fills missing descriptions using a five-step fallback chain: schema markup -> SEO meta -> excerpt -> Open Graph -> page content. Pages with descriptions already set are never touched.

What is the Quality Score?

A percentage showing how many listed pages have descriptions. It appears at the bottom of your generated llms.txt, visible to both you and any AI systems reading the file.

What does Title Override do?

Lets you set a different title for a page in your llms.txt output without changing it on your site. Useful when your WordPress title includes your site name but you want AI to see a cleaner, more descriptive title.

What does Safety Mode do?

Runs validation before every generation. If errors are found, generation is blocked and results shown immediately. Prevents broken or malformed files going live.

Which SEO plugins are supported?

Rank Math, Yoast SEO, All in One SEO, SEOPress, and The SEO Framework.

Which AI bots does the crawler log detect?

GPTBot (OpenAI), ChatGPT-User, ClaudeBot (Anthropic), PerplexityBot, Google-Extended, Applebot-Extended, Meta-ExternalAgent, Bytespider (ByteDance), CCBot (Common Crawl), Cohere, DeepSeek, and Amazonbot.

What if my host blocks file writes?

The plugin has a rewrite rule fallback that serves both files via WordPress. Choose between "Direct file", "Rewrite rule only", or "Both" (recommended) in Settings.

Does this work on multisite?

Yes. Activate network-wide from Network Admin > Plugins, or activate per-site on individual sub-sites. Each site manages its own independent llms.txt — there is no shared network file. The Network Admin overview page (Network Admin > Settings > LLMs.txt Curator) shows every site's generation status and lets you regenerate any site, or all sites at once, without leaving the network admin. On subdirectory networks (example.com, example.com/site1) each site writes a physical file at its own path. On subdomain networks (example.com, site1.example.com) sub-sites share a filesystem root, so they serve llms.txt via WordPress rewrite rule from the database instead — this is fully correct and functionally identical. WP-CLI works per-site using the standard --url= flag: wp llms-txt regenerate --url=https://site1.example.com

Is any data sent externally?

No. Everything stays on your server. No telemetry, no external API calls, no cookies. Crawler IP addresses are anonymised before storage.

更新日志:

1.4.6 1.4.5 1.4.4 1.4.0 1.3.5 1.3.3 1.3.2 1.3.1 1.3.0 1.2.2 1.2.1 1.2.0 1.1.0 1.0.4