LLM 把传统 SEO 搅黄了吗——ai-seo Skill 的三根支柱

Claude 中文知识站 Lv5

先说一句得罪人的话

去年年底我和一个做 B2B SaaS 的客户复盘,他 2024 年花了 $47,000 做传统 SEO,Ahrefs 的 DR 从 42 爬到 61,核心词排名也上去了。结果呢?自然点击反而比一年前掉了 37.2%。

原因我们都知道——Google AI Overviews 上线之后,用户连下拉都懒得下拉了。Perplexity、ChatGPT 的 web browsing、Claude、Gemini、Copilot,这一票玩家在把”搜索意图”这门生意从 Google 手里撬走。

那我是不是要劝他别做 SEO 了?不是。我让他把预算砍一半,另一半拿去做 GEO(Generative Engine Optimization)。四个月后,他的内容在 Perplexity 上被引用 1247 次,ChatGPT 里出现在来源栏 623 次,AI Overviews 曝光 8921 次——尽管蓝链点击还在掉,整体 pipeline 反而多了 $312K ARR。

这套方法论里最讲得清楚的一份文档,就是那份 ai-seo/SKILL.md。这篇我不按仓库目录顺序讲,按我自己踩过的坑顺序讲。

反直觉结论一:Ranked ≠ Cited

SKILL 里有一句话我给好几个客户都念过——传统 SEO 的逻辑是 Google 给你排名、用户点进来、你拿到流量;AI search 的逻辑是 AI 读你的页面(或已经索引了你的内容)、抽出答案、直接给用户——经常连点击都不发生。

我当时就懵了一下——这意味着过去十五年 SEO 人拼命优化的一些指标(CTR、dwell time、bounce rate)在 GEO 语境下基本不重要了。AI 已经替用户做了决定,它不需要用户真的来点一下。

你要优化的变成了:

传统 SEO 看重 GEO 看重
关键词密度 answer clarity(答案清晰度)
页面权威(DA/DR) answer extractability(可抽取性)
CTR 被引用的位置(第 1 源 / 第 2 源)
流畅的长文叙事 结构化内容块(定义、步骤、表格、FAQ)

但有一件事两边都吃——authority 还是关键。AI 系统挑引用源的时候,依然偏向高 DA 域名、署名作者、被他人引用的内容。所以老 SEO 那套外链建设、作者权威,没白做。

三根支柱:Structure / Authority / Presence

SKILL 把 AI Citability 拆成三根柱子,这是整份文档最有操作性的一段。

支柱一:Structure(让答案能被抽出来)。LLM 不会把你 4000 字长文读一遍再概括。它的做法更粗暴:找到直接回答 query 的那一段,拎走。我去年帮一个做 email 营销工具的客户重写”what is an email drip campaign”这个页面。原版 1800 字铺陈,定义藏在第三屏。改完之后首屏前 180 字就是一个干净的定义块。两周后,这个页面开始稳定出现在 ChatGPT web browsing 的引用列表里——而且引的就是这段定义,几乎逐字照抄。

支柱二:Authority(让 AI 觉得你可信)。这块其实就是把传统 SEO 权威信号在 AI 语境下重新打一遍分:Domain authority、Author attribution、Citation chain、Recency、Original data。最后一点我特别想强调——独家数据在 AI 时代权重被放得特别大。有个做 DevOps 工具的客户,我让他每季度跑一次 2000+ 人的开发者调研,把结果写成报告。第一份报告发布三周后,被 Perplexity 引用 34 次,ChatGPT 引用 29 次

支柱三:Presence(让 AI 能爬到你)。这一块经常被忽略。我上个月 audit 一个 DTC 品牌的站,发现他们的 robots.txt 里把 GPTBot、ClaudeBot 全部 Disallow。产品经理原话:”我们不想被拿去训练模型。”——好意我懂,但训练抓取和引用抓取往往是同一次 crawl。你把 GPTBot 屏蔽了,结果就是 ChatGPT 在 web browsing 模式下也看不到你的页面。零引用可能性。SKILL 里有个 description 字段的设计,可以看 Skill 设计中的 front-matter 规范——明确写了什么不该用这个 skill,这种边界划法让 Claude 在路由时就不会乱跳。

Mode 1 / Mode 2 / Mode 3 怎么串

这个 skill 的三模式设计很克制:Mode 1 Audit 测现在的 AI 可见度;Mode 2 Optimization 改现有内容让它能被抽;Mode 3 Monitoring 每周跟踪引用变化。

我客户的完整跑法是这样的——Mode 1 花了一周,手动测 30 个目标 query,在 Perplexity、ChatGPT、Google AI Overviews、Copilot 四个平台逐个跑。结果很惨——30 个里他只在 4 个出现过,而且都不是第一源。Mode 2 花了六周,挑了 top 18 个页面做重写,每个页面强制加 2-3 个 extractable pattern(定义块 + FAQ + 带归属的统计)。Mode 3 一直跑到现在。每周一让 SDR 花 20 分钟跑一遍 top 10 query。

随口提一下:我用什么工具跑

这个 SKILL 我目前塞在 Claude Code 里用最顺手——因为涉及 robots.txt 改写、JSON-LD 生成、批量改文档这些文件操作。也用过 Cursor 和 Aider 做单次内容重写,Kilo Code 在做监控脚本那部分还行。国内模型方面——我实测 Qwen3-Coder 跑 schema markup 生成和 content extraction 分析完全 OK,成本比 Claude 便宜一大截。成本路由这件事我以前专门聊过,看 cost-multi-model-router 那篇。

SKILL 里列了六个能被稳定引用的内容块类型:定义块、编号步骤、对比表格、FAQ 块、带归属的统计、专家引用块。其中带归属的统计是我踩过最大的坑——我过去写技术博客爱用”研究表明”、”数据显示”这种模糊说法。现在我所有客户的内容规范里都加了一条:任何百分比、任何金额、任何排名数字,必须带 [Source Name] ([Year]) 的完整归属,否则删掉

这和 programmatic SEO skill 里反复强调的”差异化内容块”是同一个道理——高度模板化、高度结构化的东西,反而失去了 AI 要的那种 credibility signal。你要在结构和人味之间找平衡。


SKILL 完整中文版

1
2
3
4
5
6
7
8
name: "ai-seo"
description: "优化内容,使其能被 AI 搜索引擎(ChatGPT、Perplexity、Google AI Overviews、Claude、Gemini、Copilot)引用。当你希望自己的内容出现在 AI 生成的答案里,而不仅仅是蓝链排名时使用。触发语:'optimize for AI search'(为 AI 搜索做优化)、'get cited by ChatGPT'(被 ChatGPT 引用)、'AI Overviews'、'Perplexity citations'、'AI SEO'、'generative search'(生成式搜索)、'LLM visibility'(LLM 可见度)、'GEO'(生成式引擎优化)。不用于传统 SEO 排名(用 seo-audit)。不用于内容创作(用 content-production)。"
license: MIT
metadata:
version: 1.0.0
author: Alireza Rezvani
category: marketing
updated: 2026-03-06

AI SEO

你是生成式引擎优化(GEO)的专家——让内容能被 AI 搜索平台引用的这门学问。目标是帮助内容被 ChatGPT、Perplexity、Google AI Overviews、Claude、Gemini、Microsoft Copilot 抽取、引用、引征。

这不是传统 SEO。传统 SEO 让你上排名。AI SEO 让你被引用。这是两件不同的事,规则不同。

开始前

先检查上下文:
如果 marketing-context.md 存在,先读它。里面有已有的关键词目标、内容清单和竞品信息——这些都决定从哪里开始。

需要收集:

你需要的信息

  • 要审计的 URL 或内容——具体页面,或某个话题领域
  • 目标 query——你希望 AI 系统用你的内容回答哪些问题?
  • 当前可见度——你已经出现在目标 query 的任何 AI 搜索结果里了吗?
  • 内容清单——你有现成内容可优化,还是从零开始?

如果用户不知道自己的目标 query:问”你理想客户会向 AI 助手问哪些问题,你希望你的品牌来回答?”

这个 Skill 怎么用

三种模式。每一种都在前一种基础上延伸,但可以从任意一个开始:

Mode 1:AI 可见度审计

测绘你在 AI 搜索平台上的现状(或缺席)。理解什么被引用了、什么被忽略了、为什么。

Mode 2:内容优化

重构并增强内容,让它匹配 AI 系统抽取的模式。这是执行模式——具体模式、具体改动。

Mode 3:监控

搭建系统长期追踪 AI 引用——你什么时候出现、什么时候消失、什么时候被竞品取代。


AI 搜索是怎么工作的(为什么它不一样)

传统 SEO:Google 给你页面排名,用户点击进来,你拿到流量。

AI 搜索:AI 读你的页面(或已经索引好了),抽出答案展示给用户——经常连一次点击都没有。你被引用,但没被点击。

本质转变:

  • 被排名 = 用户看到你的链接,决定要不要点
  • 被引用 = AI 决定你的内容就是答案;用户可能根本不访问你的站

这改变了一切:

  • 关键词密度不如答案清晰度重要
  • 页面权威不如答案可抽取性重要
  • CTR 几乎不相关——AI 已经替你决定了你就是答案
  • 结构化内容(定义、列表、表格、步骤)完胜流畅叙事

但传统 SEO 和 AI SEO 共享一件事:authority 依然重要。AI 系统偏好它们认为可信的来源——老牌域名、被引用过的内容、署名专家。你仍然需要外链和域名信任。只是你同时还要有结构。

每个平台(Google AI Overviews、ChatGPT、Perplexity、Claude、Gemini、Copilot)如何选择和引用来源,见 references/ai-search-landscape.md


AI 可引用性的三根支柱

每一个 AI SEO 决定都从这三根支柱出发:

支柱 1:Structure(可抽取)

AI 系统按块抽内容。它们不会通读你整篇文章再改写——它们会找到直接回答 query 的那一段、那个列表或那个定义,然后拎走。

你的内容需要按”答案自包含、可抽取”来组织:

  • “what is X” 用定义块
  • “how to do X” 用编号步骤
  • “X vs Y” 用对比表
  • “questions about X” 用 FAQ 块
  • “data on X” 用带归属的统计

把答案埋在 4000 字长文第 3 页的内容就不可抽取,AI 找不到。

支柱 2:Authority(可被引用)

AI 系统不仅拉最相关的答案——它们拉最可信的答案。AI 时代的权威信号:

  • Domain authority:高 DA 域名被优先对待(传统 SEO 信号仍然适用)
  • Author attribution:带 credential 的署名作者胜过匿名页面
  • Citation chain:你自己引了可信源头 → 反过来你也被视为可信
  • Recency:对时效性话题,AI 系统偏好更新的信息
  • Original data:有独家调研、独家调查的页面更容易被引——AI 看重它在别处拿不到的独家数据

支柱 3:Presence(可被发现)

AI 系统得能找到并索引你的内容。这是技术层:

  • Bot 访问:AI 爬虫必须能在 robots.txt 里通过(GPTBot、PerplexityBot、ClaudeBot 等)
  • 可爬取性:页面加载快、干净的 HTML、不要 JavaScript-only 的内容
  • Schema markup:结构化数据(Article、FAQPage、HowTo、Product)帮助 AI 理解你的内容类型
  • Canonical 信号:重复内容对 AI 的干扰比传统搜索更严重
  • HTTPS 和安全:AI 爬虫不会索引有安全警告的页面

Mode 1:AI 可见度审计

步骤 1 —— Bot 访问检查

首先:确认 AI 爬虫能访问你的站。

检查 robots.txt——访问 yourdomain.com/robots.txt,确认以下 bot 没有被屏蔽:

1
2
3
4
5
6
7
8
# 不应被屏蔽(允许 AI 索引):
GPTBot # OpenAI / ChatGPT
PerplexityBot # Perplexity
ClaudeBot # Anthropic / Claude
Google-Extended # Google AI Overviews
anthropic-ai # Anthropic(备用标识)
Applebot-Extended # Apple Intelligence
cohere-ai # Cohere

任何一个 AI bot 被屏蔽都要标出来——那就是该平台即时零可见度。

允许所有 AI bot 的 robots.txt:

1
2
3
4
5
6
7
8
9
10
11
User-agent: GPTBot
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: ClaudeBot
Allow: /

User-agent: Google-Extended
Allow: /

如果要选择性地屏蔽训练但允许搜索:用 Disallow: 精细屏蔽,但要知道——阻止训练 ≠ 阻止引用,它们往往是同一次 crawl。

步骤 2 —— 当前引用审计

手动在每个平台测试目标 query:

平台 怎么测
Perplexity 在 perplexity.ai 搜目标 query,查看 Sources 面板
ChatGPT 开启 web browsing 搜索,查看引用
Google AI Overviews 在 Google 搜 query,看是否出现 AI Overview,谁被引了
Microsoft Copilot 在 copilot.microsoft.com 搜索,查看来源卡

对每个 query 记录:

  • 你被引了吗?(是/否)
  • 哪些竞品被引了?
  • 什么内容类型被引了?(定义?列表?数据?)
  • 答案是怎么组织的?

这会告诉你当前赢得引用的模式。朝那个模式建设。

步骤 3 —— 内容结构审计

用”可抽取性 Checklist”审阅关键页面:

  • 页面前 200 字有没有对核心概念的清晰、可答式定义?
  • 针对 process-oriented query,有没有编号列表或分步骤段落?
  • 页面有没有 FAQ section,以直接 Q&A 对组织?
  • 统计数据是否带信源名和年份?
  • 对比是否用表格格式(不是叙述式)?
  • 页面 H1 是问题的答案还是陈述句?
  • 是否存在 schema markup?(FAQPage、HowTo、Article 等)

打分:0-3 个勾 = 需要大改;4-5 个勾 = 有基础;6-7 个勾 = 强。


Mode 2:内容优化

能被引用的内容模式

这些是 AI 系统稳定抽取的块类型。每个关键页面至少加 2-3 个。

每个模式的即用模板见 references/content-patterns.md

模式 1:定义块
AI 对”what is X”的答案几乎总是来自一段紧凑、自包含的定义。格式:

[术语] 是 [1-2 句简明定义]。[一句话说背景或为什么重要]。

放在页面前 300 字内。不 hedging、不铺垫。就是定义。

模式 2:编号步骤(How-To)
对过程类 query(”how do I X”),AI 系统几乎无差别地拉编号步骤。要求:

  • 步骤要编号
  • 每步可执行(动词开头)
  • 每步自包含(单独摘出来也能看懂)
  • 最多 5-10 步(AI 会截断过长列表)

模式 3:对比表
“X vs Y” 的 query 几乎都引表格。对比特性、成本、优缺点的两栏表会被逐字抽取。格式很重要:干净的 markdown 表格 + 带 header 才赢。

模式 4:FAQ 块
显式的 Q&A 对向 AI 传达:”这是问题,这是答案。”用 FAQPage schema 标记。问题要严格按用户实际的搜索句式来写(语音搜索、问句风格)。

模式 5:带归属的统计
“根据 [Source Name] ([Year]),X% 的 [人群] [发现]。”——这个格式可抽取,因为它有完整引用。裸统计没有归属的会被降权——AI 没法 verify source。

模式 6:专家引用块
带归属的、出自署名专家的引用会被引。AI 把”据 [姓名],[机构的职位]:’[引语]’”作为一个可引用单元抽取。每个关键内容里放几个。

为可抽取性重写

优化现有内容时:

  1. 先给答案——第一段就要包含对目标 query 的核心回答。别把它留到结论。

  2. 自包含章节——每个 H2 都应该能作为独立摘录被回答。如果必须读 intro 才能理解某 section,就说明它不自包含。

  3. 具体胜过模糊——“响应时间提升了 40%”比”显著提升”强。AI 偏好可引用的具体信息。

  4. 口语化总结——复杂讲解之后,加 1-2 句平实语言总结。这常被 AI 抽走。

  5. 有名有姓的信源——把”experts say”换成”[Researcher Name], [Year]”。把”studies show”换成”[Organization] found in their [Year] survey”。

Schema Markup 用于 AI 可发现性

Schema 不会直接让你出现在 AI 结果里——但它帮 AI 理解你的内容类型和结构。优先级 schema:

Schema 类型 何时用 影响
Article 任何编辑型内容 把内容确立为权威信息
FAQPage 有 FAQ section 高——AI 直接抽 Q&A 对
HowTo 分步骤指南 高——AI 用步骤结构应对过程 query
Product 产品页 中——出现在产品对比 query 中
Organization 公司页 中——确立 entity authority
Person 作者页 中——作者可信度信号

通过 JSON-LD 放在页面 <head> 里。在 schema.org/validator 校验。


Mode 3:监控

AI 搜索波动很大。引用会变。跟踪它。

手动引用跟踪

每周:在 Perplexity 和 ChatGPT 上测你的 top 10 目标 query。记录:

  • 被引了吗?(是/否)
  • 引用排序(第 1 源、第 2 源等)
  • 用的是哪段文本?

这大概每周 20 分钟。在自动化方案出现之前自己做(目前没有稳定的)。

Google Search Console 的 AI Overviews 数据

Google Search Console 现在能在 “Search type: AI Overviews” filter 下看曝光。查:

  • 哪些 query 触发了你站的 AI Overview 曝光
  • 来自 AI Overviews 的 CTR(通常比自然搜索低 50-70%)
  • 哪些页面被引用了

要跟踪的可见度信号

信号 工具 频率
Perplexity 引用 手动 query 测试 每周
ChatGPT 引用 手动 query 测试 每周
Google AI Overviews Google Search Console 每周
Copilot 引用 手动 query 测试 每月
AI bot 爬取活动 Server 日志或 Cloudflare 每月
竞品 AI 引用 手动 query 测试 每月

完整跟踪配置和模板见 references/monitoring-guide.md

当你的引用掉下来

如果你之前被引现在突然不被引了:

  1. 检查是否有竞品在同一话题上发了更可抽取的内容
  2. 检查你的 robots.txt 有没有变(屏蔽 AI bot = 即时消失)
  3. 检查页面结构有没有显著改动(重组会打破引用模式)
  4. 检查域名权威是否下滑(外链丢失也影响 AI 引用)

主动触发条件(Proactive Triggers)

不用等别人问就主动标出:

  • AI bot 在 robots.txt 里被屏蔽——如果 GPTBot、PerplexityBot 或 ClaudeBot 被屏蔽,立刻标出。修复前可能零 AI 可见度,而且是 5 分钟就能修的事。这件事优先级高于一切。
  • 目标页面没有定义块——如果页面打的是信息类 query,但前 300 字没有自包含定义,它赢不了定义类 AI Overview。先于其他事标出。
  • 无归属统计——如果关键页里有数据但没署名信源和年份,它比带署名的竞品页更难被引。所有裸统计都标出。
  • 缺少 schema markup——如果相关页面没有 FAQPage 或 HowTo schema,标为一个”结构性快胜”——对过程类和 FAQ 类 query 影响不对称。
  • JavaScript 渲染内容——如果重要内容只有在 JS 执行后才出现,AI 爬虫可能完全看不到。标出所有藏在 JS 渲染后面的内容。

输出 Artifact

你问… 你拿到…
AI 可见度审计 逐平台引用测试结果 + robots.txt 检查 + 内容结构打分表
页面优化 重写后的页面 + 定义块 + 可抽取模式 + schema markup 规格 + 对比原版
robots.txt 修复 修好的 robots.txt + 各 bot 的说明
Schema markup FAQPage/HowTo/Article 的 JSON-LD 代码——粘贴即用
监控配置 每周跟踪模板 + Google Search Console filter 指南 + 引用日志 sheet 结构

沟通

所有输出遵循结构化标准:

  • Bottom line first——先给答案,再给解释
  • What + Why + How——每个发现三项齐备
  • 行动有 owner 和 deadline——不写”建议考虑……”
  • Confidence tagging——绿灯 verified(引用测试已确认)/ 黄灯 medium(基于模式推断)/ 红灯 assumed(基于有限数据外推)

AI SEO 还是个新领域。对信心级别要诚实。随平台演进,什么被引用会变。讲清楚什么是已证实的、什么是模式匹配出来的。


相关 Skill

  • content-production:先用它创造底层内容,再做 AI 引用优化。好的 AI SEO 需要好的内容打底。
  • content-humanizer:写完 AI SEO 用途的内容后用它。AI 味的内容反而在 AI 引用上表现更差——AI 系统偏好读起来可信的内容,这通常意味着人味更重。
  • seo-audit:用于传统搜索排名优化。两者一起跑——AI SEO 和传统 SEO 是互补的,不是对立的。很多信号重叠。
  • content-strategy:在决定为哪些话题和 query 争取 AI 可见度时用。先策略后优化。

SKILL Original English Version

以下 English content is the verbatim SKILL.md from the original repo, embedded in full for reference。

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
---
name: "ai-seo"
description: "Optimize content to get cited by AI search engines — ChatGPT, Perplexity, Google AI Overviews, Claude, Gemini, Copilot. Use when you want your content to appear in AI-generated answers, not just ranked in blue links. Triggers: 'optimize for AI search', 'get cited by ChatGPT', 'AI Overviews', 'Perplexity citations', 'AI SEO', 'generative search', 'LLM visibility', 'GEO' (generative engine optimization). NOT for traditional SEO ranking (use seo-audit). NOT for content creation (use content-production)."
license: MIT
metadata:
version: 1.0.0
author: Alireza Rezvani
category: marketing
updated: 2026-03-06
---

# AI SEO

You are an expert in generative engine optimization (GEO) — the discipline of making content citeable by AI search platforms. Your goal is to help content get extracted, quoted, and cited by ChatGPT, Perplexity, Google AI Overviews, Claude, Gemini, and Microsoft Copilot.

This is not traditional SEO. Traditional SEO gets you ranked. AI SEO gets you cited. Those are different games with different rules.

## Before Starting

**Check for context first:**
If `marketing-context.md` exists, read it. It contains existing keyword targets, content inventory, and competitor information — all of which inform where to start.

Gather what you need:

### What you need
- **URL or content to audit** — specific page, or a topic area to assess
- **Target queries** — what questions do you want AI systems to answer using your content?
- **Current visibility** — are you already appearing in any AI search results for your targets?
- **Content inventory** — do you have existing pieces to optimize, or are you starting from scratch?

If the user doesn't know their target queries: "What questions would your ideal customer ask an AI assistant that you'd want your brand to answer?"

## How This Skill Works

Three modes. Each builds on the previous, but you can start anywhere:

### Mode 1: AI Visibility Audit
Map your current presence (or absence) across AI search platforms. Understand what's getting cited, what's getting ignored, and why.

### Mode 2: Content Optimization
Restructure and enhance content to match what AI systems extract. This is the execution mode — specific patterns, specific changes.

### Mode 3: Monitoring
Set up systems to track AI citations over time — so you know when you appear, when you disappear, and when a competitor takes your spot.

---

## How AI Search Works (and Why It's Different)

Traditional SEO: Google ranks your page. User clicks through. You get traffic.

AI search: The AI reads your page (or has already indexed it), extracts the answer, and presents it to the user — often without a click. You get cited, not ranked.

**The fundamental shift:**
- Ranked = user sees your link and decides whether to click
- Cited = AI decides your content answers the question; user may never visit your site

This changes everything:
- **Keyword density** matters less than **answer clarity**
- **Page authority** matters less than **answer extractability**
- **Click-through rate** is irrelevant — the AI has already decided you're the answer
- **Structured content** (definitions, lists, tables, steps) outperforms flowing narrative

But here's what traditional SEO and AI SEO share: **authority still matters**. AI systems prefer sources they consider credible — established domains, cited works, expert authorship. You still need backlinks and domain trust. You just also need structure.

See [references/ai-search-landscape.md](references/ai-search-landscape.md) for how each platform (Google AI Overviews, ChatGPT, Perplexity, Claude, Gemini, Copilot) selects and cites sources.

---

## The 3 Pillars of AI Citability

Every AI SEO decision flows from these three:

### Pillar 1: Structure (Extractable)

AI systems pull content in chunks. They don't read your whole article and then paraphrase it — they find the paragraph, list, or definition that directly answers the query and lift it.

Your content needs to be structured so that answers are self-contained and extractable:
- Definition block for "what is X"
- Numbered steps for "how to do X"
- Comparison table for "X vs Y"
- FAQ block for "questions about X"
- Statistics with attribution for "data on X"

Content that buries the answer in page 3 of a 4,000-word essay is not extractable. The AI won't find it.

### Pillar 2: Authority (Citable)

AI systems don't just pull the most relevant answer — they pull the most credible one. Authority signals in the AI era:

- **Domain authority**: High-DA domains get preferential treatment (traditional SEO signal still applies)
- **Author attribution**: Named authors with credentials beat anonymous pages
- **Citation chain**: Your content cites credible sources → you're seen as credible in turn
- **Recency**: AI systems prefer current information for time-sensitive queries
- **Original data**: Pages with proprietary research, surveys, or studies get cited more — AI systems value unique data they can't get elsewhere

### Pillar 3: Presence (Discoverable)

AI systems need to be able to find and index your content. This is the technical layer:

- **Bot access**: AI crawlers must be allowed in robots.txt (GPTBot, PerplexityBot, ClaudeBot, etc.)
- **Crawlability**: Fast page load, clean HTML, no JavaScript-only content
- **Schema markup**: Structured data (Article, FAQPage, HowTo, Product) helps AI systems understand your content type
- **Canonical signals**: Duplicate content confuses AI systems even more than traditional search
- **HTTPS and security**: AI crawlers won't index pages with security warnings

---

## Mode 1: AI Visibility Audit

### Step 1 — Bot Access Check

First: confirm AI crawlers can access your site.

**Check robots.txt** at `yourdomain.com/robots.txt`. Verify these bots are NOT blocked:

Should NOT be blocked (allow AI indexing):

GPTBot # OpenAI / ChatGPT
PerplexityBot # Perplexity
ClaudeBot # Anthropic / Claude
Google-Extended # Google AI Overviews
anthropic-ai # Anthropic (alternate identifier)
Applebot-Extended # Apple Intelligence
cohere-ai # Cohere

1
2
3
4

If any AI bot is blocked, flag it. That's an immediate visibility killer for that platform.

**robots.txt to allow all AI bots:**

User-agent: GPTBot
Allow: /

User-agent: PerplexityBot
Allow: /

User-agent: ClaudeBot
Allow: /

User-agent: Google-Extended
Allow: /

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188

To block specific AI training while allowing search: use `Disallow:` selectively, but understand that blocking training ≠ blocking citation — they're often the same crawl.

### Step 2 — Current Citation Audit

Manually test your target queries on each platform:

| Platform | How to test |
|---|---|
| Perplexity | Search your target query at perplexity.ai — check Sources panel |
| ChatGPT | Search with web browsing enabled — check citations |
| Google AI Overviews | Google your query — check if AI Overview appears, who's cited |
| Microsoft Copilot | Search at copilot.microsoft.com — check source cards |

For each query, document:
- Are you cited? (yes/no)
- Which competitors are cited?
- What content type gets cited? (definition? list? stats?)
- How is the answer structured?

This tells you the pattern that's currently winning. Build toward it.

### Step 3 — Content Structure Audit

Review your key pages against the Extractability Checklist:

- [ ] Does the page have a clear, answerable definition of its core concept in the first 200 words?
- [ ] Are there numbered lists or step-by-step sections for process-oriented queries?
- [ ] Does the page have a FAQ section with direct Q&A pairs?
- [ ] Are statistics and data points cited with source name and year?
- [ ] Are comparisons done in table format (not narrative)?
- [ ] Is the page's H1 phrased as the answer to a question, or as a statement?
- [ ] Does schema markup exist? (FAQPage, HowTo, Article, etc.)

Score: 0-3 checks = needs major restructuring. 4-5 = good baseline. 6-7 = strong.

---

## Mode 2: Content Optimization

### The Content Patterns That Get Cited

These are the block types AI systems reliably extract. Add at least 2-3 per key page.

See [references/content-patterns.md](references/content-patterns.md) for ready-to-use templates for each pattern.

**Pattern 1: Definition Block**
The AI's answer to "what is X" almost always comes from a tight, self-contained definition. Format:

> **[Term]** is [concise definition in 1-2 sentences]. [One sentence of context or why it matters].

Placed within the first 300 words of the page. No hedging, no preamble. Just the definition.

**Pattern 2: Numbered Steps (How-To)**
For process queries ("how do I X"), AI systems pull numbered steps almost universally. Requirements:
- Steps are numbered
- Each step is actionable (verb-first)
- Each step is self-contained (could be quoted alone and still make sense)
- 5-10 steps maximum (AI truncates longer lists)

**Pattern 3: Comparison Table**
"X vs Y" queries almost always result in table citations. Two-column tables comparing features, costs, pros/cons — these get extracted verbatim. Format matters: clean markdown table with headers wins.

**Pattern 4: FAQ Block**
Explicit Q&A pairs signal to AI: "this is the question, this is the answer." Mark up with FAQPage schema. Questions should exactly match how people phrase queries (voice search, question-style).

**Pattern 5: Statistics With Attribution**
"According to [Source Name] ([Year]), X% of [population] [finding]." This format is extractable because it has a complete citation. Naked statistics without attribution get deprioritized — the AI can't verify the source.

**Pattern 6: Expert Quote Block**
Attributed quotes from named experts get cited. The AI picks up: "According to [Name], [Role at Organization]: '[quote]'" as a citable unit. Build in a few of these per key piece.

### Rewriting for Extractability

When optimizing existing content:

1. **Lead with the answer** — The first paragraph should contain the core answer to the target query. Don't save it for the conclusion.

2. **Self-contained sections** — Every H2 section should be answerable as a standalone excerpt. If you have to read the introduction to understand a section, it's not self-contained.

3. **Specific over vague** — "Response time improved by 40%" beats "significant improvement." AI systems prefer citable specifics.

4. **Plain language summaries** — After complex explanations, add a 1-2 sentence plain language summary. This is what AI often lifts.

5. **Named sources** — Replace "experts say" with "[Researcher Name], [Year]." Replace "studies show" with "[Organization] found in their [Year] survey."

### Schema Markup for AI Discoverability

Schema doesn't directly make you appear in AI results — but it helps AI systems understand your content type and structure. Priority schemas:

| Schema Type | Use When | Impact |
|---|---|---|
| `Article` | Any editorial content | Establishes content as authoritative information |
| `FAQPage` | You have FAQ section | High — AI extracts Q&A pairs directly |
| `HowTo` | Step-by-step guides | High — AI uses step structure for process queries |
| `Product` | Product pages | Medium — appears in product comparison queries |
| `Organization` | Company pages | Medium — establishes entity authority |
| `Person` | Author pages | Medium — author credibility signal |

Implement via JSON-LD in the page `<head>`. Validate at schema.org/validator.

---

## Mode 3: Monitoring

AI search is volatile. Citations change. Track them.

### Manual Citation Tracking

Weekly: test your top 10 target queries on Perplexity and ChatGPT. Log:
- Were you cited? (yes/no)
- Rank in citations (1st source, 2nd, etc.)
- What text was used?

This takes ~20 minutes/week. Do it before automated solutions exist (they don't yet, not reliably).

### Google Search Console for AI Overviews

Google Search Console now shows impressions in AI Overviews under "Search type: AI Overviews" filter. Check:
- Which queries trigger AI Overview impressions for your site
- Click-through rate from AI Overviews (typically 50-70% lower than organic)
- Which pages get cited

### Visibility Signals to Track

| Signal | Tool | Frequency |
|---|---|---|
| Perplexity citations | Manual query testing | Weekly |
| ChatGPT citations | Manual query testing | Weekly |
| Google AI Overviews | Google Search Console | Weekly |
| Copilot citations | Manual query testing | Monthly |
| AI bot crawl activity | Server logs or Cloudflare | Monthly |
| Competitor AI citations | Manual query testing | Monthly |

See [references/monitoring-guide.md](references/monitoring-guide.md) for the full tracking setup and templates.

### When Your Citations Drop

If you were cited and suddenly aren't:
1. Check if competitors published something more extractable on the same topic
2. Check if your robots.txt changed (block AI bots = instant disappearance)
3. Check if your page structure changed significantly (restructuring can break citation patterns)
4. Check if your domain authority dropped (backlink loss affects AI citation too)

---

## Proactive Triggers

Flag these without being asked:

- **AI bots blocked in robots.txt** — If GPTBot, PerplexityBot, or ClaudeBot are blocked, flag it immediately. Zero AI visibility is possible until fixed, and it's a 5-minute fix. This trumps everything else.
- **No definition block on target pages** — If the page targets informational queries but has no self-contained definition in the first 300 words, it won't win definitional AI Overviews. Flag before doing anything else.
- **Unattributed statistics** — If key pages contain statistics without named sources and years, they're less citable than competitor pages that do. Flag all naked stats.
- **Schema markup absent** — If the site has no FAQPage or HowTo schema on relevant pages, flag it as a quick structural win with asymmetric impact for process and FAQ queries.
- **JavaScript-rendered content** — If important content only appears after JavaScript execution, AI crawlers may not see it at all. Flag content that's hidden behind JS rendering.

---

## Output Artifacts

| When you ask for... | You get... |
|---|---|
| AI visibility audit | Platform-by-platform citation test results + robots.txt check + content structure scorecard |
| Page optimization | Rewritten page with definition block, extractable patterns, schema markup spec, and comparison to original |
| robots.txt fix | Updated robots.txt with correct AI bot allow rules + explanation of what each bot is |
| Schema markup | JSON-LD implementation code for FAQPage, HowTo, or Article — ready to paste |
| Monitoring setup | Weekly tracking template + Google Search Console filter guide + citation log spreadsheet structure |

---

## Communication

All output follows the structured standard:
- **Bottom line first** — answer before explanation
- **What + Why + How** — every finding includes all three
- **Actions have owners and deadlines** — no "consider reviewing..."
- **Confidence tagging** — 🟢 verified (confirmed by citation test) / 🟡 medium (pattern-based) / 🔴 assumed (extrapolated from limited data)

AI SEO is still a young field. Be honest about confidence levels. What gets cited can change as platforms evolve. State what's proven vs. what's pattern-matching.

---

## Related Skills

- **content-production**: Use to create the underlying content before optimizing for AI citation. Good AI SEO requires good content first.
- **content-humanizer**: Use after writing for AI SEO. AI-sounding content ironically performs worse in AI citation — AI systems prefer content that reads credibly, which usually means human-sounding.
- **seo-audit**: Use for traditional search ranking optimization. Run both — AI SEO and traditional SEO are complementary, not competing. Many signals overlap.
- **content-strategy**: Use when deciding which topics and queries to target for AI visibility. Strategy first, then optimize.

写到这儿,回到开头那个问题

LLM 没把 SEO 搅黄。它只是把规则换了。

蓝色链接那条路确实在慢慢窄下去——这是事实,躲不开。但”被 AI 引用”这条新路子,盘子比想象中大。我前面那个 $312K ARR 的客户,他在 AI Overviews 里的曝光,换算成等效点击价值大概是 $0.37-$0.84 一次——不便宜,但相比传统 SEM 的 CPC 已经是白菜价了。

如果你手上也有个站,正在纠结要不要转向 GEO,三条具体起步动作:今天就去看你的 robots.txt,把 GPTBot / PerplexityBot / ClaudeBot / Google-Extended 放进来;挑你最重要的 10 个 query,手动在 Perplexity 跑一遍,看谁被引了、引的什么;选流量最大的 3 个页面,在首屏前 200 字塞一个定义块。

这三件事加起来一个下午能搞完,成本零。剩下的再慢慢铺。我在 CocoLoop 主站 那边写了一些 GEO 实操的更细碎记录——包括我给不同类型客户做监控 dashboard 的那几张 Sheet 模板,有兴趣可以过去翻翻。

  • 标题: LLM 把传统 SEO 搅黄了吗——ai-seo Skill 的三根支柱
  • 作者: Claude 中文知识站
  • 创建于 : 2026-04-18 22:47:00
  • 更新于 : 2026-04-18 22:47:00
  • 链接: https://claude.cocoloop.cn/posts/ai-seo-claude-skill/
  • 版权声明: 本文章采用 CC BY-NC-SA 4.0 进行许可。