技术 SEO 精通:审计、修复与监控 - Openclaw Skills
作者:互联网
2026-03-30
什么是 技术 SEO 精通?
技术 SEO 精通是专为开发人员和 SEO 专家设计的高级工具集,旨在确保网站针对搜索引擎进行全面优化。它提供了一种结构化方法来识别和解决阻碍页面排名的各种问题,从抓取预算优化到复杂的国际 SEO 实施。通过使用这些 Openclaw 技能,您可以自动化技术审计过程并保持最佳的网站健康状态。
此技能涵盖了技术 SEO 的整个生命周期,包括核心网页指标(Core Web Vitals)、安全头信息和结构化数据验证。无论您是准备进行网站迁移还是排查索引阻塞问题,它都提供了可操作的方案和评分系统,以有效地确定修复任务的优先级。
下载入口:https://github.com/openclaw/skills/tree/main/skills/1kalin/afrexai-technical-seo
安装与下载
1. ClawHub CLI
从源直接安装技能的最快方式。
npx clawhub@latest install afrexai-technical-seo
2. 手动安装
将技能文件夹复制到以下位置之一
全局模式~/.openclaw/skills/
工作区
/skills/
优先级:工作区 > 本地 > 内置
3. 提示词安装
将此提示词复制到 OpenClaw 即可自动安装。
请帮我使用 Clawhub 安装 afrexai-technical-seo。如果尚未安装 Clawhub,请先安装(npm i -g clawhub)。
技术 SEO 精通 应用场景
- 执行 5 分钟快速健康检查,对关键网站错误进行分类。
- 进行全面的抓取和索引审计,解决 robots.txt 或站点地图问题。
- 诊断并修复 LCP、INP 和 CLS 等核心网页指标,以改善用户体验。
- 通过详细的发布前和发布后检查清单管理复杂的网站迁移。
- 实现并验证支持双向 hreflang 的国际化 SEO。
- 初始化快速健康检查,识别 robots.txt、站点地图和 HTTPS 状态中的高优先级问题。
- 执行深度抓取和索引审计,寻找诸如 noindex 标签或重定向循环等隐藏的阻塞因素。
- 根据核心网页指标目标分析性能指标,并识别导致延迟的具体元素。
- 验证结构化数据和安全头信息,确保符合富媒体搜索结果资格并保证数据安全。
- 使用评分系统生成优先行动计划,首先处理影响最大的技术债务。
技术 SEO 精通 配置指南
技术 SEO 精通已直接集成到您的 AI 代理环境中。要开始网站审计,请提供目标域名并运行初始健康检查命令。
# 开始审计的示例命令
seo-audit "https://example.com"
确保您的代理具有访问 Web 抓取工具的权限,以便分析实时的 robots.txt 和站点地图文件。
技术 SEO 精通 数据架构与分类体系
该技能将数据组织成结构化的 YAML 和 JSON 模式,以便于报告和分析。
| 数据对象 | 描述 |
|---|---|
| quick_health | 快速分类网站可访问性和基本 SEO 信号。 |
| audit_reports | 关于 robots.txt、站点地图和索引能力的详细调查结果。 |
| performance_budget | LCP、INP 和总页面权重的量化目标。 |
| migration_map | 用于网站重构的 1:1 URL 重定向映射。 |
| scoring_matrix | 用于整体技术健康状况的加权评分系统 (0-100)。 |
Technical SEO Mastery
Complete technical SEO audit, fix, and monitoring system. From crawlability to Core Web Vitals to international SEO — everything search engines need to find, crawl, index, and rank your site.
When to Use
- "Technical SEO audit for my site"
- "Why isn't Google indexing my pages?"
- "Fix Core Web Vitals issues"
- "Pre-migration SEO checklist"
- "My site is slow / rankings dropped"
- "Check robots.txt / sitemap / schema markup"
- "International SEO / hreflang setup"
Phase 1: Quick Health Check (5-Minute Triage)
Before a full audit, run a rapid assessment. Fetch the site and check:
quick_health:
url: "[domain]"
checks:
robots_txt: "[accessible / blocked / missing]"
sitemap_xml: "[found / missing / errors]"
https: "[yes / mixed content / no]"
mobile_viewport: "[set / missing]"
page_load: "[fast <2s / moderate 2-4s / slow >4s]"
h1_present: "[yes / missing / multiple]"
canonical: "[set / missing / self-referencing]"
structured_data: "[present / missing / errors]"
severity: "[healthy / needs work / critical]"
priority_fix: "[top issue to address first]"
Severity guide:
- ?? Healthy: 0-1 issues — minor optimizations only
- ?? Needs work: 2-4 issues — schedule fixes this week
- ?? Critical: 5+ issues or any blocking issue — fix immediately
Phase 2: Crawlability Audit
2.1 Robots.txt Analysis
Fetch [domain]/robots.txt and evaluate:
robots_txt_audit:
exists: true/false
valid_syntax: true/false
issues:
- type: "[blocked_important_page / missing_sitemap / wildcard_block / syntax_error]"
detail: "[specific line or pattern]"
severity: "critical/warning/info"
fix: "[exact fix]"
checks:
- "Sitemap directive present"
- "No accidental blocking of CSS/JS/images"
- "No blocking of important page directories"
- "Correct user-agent targeting (Googlebot, Bingbot, etc.)"
- "No conflicting rules (allow + disallow same path)"
- "Crawl-delay only if needed (slows indexing)"
Common mistakes:
| Mistake | Impact | Fix |
|---|---|---|
Disallow: / blocking everything |
No pages indexed | Remove or narrow scope |
| Blocking CSS/JS | Poor rendering = ranking drop | Allow: /assets/ |
| No sitemap reference | Slower discovery | Add Sitemap: directive |
| Multiple sitemaps not declared | Partial crawling | Declare all sitemaps |
| Blocking search/filter pages poorly | Crawl waste | Use Disallow: /*? patterns |
Recommended template:
User-agent: *
Allow: /
Disallow: /admin/
Disallow: /api/
Disallow: /checkout/
Disallow: /*?sort=
Disallow: /*?filter=
Sitemap: https://example.com/sitemap.xml
Sitemap: https://example.com/sitemap-images.xml
2.2 XML Sitemap Audit
Fetch sitemap(s) and evaluate:
sitemap_audit:
url: "[sitemap URL]"
type: "[index / single / none]"
url_count: 0
issues:
- type: "[missing / too_large / includes_noindex / stale_lastmod / broken_urls]"
severity: "critical/warning/info"
fix: "[specific action]"
quality_checks:
- "All URLs return 200 (no 404s, 301s, or 5xx)"
- "No noindex pages included"
- "lastmod dates are accurate (not all same date)"
- "Under 50,000 URLs per sitemap file"
- "Under 50MB uncompressed per file"
- "Sitemap index if >50K URLs"
- "Matches canonical URLs (not alternate versions)"
- "Images/video sitemaps if applicable"
Sitemap best practices:
- One URL per
entry — canonical version only lastmodshould reflect actual content change date- Priority and changefreq are largely ignored by Google — optional
- Compress with gzip for large sitemaps
- Submit in Google Search Console AND robots.txt
2.3 Crawl Budget Optimization
crawl_budget_analysis:
total_pages: 0
indexable_pages: 0
crawl_waste_ratio: "[indexable / total — target >80%]"
waste_sources:
- source: "[faceted navigation / pagination / parameters / duplicate content / thin pages]"
page_count: 0
action: "[noindex / canonical / robots block / parameter handling / consolidate]"
optimization_priority:
1: "Remove/noindex thin and duplicate pages"
2: "Consolidate parameter variations with canonicals"
3: "Implement pagination best practices (rel=next or load-more)"
4: "Fix redirect chains (max 1 hop)"
5: "Eliminate orphan pages or link them into site structure"
Crawl budget matters when: Site has >10K pages OR crawl rate is notably low in Search Console.
Phase 3: Indexability Audit
3.1 Index Status Check
For each important page type, verify:
indexability_check:
page_type: "[homepage / product / blog / category / landing]"
sample_url: "[URL]"
signals:
meta_robots: "[index,follow / noindex / nofollow / none]"
x_robots_tag: "[present / absent — check HTTP headers]"
canonical: "[self / points to other URL / missing]"
http_status: "[200 / 301 / 302 / 404 / 410 / 5xx]"
in_sitemap: true/false
internal_links_to: "[count of internal links pointing here]"
robots_txt_allowed: true/false
verdict: "[indexable / blocked / conflicting signals]"
fix: "[action if not indexable]"
3.2 Common Indexing Blockers
| Blocker | Detection | Fix | Priority |
|---|---|---|---|
noindex meta tag |
Check |
Remove tag or move to correct pages | P0 |
X-Robots-Tag: noindex header |
Check HTTP response headers | Remove header from server config | P0 |
| Canonical pointing elsewhere | Check |
Fix to self-reference or correct target | P0 |
| Blocked in robots.txt | Cross-reference robots.txt | Update robots.txt rules | P0 |
| Not in sitemap | Check sitemap inclusion | Add to sitemap | P1 |
| No internal links (orphan) | Crawl internal link graph | Add contextual internal links | P1 |
| Soft 404 (200 with no content) | Check page content | Return proper 404 or add content | P1 |
| Duplicate content | Compare page similarity | Canonical or consolidate | P2 |
| Thin content (<200 words, no value) | Word count + quality check | Expand or merge with related page | P2 |
3.3 Redirect Audit
redirect_audit:
chains_found: 0 # A→B→C (should be A→C)
loops_found: 0 # A→B→A (broken)
temporary_redirects: 0 # 302s that should be 301s
rules:
- "Max 1 redirect hop (no chains)"
- "Use 301 for permanent moves, 308 for POST-preserving"
- "302 only for genuinely temporary redirects"
- "Update internal links to point to final destination"
- "Redirect HTTP → HTTPS at server level"
- "Redirect www ? non-www consistently"
Phase 4: Core Web Vitals & Performance
4.1 Core Web Vitals Targets
| Metric | Good | Needs Improvement | Poor |
|---|---|---|---|
| LCP (Largest Contentful Paint) | ≤2.5s | 2.5-4.0s | >4.0s |
| INP (Interaction to Next Paint) | ≤200ms | 200-500ms | >500ms |
| CLS (Cumulative Layout Shift) | ≤0.1 | 0.1-0.25 | >0.25 |
4.2 LCP Fix Playbook
lcp_diagnosis:
current_lcp: "[Xs]"
lcp_element: "[identify the largest element — usually hero image or heading]"
fix_priority:
1:
issue: "Slow server response (TTFB >800ms)"
fixes:
- "Enable server-side caching (Redis, Varnish)"
- "Use CDN for static assets"
- "Optimize database queries"
- "Upgrade hosting if shared"
2:
issue: "Render-blocking resources"
fixes:
- "Defer non-critical CSS: ``"
- "Async/defer JavaScript: `
