技术 SEO 精通:审计、修复与监控 - Openclaw Skills

作者:互联网

2026-03-30

AI教程

什么是 技术 SEO 精通?

技术 SEO 精通是专为开发人员和 SEO 专家设计的高级工具集,旨在确保网站针对搜索引擎进行全面优化。它提供了一种结构化方法来识别和解决阻碍页面排名的各种问题,从抓取预算优化到复杂的国际 SEO 实施。通过使用这些 Openclaw 技能,您可以自动化技术审计过程并保持最佳的网站健康状态。

此技能涵盖了技术 SEO 的整个生命周期,包括核心网页指标(Core Web Vitals)、安全头信息和结构化数据验证。无论您是准备进行网站迁移还是排查索引阻塞问题,它都提供了可操作的方案和评分系统,以有效地确定修复任务的优先级。

下载入口:https://github.com/openclaw/skills/tree/main/skills/1kalin/afrexai-technical-seo

安装与下载

1. ClawHub CLI

从源直接安装技能的最快方式。

npx clawhub@latest install afrexai-technical-seo

2. 手动安装

将技能文件夹复制到以下位置之一

全局模式 ~/.openclaw/skills/ 工作区 /skills/

优先级:工作区 > 本地 > 内置

3. 提示词安装

将此提示词复制到 OpenClaw 即可自动安装。

请帮我使用 Clawhub 安装 afrexai-technical-seo。如果尚未安装 Clawhub,请先安装(npm i -g clawhub)。

技术 SEO 精通 应用场景

  • 执行 5 分钟快速健康检查,对关键网站错误进行分类。
  • 进行全面的抓取和索引审计,解决 robots.txt 或站点地图问题。
  • 诊断并修复 LCP、INP 和 CLS 等核心网页指标,以改善用户体验。
  • 通过详细的发布前和发布后检查清单管理复杂的网站迁移。
  • 实现并验证支持双向 hreflang 的国际化 SEO。
技术 SEO 精通 工作原理
  1. 初始化快速健康检查,识别 robots.txt、站点地图和 HTTPS 状态中的高优先级问题。
  2. 执行深度抓取和索引审计,寻找诸如 noindex 标签或重定向循环等隐藏的阻塞因素。
  3. 根据核心网页指标目标分析性能指标,并识别导致延迟的具体元素。
  4. 验证结构化数据和安全头信息,确保符合富媒体搜索结果资格并保证数据安全。
  5. 使用评分系统生成优先行动计划,首先处理影响最大的技术债务。

技术 SEO 精通 配置指南

技术 SEO 精通已直接集成到您的 AI 代理环境中。要开始网站审计,请提供目标域名并运行初始健康检查命令。

# 开始审计的示例命令
seo-audit "https://example.com"

确保您的代理具有访问 Web 抓取工具的权限,以便分析实时的 robots.txt 和站点地图文件。

技术 SEO 精通 数据架构与分类体系

该技能将数据组织成结构化的 YAML 和 JSON 模式,以便于报告和分析。

数据对象 描述
quick_health 快速分类网站可访问性和基本 SEO 信号。
audit_reports 关于 robots.txt、站点地图和索引能力的详细调查结果。
performance_budget LCP、INP 和总页面权重的量化目标。
migration_map 用于网站重构的 1:1 URL 重定向映射。
scoring_matrix 用于整体技术健康状况的加权评分系统 (0-100)。

Technical SEO Mastery

Complete technical SEO audit, fix, and monitoring system. From crawlability to Core Web Vitals to international SEO — everything search engines need to find, crawl, index, and rank your site.

When to Use

  • "Technical SEO audit for my site"
  • "Why isn't Google indexing my pages?"
  • "Fix Core Web Vitals issues"
  • "Pre-migration SEO checklist"
  • "My site is slow / rankings dropped"
  • "Check robots.txt / sitemap / schema markup"
  • "International SEO / hreflang setup"

Phase 1: Quick Health Check (5-Minute Triage)

Before a full audit, run a rapid assessment. Fetch the site and check:

quick_health:
  url: "[domain]"
  checks:
    robots_txt: "[accessible / blocked / missing]"
    sitemap_xml: "[found / missing / errors]"
    https: "[yes / mixed content / no]"
    mobile_viewport: "[set / missing]"
    page_load: "[fast <2s / moderate 2-4s / slow >4s]"
    h1_present: "[yes / missing / multiple]"
    canonical: "[set / missing / self-referencing]"
    structured_data: "[present / missing / errors]"
  
  severity: "[healthy / needs work / critical]"
  priority_fix: "[top issue to address first]"

Severity guide:

  • ?? Healthy: 0-1 issues — minor optimizations only
  • ?? Needs work: 2-4 issues — schedule fixes this week
  • ?? Critical: 5+ issues or any blocking issue — fix immediately

Phase 2: Crawlability Audit

2.1 Robots.txt Analysis

Fetch [domain]/robots.txt and evaluate:

robots_txt_audit:
  exists: true/false
  valid_syntax: true/false
  issues:
    - type: "[blocked_important_page / missing_sitemap / wildcard_block / syntax_error]"
      detail: "[specific line or pattern]"
      severity: "critical/warning/info"
      fix: "[exact fix]"
  
  checks:
    - "Sitemap directive present"
    - "No accidental blocking of CSS/JS/images"
    - "No blocking of important page directories"
    - "Correct user-agent targeting (Googlebot, Bingbot, etc.)"
    - "No conflicting rules (allow + disallow same path)"
    - "Crawl-delay only if needed (slows indexing)"

Common mistakes:

Mistake Impact Fix
Disallow: / blocking everything No pages indexed Remove or narrow scope
Blocking CSS/JS Poor rendering = ranking drop Allow: /assets/
No sitemap reference Slower discovery Add Sitemap: directive
Multiple sitemaps not declared Partial crawling Declare all sitemaps
Blocking search/filter pages poorly Crawl waste Use Disallow: /*? patterns

Recommended template:

User-agent: *
Allow: /
Disallow: /admin/
Disallow: /api/
Disallow: /checkout/
Disallow: /*?sort=
Disallow: /*?filter=

Sitemap: https://example.com/sitemap.xml
Sitemap: https://example.com/sitemap-images.xml

2.2 XML Sitemap Audit

Fetch sitemap(s) and evaluate:

sitemap_audit:
  url: "[sitemap URL]"
  type: "[index / single / none]"
  url_count: 0
  issues:
    - type: "[missing / too_large / includes_noindex / stale_lastmod / broken_urls]"
      severity: "critical/warning/info"
      fix: "[specific action]"
  
  quality_checks:
    - "All URLs return 200 (no 404s, 301s, or 5xx)"
    - "No noindex pages included"
    - "lastmod dates are accurate (not all same date)"
    - "Under 50,000 URLs per sitemap file"
    - "Under 50MB uncompressed per file"
    - "Sitemap index if >50K URLs"
    - "Matches canonical URLs (not alternate versions)"
    - "Images/video sitemaps if applicable"

Sitemap best practices:

  • One URL per entry — canonical version only
  • lastmod should reflect actual content change date
  • Priority and changefreq are largely ignored by Google — optional
  • Compress with gzip for large sitemaps
  • Submit in Google Search Console AND robots.txt

2.3 Crawl Budget Optimization

crawl_budget_analysis:
  total_pages: 0
  indexable_pages: 0
  crawl_waste_ratio: "[indexable / total — target >80%]"
  
  waste_sources:
    - source: "[faceted navigation / pagination / parameters / duplicate content / thin pages]"
      page_count: 0
      action: "[noindex / canonical / robots block / parameter handling / consolidate]"
  
  optimization_priority:
    1: "Remove/noindex thin and duplicate pages"
    2: "Consolidate parameter variations with canonicals"
    3: "Implement pagination best practices (rel=next or load-more)"
    4: "Fix redirect chains (max 1 hop)"
    5: "Eliminate orphan pages or link them into site structure"

Crawl budget matters when: Site has >10K pages OR crawl rate is notably low in Search Console.


Phase 3: Indexability Audit

3.1 Index Status Check

For each important page type, verify:

indexability_check:
  page_type: "[homepage / product / blog / category / landing]"
  sample_url: "[URL]"
  
  signals:
    meta_robots: "[index,follow / noindex / nofollow / none]"
    x_robots_tag: "[present / absent — check HTTP headers]"
    canonical: "[self / points to other URL / missing]"
    http_status: "[200 / 301 / 302 / 404 / 410 / 5xx]"
    in_sitemap: true/false
    internal_links_to: "[count of internal links pointing here]"
    robots_txt_allowed: true/false
  
  verdict: "[indexable / blocked / conflicting signals]"
  fix: "[action if not indexable]"

3.2 Common Indexing Blockers

Blocker Detection Fix Priority
noindex meta tag Check Remove tag or move to correct pages P0
X-Robots-Tag: noindex header Check HTTP response headers Remove header from server config P0
Canonical pointing elsewhere Check Fix to self-reference or correct target P0
Blocked in robots.txt Cross-reference robots.txt Update robots.txt rules P0
Not in sitemap Check sitemap inclusion Add to sitemap P1
No internal links (orphan) Crawl internal link graph Add contextual internal links P1
Soft 404 (200 with no content) Check page content Return proper 404 or add content P1
Duplicate content Compare page similarity Canonical or consolidate P2
Thin content (<200 words, no value) Word count + quality check Expand or merge with related page P2

3.3 Redirect Audit

redirect_audit:
  chains_found: 0  # A→B→C (should be A→C)
  loops_found: 0   # A→B→A (broken)
  temporary_redirects: 0  # 302s that should be 301s
  
  rules:
    - "Max 1 redirect hop (no chains)"
    - "Use 301 for permanent moves, 308 for POST-preserving"
    - "302 only for genuinely temporary redirects"
    - "Update internal links to point to final destination"
    - "Redirect HTTP → HTTPS at server level"
    - "Redirect www ? non-www consistently"

Phase 4: Core Web Vitals & Performance

4.1 Core Web Vitals Targets

Metric Good Needs Improvement Poor
LCP (Largest Contentful Paint) ≤2.5s 2.5-4.0s >4.0s
INP (Interaction to Next Paint) ≤200ms 200-500ms >500ms
CLS (Cumulative Layout Shift) ≤0.1 0.1-0.25 >0.25

4.2 LCP Fix Playbook

lcp_diagnosis:
  current_lcp: "[Xs]"
  lcp_element: "[identify the largest element — usually hero image or heading]"
  
  fix_priority:
    1:
      issue: "Slow server response (TTFB >800ms)"
      fixes:
        - "Enable server-side caching (Redis, Varnish)"
        - "Use CDN for static assets"
        - "Optimize database queries"
        - "Upgrade hosting if shared"
    2:
      issue: "Render-blocking resources"
      fixes:
        - "Defer non-critical CSS: ``"
        - "Async/defer JavaScript: `