AIsa Provider: 访问中国大语言模型 (通义千问, Kimi, DeepSeek) - Openclaw Skills
作者:互联网
2026-03-30
什么是 AIsa Provider?
AIsa Provider 是 Openclaw Skills 的专用集成插件,将开发者连接到中国最强大的 AI 模型,包括通义千问全系列、DeepSeek V3.1 和月之暗面的 Kimi K2.5。作为通义千问的大客户合作伙伴以及阿里云和字节跳动的官方合作伙伴,AIsa 提供的模型价格显著低于零售渠道。
除了节省成本外,此技能还提供技术优势,如多区域支持(美国、中国、新加坡)和企业级零数据保留 (ZDR) 协议。这确保了通过 Openclaw Skills 处理的敏感数据保持私密,绝不用于模型训练。该提供商使用兼容 OpenAI 的 API,使其成为标准模型后端的无缝替代方案,同时增加了先进的推理和高上下文处理能力。
下载入口:https://github.com/openclaw/skills/tree/main/skills/0xjordansg-yolo/openclaw-aisa-chinese-llm-models
安装与下载
1. ClawHub CLI
从源直接安装技能的最快方式。
npx clawhub@latest install openclaw-aisa-chinese-llm-models
2. 手动安装
将技能文件夹复制到以下位置之一
全局模式~/.openclaw/skills/
工作区
/skills/
优先级:工作区 > 本地 > 内置
3. 提示词安装
将此提示词复制到 OpenClaw 即可自动安装。
请帮我使用 Clawhub 安装 openclaw-aisa-chinese-llm-models。如果尚未安装 Clawhub,请先安装(npm i -g clawhub)。
AIsa Provider 应用场景
- 部署通义千问3 Max 等高性能中国大语言模型,用于复杂的多个步骤推理和编程任务。
- 通过使用专用的 Openclaw Skills 集成,为高 Token 消耗的生产工作负载降低高达 50% 的 API 成本。
- 利用 Kimi K2.5 的零数据保留 (ZDR) 功能,为敏感数据处理实施企业级隐私工作流。
- 通过美国弗吉尼亚或新加坡等特定全球区域路由 AI 请求,以满足合规性或延迟要求。
- 访问高达 256,000 个 Token 的超大上下文窗口,用于大规模文档分析和研究。
- 该技能在环境变量或 OpenClaw 配置文件中识别 AISA_API_KEY。
- OpenClaw 使用 OpenAI-completions API 标准将 AIsa 注册为主要提供商。
- 用户通过 CLI 或聊天界面使用 aisa/ 前缀(例如 aisa/qwen3-max)选择模型。
- 请求通过 AIsa 的高速网关路由到阿里云或月之暗面等官方合作伙伴基础设施。
- 该技能管理特定模型的特殊要求,例如为 Kimi K2.5 推理模型强制设置 1.0 的温度值。
- 流式响应和推理 Token 将返回给智能体,以便立即执行或显示。
AIsa Provider 配置指南
要在您的 Openclaw Skills 环境中激活此提供商,请按照以下步骤操作:
选项 1:快速开始(环境变量)
export AISA_API_KEY="your-key-here"
选项 2:交互式引导
openclaw onboard --auth-choice aisa-api-key
选项 3:手动配置
编辑您的 ~/.openclaw/openclaw.json 以包含 AIsa 模型块:
{
"models": {
"providers": {
"aisa": {
"baseUrl": "https://api.aisa.one/v1",
"apiKey": "${AISA_API_KEY}",
"api": "openai-completions",
"models": [
{
"id": "aisa/qwen3-max",
"name": "Qwen3 Max",
"reasoning": true,
"contextWindow": 256000
}
]
}
}
}
}
AIsa Provider 数据架构与分类体系
该技能利用结构化配置将 AIsa 模型 ID 映射到 OpenClaw 功能。
| 字段 | 描述 | 备注 |
|---|---|---|
| id | 特定的 AIsa 模型标识符 | 必须包含 aisa/ 前缀 |
| reasoning | 推理模型的布尔标记 | 通义千问 Max 和 Kimi K2.5 已启用 |
| contextWindow | 允许的最大 Token 数 | 通义千问为 256k;Kimi/DeepSeek 为 128k |
| supportsDeveloperRole | 系统提示词能力 | 通义千问模型设置为 false |
| cost | 计费对象 | 按每 100 万 Token 衡量 |
name: aisa-provider
description: Configure AIsa as a first-class model provider for OpenClaw, enabling production access to major Chinese AI models (Qwen, DeepSeek, Kimi K2.5, Doubao) through official partnerships with Alibaba Cloud, BytePlus, and Moonshot. Use this skill when the user wants to set up Chinese AI models, configure AIsa API access, compare pricing between AIsa and other providers (OpenRouter, Bailian), switch between Qwen/DeepSeek/Kimi models, or troubleshoot AIsa provider configuration in OpenClaw. Also use when the user mentions AISA_API_KEY, asks about Chinese LLM pricing, Kimi K2.5 setup, or needs help with Qwen Key Account setup.
metadata:
openclaw:
emoji: "????"
requires:
env:
- AISA_API_KEY
primaryEnv: AISA_API_KEY
homepage: "https://marketplace.aisa.one"
AIsa Provider for OpenClaw
AIsa is a unified API gateway providing production access to China's leading AI models through official partnerships with all major Chinese AI platforms. It is an Alibaba Cloud Qwen Key Account partner, offering the full Qwen model family at discounted pricing, plus models on the Alibaba Bailian aggregation platform (DeepSeek, Kimi, GLM).
AIsa also provides access to Kimi K2.5 (Moonshot AI's flagship reasoning model) at approximately 80% of official pricing.
?? All pricing listed below is for reference. Real-time pricing is subject to change — always check https://marketplace.aisa.one/pricing for the latest rates.
Quick Setup
Option 1: Environment Variable (fastest)
export AISA_API_KEY="your-key-here"
OpenClaw auto-detects AISA_API_KEY and registers AIsa as a provider. No config file changes needed.
Option 2: Interactive Onboarding
openclaw onboard --auth-choice aisa-api-key
Option 3: CLI with Key
openclaw onboard --auth-choice aisa-api-key --aisa-api-key "your-key-here"
Option 4: Manual Config in ~/.openclaw/openclaw.json
{
"models": {
"providers": {
"aisa": {
"baseUrl": "https://api.aisa.one/v1",
"apiKey": "${AISA_API_KEY}",
"api": "openai-completions",
"models": [
{
"id": "aisa/qwen3-max",
"name": "Qwen3 Max",
"reasoning": true,
"input": ["text", "image"],
"contextWindow": 256000,
"maxTokens": 16384,
"supportsDeveloperRole": false,
"cost": {
"input": 1.20,
"output": 4.80,
"cacheRead": 0,
"cacheWrite": 0
}
},
{
"id": "aisa/qwen-plus-2025-12-01",
"name": "Qwen Plus",
"reasoning": true,
"input": ["text", "image"],
"contextWindow": 256000,
"maxTokens": 16384,
"supportsDeveloperRole": false,
"cost": {
"input": 0.30,
"output": 0.90,
"cacheRead": 0,
"cacheWrite": 0
}
},
{
"id": "aisa/qwen-mt-flash",
"name": "Qwen MT Flash",
"reasoning": true,
"input": ["text"],
"contextWindow": 256000,
"maxTokens": 8192,
"supportsDeveloperRole": false,
"cost": {
"input": 0.05,
"output": 0.30,
"cacheRead": 0,
"cacheWrite": 0
}
},
{
"id": "aisa/deepseek-v3.1",
"name": "DeepSeek V3.1",
"reasoning": true,
"input": ["text"],
"contextWindow": 131072,
"maxTokens": 8192,
"supportsDeveloperRole": false,
"cost": {
"input": 0.27,
"output": 1.10,
"cacheRead": 0.07,
"cacheWrite": 0
}
},
{
"id": "aisa/kimi-k2.5",
"name": "Kimi K2.5",
"reasoning": true,
"input": ["text"],
"contextWindow": 131072,
"maxTokens": 8192,
"supportsDeveloperRole": false,
"cost": {
"input": 0.60,
"output": 2.40,
"cacheRead": 0,
"cacheWrite": 0
}
}
]
}
}
},
"agents": {
"defaults": {
"model": {
"primary": "aisa/qwen3-max"
}
}
}
}
Available Models
Default Models (pre-configured, API-verified ?)
| Model | Model ID | Best For | Context | Reasoning | Verified |
|---|---|---|---|---|---|
| Qwen3 Max | aisa/qwen3-max |
Complex reasoning, flagship tasks | 256K | ? | ? |
| Qwen Plus | aisa/qwen-plus-2025-12-01 |
Main production model | 256K | ? | ? |
| Qwen MT Flash | aisa/qwen-mt-flash |
High-frequency, lightweight tasks | 256K | ? | ? |
| DeepSeek V3.1 | aisa/deepseek-v3.1 |
Cost-effective reasoning | 128K | ? | ? |
| Kimi K2.5 | aisa/kimi-k2.5 |
Moonshot's flagship reasoning model | 128K | ? | ? |
? Kimi K2.5 — Moonshot AI's Flagship Model
Kimi K2.5 is Moonshot AI's latest reasoning model, accessible via AIsa at approximately 80% of official Moonshot pricing.
Key features:
- Strong reasoning and multi-step problem solving
- Competitive performance on coding and math benchmarks
- Available through AIsa's Moonshot partnership
?? Zero Data Retention (ZDR) — Enterprise-Grade Privacy
Accessing Kimi K2.5 through AIsa comes with enterprise-level Zero Data Retention guarantees. Users do not need to worry about data privacy — AIsa has executed a formal ZDR agreement with Moonshot AI.
Under the Supplemental Enterprise Service Agreement (dated 10 February 2026) between AIsa and Kimi (Moonshot AI PTE. LTD.):
- Customer data is NOT retained by Moonshot after processing
- Generated outputs are NOT stored on Moonshot's infrastructure
- No data is used for model training — your prompts and completions remain private
- Processing is governed by contractual enterprise terms, not consumer ToS
This makes AIsa the recommended path for privacy-sensitive or enterprise workloads that require access to Kimi K2.5. When calling Kimi K2.5 directly via Moonshot's consumer API, standard consumer data policies apply — by routing through AIsa, your data benefits from the negotiated ZDR protections.
?? Important: Temperature restriction
Kimi K2.5 only accepts temperature=1.0. Using any other value will return an error:
Error: invalid temperature: only 1 is allowed for this model
If your OpenClaw config or agent sets a different temperature, override it for Kimi:
/model aisa/kimi-k2.5
OpenClaw will use the model's default temperature when not explicitly set.
Kimi K2.5 Pricing Comparison (per 1M tokens):
| Metric | AIsa | Moonshot Official | Savings |
|---|---|---|---|
| Input/1M | ~$0.60 | ~$0.75 | ~20% off |
| Output/1M | ~$2.40 | ~$3.00 | ~20% off |
Actual pricing may vary. Check https://marketplace.aisa.one/pricing for real-time rates.
Additional Models Available via AIsa
Users can add any model supported by AIsa to their config. The full catalog includes 49+ models:
Qwen family (8 models):
qwen3-max,qwen3-max-2026-01-23,qwen-plus-2025-12-01qwen-mt-flash,qwen-mt-liteqwen-vl-max,qwen3-vl-flash,qwen3-vl-plus(vision models)
DeepSeek (4 models):
deepseek-v3.1,deepseek-v3,deepseek-v3-0324,deepseek-r1
Kimi / Moonshot (2 models):
kimi-k2.5,kimi-k2-thinking
Also available: Claude series (10), GPT series (9), Gemini series (5), Grok series (2), and more.
List all available models:
curl https://api.aisa.one/v1/models -H "Authorization: Bearer $AISA_API_KEY"
Model ID Versioning
AIsa uses versioned model IDs for some models. If you encounter a 503 - No available channels error, the model ID may need updating.
Known model ID mappings:
| Common Name | Correct AIsa Model ID | ? Does NOT work |
|---|---|---|
| Qwen Plus | qwen-plus-2025-12-01 |
qwen3-plus, qwen-plus, qwen-plus-latest |
| Qwen Flash | qwen-mt-flash |
qwen3-flash, qwen-turbo, qwen-turbo-latest |
| Qwen Max | qwen3-max |
(works as-is) |
| DeepSeek V3.1 | deepseek-v3.1 |
(works as-is) |
| Kimi K2.5 | kimi-k2.5 |
(works as-is) |
To check the latest available model IDs:
curl https://api.aisa.one/v1/models -H "Authorization: Bearer $AISA_API_KEY"
Switching Models
In chat (TUI):
/model aisa/qwen3-max
/model aisa/deepseek-v3.1
/model aisa/kimi-k2.5
Via CLI:
openclaw models set aisa/qwen3-max
Pricing Comparison (per 1M tokens)
All pricing below is for reference. Real-time pricing is subject to change — always check https://marketplace.aisa.one/pricing for the latest rates.
Qwen MT Flash (lightweight)
- AIsa: $0.05 input / $0.30 output (~50% off retail)
- Bailian Official: $0.10 / $0.40
- OpenRouter: $0.11-0.13 / $0.45-0.50
Qwen Plus (production)
- AIsa: $0.30 input / $0.90 output (~25% off retail)
- Bailian Official: $0.40 / $1.20
- OpenRouter: $0.45-0.50 / $1.35-1.50
Qwen3 Max (flagship)
- AIsa: $1.20 input / $4.80 output (~40% off retail)
- Bailian Official: $2.00 / $8.00
- OpenRouter: $2.20-2.50 / $9.00-10.00
Kimi K2.5 (Moonshot flagship)
- AIsa:
$0.60 input / ~$2.40 output (**20% off** official Moonshot pricing) - Moonshot Official: ~$0.75 / ~$3.00
- OpenRouter: Limited availability
Cost at scale: 500M tokens/month on Qwen-Max
- OpenRouter: ~$4,000-4,250/month
- Bailian Official: ~$3,400/month
- AIsa: ~$2,040/month (saves $16,320-26,520/year)
Official Partnerships
AIsa maintains verified partnerships with:
- Alibaba Cloud — Qwen Key Account (full model family, 3 global regions: CN, US-Virginia, Singapore)
- BytePlus — Doubao by ByteDance
- DeepSeek — via Alibaba Cloud integration
- Moonshot — Kimi K2.5 integration, with enterprise Zero Data Retention (ZDR) agreement (effective Feb 10, 2026)
Qwen Region Support
AIsa provides access to Qwen models across 3 global regions via Alibaba Cloud:
- ???? China (default)
- ???? US (Virginia)
- ???? Singapore
This is unique to AIsa's Key Account status. Other providers like OpenRouter or the free Qwen Portal typically route through CN only.
Response Latency (tested Feb 2026)
| Model | Avg Latency | Rating |
|---|---|---|
| Qwen3 Max | ~1,577 ms | ????? Fastest |
| Qwen MT Flash | ~1,918 ms | ???? Fast |
| Kimi K2.5 | ~2,647 ms | ??? Medium |
| DeepSeek V3.1 | ~3,002 ms | ??? Medium |
| Qwen Plus | ~8,207 ms | ?? Slower |
Troubleshooting
"503 - No available channels" error
The model ID may be incorrect or outdated. Check the Model ID Versioning section above for correct IDs. Common fixes:
qwen3-plus→ useqwen-plus-2025-12-01qwen3-flash→ useqwen-mt-flash
"Model not found" error
Ensure the model ID uses the aisa/ prefix in OpenClaw config:
? aisa/qwen3-max
? qwen3-max
Kimi K2.5 "invalid temperature" error
Kimi K2.5 only accepts temperature=1.0. If your config sets a different temperature, add a model-specific override or let OpenClaw use the default.
Kimi K2.5 empty response
In rare cases Kimi K2.5 may return empty content while consuming output tokens. Retry the request — this is typically transient.
API key not detected
- Check env var:
echo $AISA_API_KEY - Or verify in config:
openclaw config get auth.profiles - Re-run onboarding:
openclaw onboard --auth-choice aisa-api-key
Streaming not working
AIsa uses the OpenAI-compatible API (openai-completions). Ensure your config has:
"api": "openai-completions"
Rate limits or daily caps
AIsa has no daily request limits (unlike the free Qwen Portal which caps at 2,000 req/day).
Get an API Key
- Visit https://marketplace.aisa.one/
- Sign up and create an API key
- Set it as
AISA_API_KEYor use the onboarding wizard
Notes
- AIsa's endpoint is OpenAI-compatible (
https://api.aisa.one/v1) - All models support streaming and function calling
supportsDeveloperRoleis set tofalsefor Qwen models- Default context window: 256,000 tokens (Qwen) or 131,072 tokens (DeepSeek/Kimi)
- Reasoning (thinking) is enabled for all default models
- Kimi K2.5 requires
temperature=1.0— other values cause API errors - Kimi K2.5 via AIsa is covered by enterprise Zero Data Retention (ZDR) — data is not retained or used for training
- Image/Video generation models (WAN) are available but require separate configuration
- AIsa API supports 49+ models total — use the models endpoint to discover all available options
相关推荐
专题
+ 收藏
+ 收藏
+ 收藏
+ 收藏
+ 收藏
最新数据
相关文章
每日语音名言:AI 驱动的励志与数字人视频 - Openclaw Skills
Zoho Bookings API 集成:管理预约与员工 - Openclaw Skills
工程经理操作系统:技术领导力与团队扩展 - Openclaw Skills
Markdown 报告生成器:专业文档制作 - Openclaw Skills
Business Claw:ERPNext MCP 自动化与工作流 - Openclaw 技能
Context Engineer:Openclaw 技能的 Token 优化专家
韩国发票:自动化估价单与税务发票 - Openclaw Skills
小红书文案教练:爆款笔记生成器 - Openclaw Skills
慕尼黑 MVG & S-Bahn 实时追踪命令行工具 - Openclaw Skills
Reddit 研究技能:自动化社群洞察 - Openclaw Skills
AI精选
