model-fallback
Multi-model automatic fallback system. Monitors model availability and automatically falls back to backup models when the primary model fails. Supports MiniMax, Kimi, Zhipu and other OpenAI-compatible APIs. Use when: (1) Primary model API is unavailable, (2) Model response time is too slow, (3) Rate limit exceeded, (4) Need to optimize costs by using cheaper models for simple tasks.
Why use this skill?
Ensure continuous AI agent operation with OpenClaw Model Fallback. Automatically switch between MiniMax, Kimi, and Zhipu APIs to prevent downtime.
Install via CLI (Recommended)
clawhub install openclaw/skills/skills/azure5100/model-fallbackWhat This Skill Does
The model-fallback skill serves as a robust safety net for OpenClaw agents by automating the selection and failover process for Large Language Models. In a production environment, reliance on a single model provider is risky due to potential downtime, sudden API rate limits, or latency spikes. This skill mitigates these issues by maintaining a user-defined fallback chain. When the primary model fails to respond—whether due to a 503 error, a timeout, or a rate limit—the agent intelligently transitions to a secondary or tertiary model in the configuration. It supports seamless integration with providers like MiniMax, Kimi, and Zhipu, ensuring your AI workflows remain uninterrupted and high-performing even when individual services falter.
Installation
To integrate the model-fallback skill into your OpenClaw environment, execute the following command in your terminal: clawhub install openclaw/skills/skills/azure5100/model-fallback. Once installed, navigate to the config.json file in your agent's root directory to define your fallback_chain priority. Ensure you have the necessary API keys configured in your environment variables for each provider included in your chain, as the skill will reference these keys during the automatic switching process. After configuration, you can verify that the service is running by executing /scripts/model-fallback.sh --status to inspect the connectivity of your listed models.
Use Cases
This skill is ideal for mission-critical automation tasks where high availability is non-negotiable. Use it for: (1) High-traffic customer support agents that cannot afford downtime; (2) Complex research agents that require large context windows; (3) Cost-sensitive projects where you prefer to use cheaper models for mundane tasks and reserve flagship models for complex reasoning; and (4) Robust data extraction pipelines that need to handle occasional API instability gracefully without manual intervention.
Example Prompts
- "Analyze these log files and summarize the errors using the primary model, or switch to the fallback if the primary is unresponsive."
- "Execute the documentation generation task; if the current model hits a rate limit, please continue using the next available model in the priority list."
- "Summarize this 100k-word document. Use the high-capacity model first, but switch to the cost-efficient model if the primary returns a 5xx error."
Tips & Limitations
To maximize the utility of this skill, prioritize your models based on both performance and cost. Place your most capable, reasoning-heavy models first, and keep faster, cheaper models as final backups. Note that switching models mid-stream may lead to slight variations in tone or style since different model architectures interpret system instructions uniquely. Always test your fallback chain by temporarily disabling your primary API to ensure the agent correctly routes tasks to your designated backups. Be aware that the skill is currently limited to OpenAI-compatible APIs; custom logic may be required for proprietary provider architectures.
Metadata
Not sure this is the right skill?
Describe what you want to build — we'll match you to the best skill from 16,000+ options.
Find the right skillPaste this into your clawhub.json to enable this plugin.
{
"plugins": {
"official-azure5100-model-fallback": {
"enabled": true,
"auto_update": true
}
}
}Tags
Flags: external-api
Related Skills
openclaw-health-audit
OpenClaw 系统健康审计与自动修复 — 监控 prompt 体积、Cron 合规、Session 清理、Token 消耗
astrai-inference-router
Route all LLM calls through Astrai for 40%+ cost savings with intelligent routing and privacy controls
afrexai-observability-engine
Complete observability & reliability engineering system. Use when designing monitoring, implementing structured logging, setting up distributed tracing, building alerting systems, creating SLO/SLI frameworks, running incident response, conducting post-mortems, or auditing system reliability. Covers all three pillars (logs/metrics/traces), alert design, dashboard architecture, on-call operations, chaos engineering, and cost optimization.
semantic-router
让 AI 代理根据对话内容自动选择最合适的模型。四层识别(系统过滤→关键词→指示词→语义相似度),四池架构(高速/智能/人文/代理),五分支路由,全自动 Fallback 回路。支持 trigger_groups_all 非连续词组命中。
astrai-code-review
AI-powered code review with intelligent model routing — saves 40%+ vs always using the most expensive model