cc-compatible-models

skill
Guvenlik Denetimi
Uyari
Health Uyari
  • No license — Repository has no license file
  • Description — Repository has a description
  • Active repo — Last push 0 days ago
  • Community trust — 23 GitHub stars
Code Uyari
  • Code scan incomplete — No supported source files were scanned during light audit
Permissions Gecti
  • Permissions — No dangerous permissions requested
Purpose: This is a reference guide that helps developers configure Claude Code to work with various alternative AI providers (like DeepSeek, Qwen, and OpenRouter). It provides setup instructions, pricing tiers, and configuration snippets.

Security Assessment: Overall risk is Low. The project is a documentation repository rather than executable code, which means it does not run shell commands or access local files. The automated code scan found no hardcoded secrets or dangerous permissions. However, it instructs users to manually paste their own API keys into configuration files (like `~/.claude/settings.json`). While standard practice for CLI tools, developers should always ensure these configuration files are secured and kept out of public version control.

Quality Assessment: The project is actively maintained, with its most recent push happening today. It has earned 23 GitHub stars, indicating a modest level of community trust and usefulness. The primary downside is the complete lack of an open-source license. This means that technically, all rights are reserved by the author, and the legal terms for using or distributing the content are undefined.

Verdict: Safe to use as a helpful configuration reference, but be aware that the guide lacks a formal software license.
SUMMARY

Complete guide for using alternative AI models with Claude Code — including DeepSeek, Qwen, MiniMax, Kimi, GLM, MiMo, StepFun, and more. Pricing, configs, and coding plans.

README.md

Claude Code: Compatible Models & Providers (2025-2026 Edition)

A comprehensive reference guide for developers using Claude Code with official and alternative AI providers. This guide covers the latest flagship models, pricing (Input/Output per 1M tokens), configuration for Anthropic, Alibaba Qwen, DeepSeek, MiniMax, Moonshot AI (Kimi), Zhipu GLM, Xiaomi MiMo, StepFun, and multi-provider gateways.

Compatible Providers

Provider Type Notable Models Coding Plan
Anthropic Official Claude Opus 4.7, Sonnet 4.6 Max subscription
Alibaba (DashScope) Direct Qwen 3.6-Plus, Qwen 3.5-Coder
DeepSeek Direct DeepSeek V3.2, DeepSeek-R2
MiniMax Direct MiniMax-M2.7, M2.7-Highspeed Token Plan (from $10/mo)
Moonshot (Kimi) Direct Kimi K2.5, K2.6 (beta)
Zhipu (Z.ai) Direct GLM-5.1, GLM-5, GLM-4.7 Coding Plan (from $18/mo)
Xiaomi (MiMo) Direct MiMo-V2-Pro, V2-Omni, V2-Flash Token Plan (from ~$6/mo)
StepFun Direct Step-3.5-Flash, Step-3 Step Plan (from $6.99/mo)
SiliconFlow Gateway Hosted DeepSeek, GLM, Qwen, Kimi
OpenRouter Gateway 200+ models from all providers
Ollama Local / Cloud Open-source models, free Free (local)

Provider Setup Guides

Anthropic (Official)

Claude Code officially supports Claude Opus 4.7 and Sonnet 4.6, excelling at agentic coding tasks and tool use.

Installation:

npm install -g @anthropic-ai/claude-code

Authentication:

  • OAuth: Run claude and follow the browser login.
  • API Key: Set ANTHROPIC_API_KEY in your environment.

Official Documentation


Alibaba Cloud - DashScope (Qwen)

Qwen 3.6-Plus is the latest flagship (April 2026), with 1M context window and strong agentic coding capabilities.

Configuration:
Edit ~/.claude/settings.json:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://dashscope-intl.aliyuncs.com/apps/anthropic",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_DASHSCOPE_API_KEY",
    "ANTHROPIC_MODEL": "qwen3.6-plus"
  }
}

Pricing tiers (per 1M tokens):

Context Band Input Output
0 - 256K $0.28 $1.65
256K - 1M $1.10 $6.60

DeepSeek

DeepSeek V3.2 and R2 offer incredibly cost-effective coding and reasoning.

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.deepseek.com/anthropic",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_DEEPSEEK_API_KEY",
    "ANTHROPIC_MODEL": "deepseek-reasoner"
  }
}

Pricing (per 1M tokens):

Model Input Cache Hit Output
DeepSeek V3.2 $0.27 $0.07 $1.10
DeepSeek-R2 $0.55 $0.14 $2.19

Note: deepseek-reasoner maps to the reasoning model with thinking processes.


MiniMax

MiniMax-M2.7 (March 2026) is open-source, 230B params MoE, with SOTA on real-world software engineering benchmarks.

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.minimax.io/anthropic",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_MINIMAX_API_KEY",
    "ANTHROPIC_MODEL": "minimax-m2.7"
  }
}

China endpoint: https://api.minimaxi.com/anthropic

Pricing (per 1M tokens):

Model Input Cache Read Output
MiniMax-M2.7 $0.30 $0.059 $1.20
M2.7-Highspeed $0.60 $2.40

Moonshot AI - Kimi

Kimi K2.5 is a 1T-parameter MoE model optimized for long context. K2.6 is in beta (April 2026).

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.moonshot.ai/anthropic",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_MOONSHOT_API_KEY",
    "ANTHROPIC_MODEL": "kimi-k2.5"
  }
}

Note: K2.6 is currently beta-only via Kimi CLI/IDE integrations. Public API expected May 2026.


Zhipu (Z.ai) - GLM

GLM-5.1 (April 2026) is a 744B MoE model capable of 8+ hour autonomous tasks, MIT-licensed.

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.z.ai/api/anthropic",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_ZAI_API_KEY",
    "ANTHROPIC_MODEL": "glm-5.1"
  }
}

Pricing (per 1M tokens):

Model Input Output
GLM-5.1 $0.95 - $1.40 $3.15 - $4.40
GLM-5 $1.00 $3.20
GLM-4.7 $0.60 $2.20

Note: GLM-5.1 consumes 3x quota during peak hours, 2x off-peak on Coding Plan.


Xiaomi - MiMo

MiMo-V2-Pro ranks top-5 globally on OpenRouter Arena. 1M context, MoE architecture.

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.xiaomimimo.com",
    "ANTHROPIC_AUTH_TOKEN": "YOUR_MIMO_API_KEY",
    "ANTHROPIC_MODEL": "mimo-v2-pro"
  }
}

Pricing (per 1M tokens):

Model Context Input Output Cache Read
MiMo-V2-Pro ≤256K $1.00 $3.00 $0.20
MiMo-V2-Pro 256K-1M $2.00 $6.00 $0.40
MiMo-V2-Omni 256K $0.40 $2.00 $0.08
MiMo-V2-Flash 256K $0.09 $0.29 $0.01

Supported integrations: Claude Code, OpenCode, OpenClaw, Cline, Kilo Code, Roo Code, Codex, Cherry Studio, Zed.


StepFun

Step-3.5-Flash is a powerful reasoning model at the lowest price point among major providers.

Configuration (Pay-as-you-go):

{
  "env": {
    "ANTHROPIC_AUTH_TOKEN": "YOUR_STEPFUN_API_KEY",
    "ANTHROPIC_BASE_URL": "https://api.stepfun.ai/",
    "ANTHROPIC_MODEL": "step-3.5-flash"
  }
}

Configuration (Step Plan subscription):

{
  "env": {
    "ANTHROPIC_API_KEY": "YOUR_STEPFUN_API_KEY",
    "ANTHROPIC_BASE_URL": "https://api.stepfun.ai/step_plan",
    "ANTHROPIC_MODEL": "step-3.5-flash"
  }
}

China endpoints: Replace .ai with .com (e.g., https://api.stepfun.com/).

Pricing (per 1M tokens):

Model Input Output Cache Read
Step-3.5-Flash $0.10 $0.30 $0.02
Step-3 ~$0.21 ~$0.55 - $1.10 ~$0.04
Step-2-Mini ~$0.14 ~$0.28 ~$0.03

Supported integrations: Claude Code, OpenCode, OpenClaw, Cline, Kilo Code, Roo Code, Trae, Cursor, Zed, Cherry Studio, Goose.


SiliconFlow (Multi-Provider)

Hosted access to DeepSeek, GLM, Qwen, Kimi, and more via Anthropic-compatible API.

Configuration:

{
  "env": {
    "ANTHROPIC_BASE_URL": "https://api.siliconflow.com/",
    "ANTHROPIC_API_KEY": "YOUR_SILICONFLOW_KEY",
    "ANTHROPIC_MODEL": "deepseek-ai/DeepSeek-V3"
  }
}

Note: Currently does not support "thinking" model variants — only non-thinking.


Ollama (Local / Cloud)

Run models locally for free since January 2026 (v0.14.0+). Also supports cloud models via ollama.com.

Configuration:

{
  "env": {
    "ANTHROPIC_AUTH_TOKEN": "ollama",
    "ANTHROPIC_BASE_URL": "http://localhost:11434",
    "ANTHROPIC_API_KEY": ""
  }
}

Quick start: ollama launch claude --config provides guided setup. Requires models with 32K+ context.

Local models: qwen3-coder, glm-4.7, minimax-m2.1, and more.
Cloud models: glm-5.1:cloud, kimi-k2.5:cloud, qwen3.5:cloud via ollama.com.


OpenRouter (Multi-Provider)

Best for comparing models or failover. Supports all providers via one token, including Fast Mode (2.5x faster output at premium pricing).

Basic Setup:

export ANTHROPIC_BASE_URL="https://openrouter.ai/api"
export ANTHROPIC_AUTH_TOKEN="YOUR_OPENROUTER_KEY"

Model environment variables:

export ANTHROPIC_DEFAULT_OPUS_MODEL="anthropic/claude-opus-4.7"
export ANTHROPIC_DEFAULT_SONNET_MODEL="anthropic/claude-sonnet-4.6"

Coding Plans & Subscriptions

Several providers offer subscription plans with flat-rate usage for Claude Code, typically measured in requests within a 5-hour rolling window.

StepFun — Step Plan

Plan Price Prompts / 5 hrs Est. Model Calls
Flash Mini $6.99/mo 100 ~1,500
Flash Plus $9.99/mo 400 ~6,000
Flash Pro $29/mo 1,500 ~22,500
Flash Max $99/mo 5,000 ~75,000

All plans include step-3.5-flash, multi-device login, and concurrent agent execution. Plus and above get priority API access and support.

Step Plan pricing →


Xiaomi MiMo — Token Plan

Plan Price Credits Est. Complex Tasks
Lite ~$6/mo (39 RMB) 60M ~120
Standard ~$16/mo (99 RMB) 200M ~400
Pro ~$50/mo (329 RMB) 700M ~1,400
Max ~$100/mo (659 RMB) 1.6B ~3,200

Credit multipliers: MiMo-V2-Omni 1x, MiMo-V2-Pro (≤256K) 2x, MiMo-V2-Pro (256K-1M) 4x. First purchase gets 12% off. No 5-hour limit.

Token Plan pricing →


MiniMax — Token Plan

Plan Price Requests / 5 hrs
Starter $10/mo 1,500
Plus $20/mo 4,500
Max $50/mo 15,000
Max-Highspeed $80/mo 15,000
Ultra-Highspeed $150/mo 30,000

All tiers include MiniMax-M2.7. Highspeed plans use the faster M2.7-Highspeed variant. Yearly plans available: Starter $100/yr, Plus $200/yr, Max $500/yr.

Token Plan pricing →


Zhipu (Z.ai) — Coding Plan

Plan Price Quota
Lite $18/mo 3x Claude Pro usage
Pro $72/mo 5x Lite Plan usage
Max $160/mo 4x Pro Plan usage

GLM-5.1 consumes 3x quota during peak hours, 2x off-peak. Pro and above include Vision and Web Search.

Coding Plan pricing →


Alibaba Cloud — Model Studio Coding Plan

Plan Price Quota Status
Lite ~$10/mo Lower quotas Discontinued for new purchases (Mar 2026)
Pro $50/mo 6,000 req/5hrs, 90K req/mo Available

Pro plan includes Qwen 3.5-Plus, Kimi K2.5, GLM-5, MiniMax-M2.5, and Qwen 3-Coder variants. Each query costs 5-30+ requests depending on complexity.

Coding Plan pricing →


Moonshot (Kimi) — Kimi Code

Plan Price Quota
Free $0 Basic chat, daily/weekly quotas
Moderato ~$19/mo 2,048 Kimi Code requests/week
Allegretto ~$49/mo Higher quotas
Vivace ~$129/mo Highest quotas, premium features

Plans include CLI and developer tools, priority compute, and 5-hour rolling token quotas.

Kimi Code →


Ollama — Cloud Plans

Plan Price Cloud Usage Concurrent Cloud Models
Free $0 Light 1
Pro $20/mo ($200/yr) 50x Free 3
Max $100/mo 5x Pro 10

Local model execution is always unlimited and free on all plans. Cloud usage resets every 5 hours (session) and weekly.

Ollama Cloud pricing →


Quick Tips

  1. Thinking Mode: For complex refactors, use models like deepseek-reasoner, kimi-k2-thinking, or mimo-v2-flash-thinking.
  2. Context Caching: DeepSeek, Anthropic, and MiMo support caching — reduces input costs by up to 90%.
  3. Regional Endpoints: In China, use open.bigmodel.cn for GLM, dashscope.aliyuncs.com for Qwen, api.stepfun.com for StepFun, or api.minimaxi.com for MiniMax for lower latency.
  4. Step Plan API: Use https://api.stepfun.ai/step_plan (not the standard endpoint) for Step Plan subscriptions.
  5. Ollama Local: Run ollama launch claude --config for guided zero-cost local setup with open-source models.
  6. Free Tiers: Step-3.5-Flash and MiMo-V2-Flash are available free (rate-limited) via OpenRouter.

Yorumlar (0)

Sonuc bulunamadi