Skip to content

Appendix E: Model Provider Selection Guide

The core capability of OpenClaw comes from large language models. Choosing the right model provider directly determines how intelligent your lobster is, how fast it responds, and how much you spend each month. This appendix systematically covers all major model providers to help you quickly select based on your budget, network environment, and use case.

Table of Contents


1. Landscape Overview

Model providers can be divided into four categories:

CategoryRepresentativesBest For
Aggregation GatewaysOpenRouter, SiliconFlowOne key for multiple models, ideal for beginners
Domestic DirectDeepSeek, Qwen, GLM, Kimi, Doubao, Hunyuan, etc.Direct connection in China, Alipay payments, low latency
International DirectOpenAI, Anthropic, Google, xAI, MistralPursuing top model capabilities, requires VPN or overseas nodes
Local DeploymentOllama, LM StudioCompletely offline, zero cost, data stays on device

Tutorial recommended path: Zero-cost onboarding with OpenRouter free models (Chapter 2) → Switch to SiliconFlow for daily use (top domestic paid choice) → Go direct with DeepSeek / OpenAI / Anthropic for maximum performance.


2. Quick Selection: Find Your Solution in 30 Seconds

Choose based on your situation:

Your SituationRecommended SolutionReason
Zero cost, don't want to spend anythingOpenRouter + stepfun/step-3.5-flash:freeFree model, sign up and use immediately
China-based user, want to spend the leastDeepSeek directBest domestic value, ¥1/million tokens
China-based user, want an all-in-one experienceSiliconFlow200+ models aggregated, new users get ¥16
Need the strongest reasoning capabilityOpenAI o3 / GPT-5Currently best overall, requires VPN
Need the strongest coding capabilityAnthropic Claude Opus 4.5Benchmark in programming
Need ultra-long contextGoogle Gemini 2.5 Pro1M token context window
Need search augmentationPerplexityBuilt-in real-time search, answers with citations
Data cannot leave the deviceOllama + DeepSeek R1Completely local, zero network dependency
Want free unlimited usageHunyuan hunyuan-liteTencent free model, no limits
Budget-conscious + quality requiredDeepSeek V3¥1/million tokens, close to GPT-4 level
Programming + code completionMistral CodestralModel optimized specifically for code

Don't want to register for a bunch of accounts? Aggregation gateways let you access multiple models with a single API Key, and switch between models freely.

3.1 OpenRouter

The default recommended solution in Chapter 2 of this tutorial.

OpenRouter is the world's largest model aggregation gateway, connecting to 300+ models including OpenAI, Anthropic, Google, Meta, Mistral, DeepSeek, and virtually all major providers.

Key Advantages:

  • Free models available: stepfun/step-3.5-flash:free, google/gemma-3-4b-it:free, etc. — zero-cost onboarding
  • One key for all models: Generate one API Key after registration and switch models using the provider/model-name format
  • Transparent pricing: Prices displayed directly on each model page, pay-as-you-go, no monthly fees
  • Accessible in China: Directly accessible in some regions, or via domestic proxies like SiliconFlow

Pricing Model:

  • Free models: $0 (rate limits apply)
  • Paid models: Each provider's original price + small markup (usually 5–15%)
  • Payment methods: International credit card, Crypto

OpenClaw Configuration:

bash
# Environment variable
export OPENROUTER_API_KEY="sk-or-v1-..."

# Set model
# openclaw models set openrouter/stepfun/step-3.5-flash:free

Link: openrouter.ai

3.2 SiliconFlow

Top choice for domestic paid use cases.

SiliconFlow is China's leading model aggregation platform, providing a unified API for 200+ open-source and commercial models.

Key Advantages:

  • New users get ¥16: Credited upon registration, enough for several days of use
  • Direct connection in China: No VPN required, low latency
  • Alipay payments: Most convenient for China-based users
  • Rich model selection: Full coverage of DeepSeek, Qwen, GLM, Llama, Mistral, and more
  • Free models available: Some open-source models offer free inference quotas

Pricing Model:

  • New users: ¥16 free credit
  • Pay-as-you-go: Price varies by model, usually slightly higher than direct connection
  • Top-up methods: Alipay, WeChat Pay

OpenClaw Configuration:

bash
export SILICONFLOW_API_KEY="sk-..."

# Model format: siliconflow:provider/model-name
# Example: siliconflow:deepseek-ai/DeepSeek-V3

Link: cloud.siliconflow.cn

3.3 Vercel AI Gateway

Vercel AI Gateway is Vercel's unified AI gateway, primarily targeting developers.

Features:

  • Unified SDK (ai npm package) to access multiple models
  • Built-in load balancing, failover, caching
  • Deep integration with the Vercel deployment platform
  • Better suited for application development rather than direct use with OpenClaw

Vercel AI Gateway is more of a development framework than an API proxy. For OpenClaw users, OpenRouter and SiliconFlow are more straightforward choices.

3.4 Aggregation Gateway Comparison

OpenRouterSiliconFlowVercel AI Gateway
Model Count300+200+Depends on config
Free Models✅ Multiple✅ Some
New User Benefits¥16 free credit
Direct Access in ChinaSome regions
Alipay
OpenClaw Integration✅ Native support✅ Native supportManual configuration required
Best ForOverseas users / free model seekersChina-based users / paid first choiceDevelopers

4. Domestic Model Providers

Common advantages of domestic providers: direct connection without VPN, Alipay/WeChat Pay, optimized for Chinese.

4.1 Quick Comparison

ProviderFlagship ModelInput Price (¥/million tokens)Output Price (¥/million tokens)Free QuotaContextHighlights
DeepSeekDeepSeek V312New user credit128KBest value, strong reasoning
DeepSeekDeepSeek R1416128KDeep reasoning, math/coding
QwenQwen3.5-plus26Free quota available128KAlibaba, excellent Chinese comprehension
GLMGLM-555New user credit128KAcademic background, strong tool calling
MoonshotKimi K2.588Limited free256KUltra-long context
DoubaoDoubao-Seed-2.00.30.6Included in Coding Plan128KByteDance, ultra-low price
Hunyuanhunyuan-liteFreeFreeUnlimited32KTencent free model
Hunyuanhunyuan-pro1550128KTencent flagship
MiniMaxMiniMax-M2.514Welcome credits200K+MoE architecture, multimodal
StepFunStep-3.5-flashFree (via OR)Free (via OR)Free via OpenRouter128KFree through OpenRouter
ERNIEERNIE-4.5-turbo412Free quota available128KBaidu ecosystem

Pricing note: The above are direct API prices (March 2026). Prices may differ slightly when using aggregation gateways (SiliconFlow/OpenRouter).

4.2 Provider Details

DeepSeek — Best Value King

DeepSeek has become one of the most popular domestic API providers thanks to its open-source strategy and exceptional value.

Model Lineup:

  • DeepSeek V3: General-purpose flagship, 128K context, MoE architecture (671B params, 37B active), performance close to GPT-4o at 1/30 the price
  • DeepSeek R1: Deep reasoning model, excels at math, coding, logical reasoning, chain-of-thought output
  • DeepSeek Coder: Code-specialized model

Advantages:

  • Among the strongest comprehensive capabilities of domestic models
  • Extremely low price (V3 input ¥1/million tokens)
  • Alipay top-up supported
  • Open-source model can be deployed locally (via Ollama)
  • API compatible with OpenAI format, easy integration

Notes:

  • May queue during peak hours (highly popular)
  • R1 model output is slower (deep reasoning takes time)

OpenClaw Configuration:

bash
export DEEPSEEK_API_KEY="sk-..."
# model: "deepseek:deepseek-chat"      # V3
# model: "deepseek:deepseek-reasoner"  # R1

Link: platform.deepseek.com

Qwen (Alibaba Cloud Bailian)

Qwen is Alibaba Cloud's large model platform, providing API services through the Bailian (DashScope) platform.

Model Lineup:

  • Qwen3.5-plus: Flagship general model, 128K context
  • Qwen3.5-turbo: High-value version
  • Qwen-VL: Visual understanding model
  • Qwen-Audio: Audio understanding model
  • Qwen-Coder: Code-specialized

Advantages:

  • Excellent Chinese comprehension (trained on Alibaba e-commerce + search data)
  • Complete model lineup (text/vision/audio/code)
  • Free quota for new users
  • Function Calling supported
  • Coding Plan Lite ¥10/month (18,000 requests)

OpenClaw Configuration:

bash
export DASHSCOPE_API_KEY="sk-..."
# model: "qwen:qwen-max"
# model: "qwen:qwen-plus"

Link: dashscope.console.aliyun.com | Bailian Platform

GLM (Zhipu AI)

Zhipu AI was incubated by Tsinghua University and is one of China's earliest large model companies.

Model Lineup:

  • GLM-5: Latest flagship, overall capability benchmarked against GPT-4o
  • GLM-4-plus: Balanced value
  • GLM-4-flash: Low-cost fast response
  • CogView: Image generation
  • CogVideoX: Video generation

Advantages:

  • Strong academic background (Tsinghua KEG Lab)
  • Outstanding Tool Use capability — OpenClaw's skill system depends on this
  • AutoClaw's built-in Pony-Alpha-2 is based on GLM architecture
  • Free quota for new users

OpenClaw Configuration:

bash
export ZHIPUAI_API_KEY="..."
# model: "glm:glm-4-plus"

Link: open.bigmodel.cn

Moonshot / Kimi

Moonshot is known for ultra-long context, with Kimi K2.5 supporting a 256K token context window.

Model Lineup:

  • Kimi K2.5: Flagship model, 256K context
  • Moonshot-v1-128k: 128K context version
  • Moonshot-v1-32k: 32K economy version

Advantages:

  • Ultra-long context (256K) — ideal for processing long documents and large codebases
  • Kimi Claw managed service (see Appendix C)
  • Search-augmented capability (integrates Yahoo Finance and other data sources)

Notes:

  • Price is on the higher end among domestic providers
  • Large token consumption in long-context scenarios

OpenClaw Configuration:

bash
export MOONSHOT_API_KEY="sk-..."
# model: "moonshot:moonshot-v1-128k"

Link: platform.moonshot.cn | Kimi.com

Doubao (Volcengine Ark)

Volcengine Ark is ByteDance's model service platform, providing the Doubao series of models.

Model Lineup:

  • Doubao-Seed-2.0: Latest flagship, MoE architecture
  • Doubao-pro: General high-performance
  • Doubao-lite: Lightweight and fast

Advantages:

  • Extremely low price: Input ¥0.3/million tokens, output ¥0.6/million tokens (possibly the lowest among mainstream models)
  • Deep integration with ArkClaw fully managed service (see Appendix C)
  • Coding Plan: Starting from ¥9.9/first month, supports ArkClaw + Claude Code + Cursor simultaneously
  • Deep Feishu integration

Notes:

  • Model capability is in the top tier domestically but not the absolute best
  • Some features require a Coding Plan subscription

OpenClaw Configuration:

bash
export ARK_API_KEY="..."
# Need to create inference endpoint in Volcengine Ark console
# model: "doubao:doubao-seed-2.0"

Link: console.volcengine.com/ark | Coding Plan

Hunyuan (Tencent)

Tencent Hunyuan offers a complete model lineup from free to flagship.

Model Lineup:

  • hunyuan-lite: Free and unlimited — currently the only mainstream model that is completely free with no usage limits
  • hunyuan-standard: Standard version
  • hunyuan-pro: Flagship version

Advantages:

  • hunyuan-lite free and unlimited — run OpenClaw 24/7 at zero cost
  • Deep integration with Tencent Cloud deployment (see Appendix C)
  • QQ/WeCom/WeChat ecosystem synergy
  • Function Calling supported

Notes:

  • hunyuan-lite has limited capability (32K context, suitable for simple tasks)
  • Flagship version is priced higher among domestic providers

OpenClaw Configuration:

bash
export HUNYUAN_SECRET_ID="..."
export HUNYUAN_SECRET_KEY="..."
# model: "hunyuan:hunyuan-lite"   # free
# model: "hunyuan:hunyuan-pro"    # flagship

Link: cloud.tencent.com/product/tclm

MiniMax

MiniMax is known for its MoE architecture and multimodal capabilities.

Model Lineup:

  • MiniMax-M2.5: Flagship MoE model (229B params, ~10B active), 200K+ context
  • MiniMax-Text: Text-specialized
  • MiniMax-VL: Visual understanding

Advantages:

  • MoE architecture keeps costs low (fewer active parameters, cheaper inference)
  • Built-in multimodal (image/video understanding, text-to-image/video)
  • MaxClaw fully managed service (see Appendix C)
  • 200K+ long context

OpenClaw Configuration:

bash
export MINIMAX_API_KEY="..."
# model: "minimax:MiniMax-M2.5"

Link: platform.minimaxi.com | MaxClaw

StepFun

StepFun specializes in high-performance reasoning models.

Model Lineup:

  • Step-3.5: Flagship model
  • Step-3.5-flash: Lightweight fast version — available for free via OpenRouter

Advantages:

  • stepfun/step-3.5-flash:free is free via OpenRouter (the onboarding solution in Chapter 2 of this tutorial)
  • Direct API pricing is also competitive
  • Fast inference speed

If you use StepFun's free model via OpenRouter, you do not need to register a separate StepFun account.

Link: platform.stepfun.com

ERNIE (Baidu Qianfan)

Baidu Qianfan provides the ERNIE series of models.

Model Lineup:

  • ERNIE-4.5-turbo: Flagship version
  • ERNIE-4.0: Previous-generation flagship
  • ERNIE-Speed/Lite: Economy versions

Advantages:

  • Backed by Baidu search data, rich Chinese knowledge base
  • Qianfan platform provides 7 official OpenClaw skills
  • Free trial quota available
  • Integrated with Baidu Cloud deployment solutions (see Appendix C)

Link: console.bce.baidu.com/qianfan | Qianfan Documentation

Z.AI

Z.AI provides model API services.

Features:

  • One of the officially supported providers for OpenClaw
  • For specific pricing and model details, please refer to the official website

Link: z.ai


5. International Model Providers

International providers generally offer the most cutting-edge model capabilities, but require a VPN and international credit card payments. If you are in China and cannot connect directly, you can access these models indirectly through OpenRouter or SiliconFlow.

5.1 Quick Comparison

ProviderFlagship ModelInput Price ($/million tokens)Output Price ($/million tokens)Free QuotaContextHighlights
OpenAIGPT-5~30~60128KBest overall, most complete ecosystem
OpenAIGPT-4o~2.5~10128KHigh-value multimodal
OpenAIo3~15~60200KDeep reasoning
AnthropicClaude Opus 4.5~15~75200KBest for coding, extra-long output
AnthropicClaude Sonnet 4.5~3~15200KValue coding
GoogleGemini 2.5 Pro~1.25~10✅ Free tier1MUltra-long context champion
GoogleGemini 2.5 Flash~0.15~0.6✅ Free tier1MExtreme value
xAIGrok 4~5~15✅ Free credits128KReal-time info (X/Twitter)
MistralMistral Large~2~6✅ Free tier128KEuropean open-source leader
MistralCodestral~0.3~0.9✅ Free tier256KCode-specialized, FIM support
PerplexitySonar Pro~3~15128KSearch-augmented, with citations

Pricing note: The above are reference prices as of March 2026. Please refer to each provider's official website for actual prices. OpenAI/Anthropic prices change frequently.

5.2 Provider Details

OpenAI — Industry Standard

OpenAI is the pioneer of the large model industry. The GPT series has long led in overall capability.

Model Lineup:

  • GPT-5: Latest flagship, strongest overall capability
  • GPT-4o: Multimodal flagship, supports text/image/audio input
  • GPT-4o-mini: Lightweight and high-value
  • o3 / o3-mini: Deep reasoning models (similar to DeepSeek R1, but stronger)
  • o1: Previous-generation reasoning model

Advantages:

  • Consistently leads in overall capability
  • Most complete ecosystem (function calling, JSON mode, structured output)
  • Best native OpenClaw support (many skills default to OpenAI format)
  • Strong multimodal capabilities (image understanding, audio)

Notes:

  • VPN required
  • International credit card required (Visa/Mastercard)
  • Price is on the higher end among mainstream providers
  • Chinese phone numbers cannot register (overseas or virtual number required)

OpenClaw Configuration:

bash
export OPENAI_API_KEY="sk-..."
# model: "openai:gpt-4o"
# model: "openai:o3"

Link: platform.openai.com

Anthropic (Claude) — King of Coding

Anthropic's Claude series excels in coding, long-text processing, and safety.

Model Lineup:

  • Claude Opus 4.5: Flagship, industry-best coding capability
  • Claude Sonnet 4.5: Value coding option
  • Claude Haiku 4.5: Lightweight and fast

Advantages:

  • Industry benchmark for coding capability (consistently leads SWE-bench rankings)
  • 200K context window
  • Outstanding safety design (Constitutional AI)
  • Extra-long output capability (generate thousands of lines of code in one request)

Notes:

  • VPN required
  • International credit card required
  • API price is high (Opus 4.5 output $75/million tokens)
  • Rate limits are relatively strict

OpenClaw Configuration:

bash
export ANTHROPIC_API_KEY="sk-ant-..."
# model: "anthropic:claude-sonnet-4-5-20250514"

Link: console.anthropic.com

Google (Gemini) — Ultra-Long Context

Google AI Studio provides the Gemini series of models, renowned for their ultra-long context windows.

Model Lineup:

  • Gemini 2.5 Pro: Flagship, 1M token context (one million!)
  • Gemini 2.5 Flash: Lightweight and fast, also supports 1M context
  • Gemini 2.5 Flash-8B: Ultra-lightweight

Advantages:

  • 1M token context — can ingest an entire book or codebase in one go
  • Free tier available (AI Studio is free with rate limits)
  • Strong multimodal capabilities (native support for image, video, audio input)
  • Gemini 2.5 Flash is extremely affordable ($0.15/million tokens input)

Notes:

  • VPN required (AI Studio)
  • Chinese language support not as strong as domestic models
  • Free tier has rate limits

OpenClaw Configuration:

bash
export GOOGLE_API_KEY="..."
# model: "google:gemini-2.5-pro"
# model: "google:gemini-2.5-flash"

Link: aistudio.google.com | Vertex AI

xAI (Grok)

xAI, founded by Elon Musk, has Grok models deeply integrated with the X/Twitter platform.

Model Lineup:

  • Grok 4: Latest flagship
  • Grok 3: Previous-generation flagship

Advantages:

  • Real-time information access (integrates X/Twitter data stream)
  • Free API credits for new users
  • Humorous conversational experience

Notes:

  • VPN required
  • International credit card required
  • Model ecosystem and tool calling support not as complete as OpenAI

Link: console.x.ai

Mistral — European Open-Source Leader

Mistral is Europe's most important AI company, known for open-source models and coding capabilities.

Model Lineup:

  • Mistral Large: Flagship general model, 128K context
  • Codestral: Designed specifically for code, 256K context, supports Fill-in-the-Middle (FIM)
  • Mistral Small: Lightweight economy version
  • Pixtral: Visual understanding model

Advantages:

  • Free tier available (some models free on La Plateforme)
  • Codestral excels at code: Optimized for programming, supports 80+ languages
  • Open-source models can be deployed locally
  • European data compliance (GDPR)

OpenClaw Configuration:

bash
export MISTRAL_API_KEY="..."
# model: "mistral:mistral-large-latest"
# model: "mistral:codestral-latest"

Link: mistral.ai | La Plateforme

Perplexity — Search Augmented

Perplexity provides search-augmented model APIs, where answers automatically include web citation sources.

Model Lineup:

  • Sonar Pro: Flagship search-augmented model
  • Sonar: Standard version

Advantages:

  • Built-in real-time web search — no need to configure search skills separately
  • Answers automatically include citation sources (URLs)
  • Ideal for scenarios requiring real-time information (news, research, fact-checking)

Notes:

  • Higher price (search cost included in token price)
  • Not suitable for pure creative/coding scenarios
  • VPN required

Link: docs.perplexity.ai


6. Local Deployment

Don't want to send your data to the cloud? Local deployment runs models entirely on your computer — zero cost, zero latency, complete privacy. The trade-off is that sufficient hardware is required, and model capability is usually weaker than cloud flagships.

OllamaLM Studio
TypeCLI toolGUI application
Supported PlatformsmacOS / Linux / WindowsmacOS / Linux / Windows
InterfaceCommand lineGraphical UI (beginner-friendly)
Model FormatGGUF (llama.cpp)GGUF (llama.cpp)
Model Libraryollama.com/libraryBuilt-in model search and download
API Compatibility✅ OpenAI format (localhost:11434)✅ OpenAI format (localhost:1234)
Resource UsageLow (inference only)Medium (includes GUI)
Best ForTechnical users / experienced with terminalBeginners / prefer GUI
Ollama Quick Start
bash
# Install
curl -fsSL https://ollama.com/install.sh | sh   # Linux/macOS
# Windows: Download installer from ollama.com

# Download and run models
ollama pull deepseek-r1:8b        # 8B param version, requires ~6GB VRAM/RAM
ollama pull qwen2.5:14b           # 14B param version, requires ~10GB
ollama pull llama3.3:8b           # Meta Llama 3.3

# Start API server (default http://localhost:11434)
ollama serve

OpenClaw Configuration:

bash
# Set local model
# openclaw models set ollama/deepseek-r1:8b

Hardware Recommendations:

Model SizeMinimum RAMRecommended GPURecommended Use Case
1–3B4GBNone neededSimple Q&A
7–8B8GB6GB VRAMDaily chat, simple coding
14B16GB12GB VRAMMore complex tasks
32B+32GB+24GB+ VRAMNear cloud quality

Apple Silicon users: The unified memory in M-series chips is especially well-suited for local models. A 16GB M4 can run 8B models smoothly; 24GB+ can run 14B models.

LM Studio Quick Start
  1. Download and install from lmstudio.ai
  2. Open the app and search for a model (e.g., "deepseek")
  3. One-click download, click "Start" to launch
  4. Enable "Local Server" in settings (default http://localhost:1234)

Advantage: Pure GUI operation, ideal for users unfamiliar with the command line. Built-in model performance testing lets you see inference speed at a glance.

Hugging Face Open-Source Models

Hugging Face is the world's largest open-source model hosting platform, where virtually all open-source large models are published.

Usage Methods:

  • Direct download: Download GGUF format model files and load with Ollama or LM Studio
  • Inference API: Hugging Face provides cloud inference API (free tier available)
  • Inference Endpoints: Paid dedicated inference instance deployment

Recommended Open-Source Models:

ModelParametersFeaturesHF Link
DeepSeek R11.5B–671BDeep reasoning, multiple sizesdeepseek-ai/DeepSeek-R1
Qwen 2.50.5B–72BBalanced general-purposeQwen/Qwen2.5
Llama 3.38B–70BMeta open-source flagshipmeta-llama/Llama-3.3
Mistral7B–24BEuropean open-sourcemistralai/
Gemma 32B–27BGoogle open-sourcegoogle/gemma-3

Open-source models typically need quantization (e.g., Q4_K_M) to run on consumer-grade hardware. Ollama provides quantized versions by default.


7. Selection Decision Framework

7.1 Four-Dimension Evaluation

Choosing a model provider requires balancing four dimensions:

DimensionSuggested WeightDescription
CapabilityCoreThe model's reasoning, coding, Chinese language, and tool calling abilities
CostImportantAPI price × your usage volume = monthly expenditure
AccessibilityPrerequisiteCan you access it directly from your network, and is the payment method available
EcosystemBonusLevel of integration with OpenClaw, community support

7.2 Scenario-Based Recommendations

ScenarioFirst ChoiceAlternativeReason
Zero-cost onboardingOpenRouter free modelsHunyuan hunyuan-liteFree to use, quick experience
Daily domestic useSiliconFlow + DeepSeek V3QwenBest value, all-in-one
Deep reasoning / mathDeepSeek R1OpenAI o3R1 is direct in China and cheaper
Programming / developmentAnthropic Claude Sonnet 4.5Mistral CodestralStrongest coding capability
Ultra-long document processingGoogle Gemini 2.5 ProMoonshot Kimi K2.51M context
Real-time information queriesPerplexity SonarxAI GrokBuilt-in search
Multimodal (image/video)OpenAI GPT-4oGoogle Gemini 2.5Native multimodal
Enterprise complianceAnthropic ClaudeMistral (GDPR)Safety/compliance design
Complete privacyOllama + local modelsLM StudioData stays on device
24/7 low-cost operationHunyuan hunyuan-liteDoubao Doubao-liteFree / ultra-low price

7.3 Model Capability Tiers (2026 Q1 Reference)

The following image, widely circulated in the community as the "LLM Throne" chart, visually shows the standings of today's mainstream models:

LLM Rankings — Community-recognized model capability tier chart

Image source: Community creation, for reference only. Actual rankings vary by benchmark and use case.

T0 (Top Tier): GPT-5, Claude Opus 4.5, Gemini 2.5 Pro, o3

Strongest overall capability, suitable for complex reasoning, high-quality creation, and difficult coding tasks. Highest price.

T1 (Main Force): GPT-4o, Claude Sonnet 4.5, DeepSeek V3/R1, Grok 4, Qwen3.5-plus, GLM-5

Fully capable for daily use, excellent value. DeepSeek V3 achieves near-T0 capability at 1/30 the price.

T2 (Economy): GPT-4o-mini, Gemini 2.5 Flash, Doubao-Seed-2.0, DeepSeek V3 (via aggregation), Step-3.5-flash

Best choice for simple tasks, extremely low price, fast response.

T3 (Free / Local): Hunyuan hunyuan-lite, OpenRouter free models, Ollama local models

Zero-cost operation, limited capability, suitable for exploration and lightweight scenarios.


8. Pricing Quick Reference

The following are API portal addresses and key information for each provider. Prices are for reference only — please check each provider's official website for real-time pricing.

8.1 Domestic Providers

ProviderPortalPaymentFree QuotaNotes
SiliconFlowcloud.siliconflow.cnAlipay/WeChat¥16 new userAggregation platform, widest model selection
DeepSeekplatform.deepseek.comAlipayNew user creditBest value
Qwendashscope.console.aliyun.comAlipayFree quota availableAlibaba Cloud
GLMopen.bigmodel.cnAlipay/WeChatNew user creditTsinghua affiliated
Moonshotplatform.moonshot.cnAlipayLimited freeUltra-long context
Doubaoconsole.volcengine.com/arkAlipayIncluded in Coding PlanByteDance, ultra-low price
Hunyuancloud.tencent.com/product/tclmWeChat/Alipaylite free unlimitedTencent
MiniMaxplatform.minimaxi.comAlipayWelcome creditsMoE multimodal
StepFunplatform.stepfun.comAlipayOr use OpenRouter for free
ERNIEconsole.bce.baidu.com/qianfanAlipayFree quota availableBaidu ecosystem

8.2 International Providers

ProviderPortalPaymentFree QuotaAccessible in China
OpenRouteropenrouter.aiInternational card/Crypto✅ Free modelsSome regions
OpenAIplatform.openai.comInternational card❌ VPN required
Anthropicconsole.anthropic.comInternational card❌ VPN required
Googleaistudio.google.comInternational card✅ Free tier❌ VPN required
xAIconsole.x.aiInternational card✅ Free credits❌ VPN required
Mistralconsole.mistral.aiInternational card✅ Free tier❌ VPN required
Perplexitydocs.perplexity.aiInternational card❌ VPN required

8.3 Local Deployment

ToolDownloadCostNotes
Ollamaollama.comFreeCLI tool, first choice for technical users
LM Studiolmstudio.aiFreeGUI application, beginner-friendly
Hugging Facehuggingface.coFree (download)Open-source model hosting platform

Licensed under CC BY-NC-SA 4.0