Skip to content

Models vs Providers

Models determine capability: how smart the AI is, how fast it responds, and what kinds of work it can do.

Providers determine operating constraints: privacy policy, legal jurisdiction, availability, and access model.

Models are the AI systems themselves. In practice, it helps more to think in capability classes than in winner lists.

Model classWhat it is best at
frontier reasoning modelscomplex agentic work, debugging, and multi-file changes
strong general-purpose coding modelseveryday implementation and review loops
multimodal modelsscreenshots, mockups, diagrams, and visual/UI work
long-context modelslarge investigations when context is still selective
hosted open-weight modelsmore control without going fully local

The same model can be available through multiple providers:

ModelDirect ProviderAlso Available Via
Claude Sonnet 4.6AnthropicAWS Bedrock, Google Vertex
GPT-5.2OpenAIAzure OpenAI
Llama 4Meta (weights only)AWS Bedrock, Azure, Together AI
Mistral LargeMistralAWS Bedrock, Azure, Google Vertex

Different providers have different data policies:

  • Anthropic direct: US jurisdiction, 30-day retention
  • AWS Bedrock: Your AWS account, your retention settings
  • Azure OpenAI: EU data residency available
  • AWS Bedrock: Private endpoints, VPC integration
  • Direct providers: best when you want the vendor’s native tooling and latest releases
  • Aggregators: useful when you want one key, fallback routing, or easy model switching
  • Cloud platforms: useful when procurement, networking, or compliance requirements matter more than raw convenience
PlatformTypeModelsWhy Use It
OpenRouterAggregator200+One key for all models, fallback routing
Together AIAggregator100+Open models, fine-tuning
AWS BedrockCloudClaude, Llama, MistralEnterprise, VPC integration
Azure OpenAICloudGPT, DALL-EEU regions, compliance
Google VertexCloudGemini, ClaudeGCP integration

“Open weight” means the model weights are publicly available, so you can download and run them yourself. That is not the same thing as open source, which would also include training code and data.

ModelParamsWhy It Matters
Llama 4405BMeta’s flagship, runs locally or hosted
DeepSeek V3.2671B MoEStrong open-weight coding model
Qwen 3235BStrong multilingual, tool use
Kimi K2.51T MoEMoonshot’s flagship, massive scale
GLM 4.7320BZhipu’s latest, strong reasoning
Mistral Large123BEU-based, good reasoning

Self-hosting: Run on your own hardware. Zero data leaves your network and provider boundaries stay under your control.

No vendor lock-in: Switch providers freely. Model weights are yours.

Fine-tuning: Train on your codebase for domain-specific improvements.

Offline: Works without internet. Air-gapped environments.

Open weight models lag ~6-12 months behind frontier closed models (Claude, GPT). For most coding tasks, this gap is negligible. For cutting-edge agentic workflows, closed models still lead.

Tool typeWhat it does
local runnerone-command local inference for individuals
desktop runtimeGUI-based local model management
serving stackproduction-grade local or self-hosted inference

Pair with the editor or terminal workflow you already use. Note: “free” local models still require capable hardware (16GB+ RAM, GPU with 8GB+ VRAM recommended).