> cursor-model-selection
Configure and select AI models in Cursor for Chat, Composer, and Agent mode. Triggers on "cursor model", "cursor gpt", "cursor claude", "change cursor model", "cursor ai model", "cursor auto mode".
curl "https://skillshub.wtf/jeremylongshore/claude-code-plugins-plus-skills/cursor-model-selection?format=md"Cursor Model Selection
Configure AI models for Chat, Composer, and Agent mode. Cursor supports models from OpenAI, Anthropic, Google, and its own proprietary models. Choosing the right model per task is a major productivity lever.
Available Models
Included with Cursor Subscription
| Model | Provider | Best For | Context |
|---|---|---|---|
| GPT-4o | OpenAI | General coding, fast responses | 128K |
| GPT-4o-mini | OpenAI | Simple tasks, cost-efficient | 128K |
| Claude Sonnet | Anthropic | Code quality, detailed explanations | 200K |
| Claude Haiku | Anthropic | Fast simple tasks | 200K |
| cursor-small | Cursor | Quick completions, simple edits | 8K |
| Auto | Cursor | Automatic model selection per query | Varies |
Premium Models (count against fast request quota)
| Model | Provider | Best For | Context |
|---|---|---|---|
| Claude Opus | Anthropic | Complex architecture, hard bugs | 200K |
| GPT-5 | OpenAI | Advanced reasoning, complex code | 128K+ |
| o1 / o3 | OpenAI | Deep reasoning, mathematical logic | 128K |
| Gemini 2.5 Pro | Design, large context analysis | 1M |
Model Selection by Task
Quick Reference
Bug fix in one file → GPT-4o or Claude Sonnet
Multi-file refactoring → Claude Sonnet or Opus
Architecture planning → Claude Opus or GPT-5
Test generation → GPT-4o (fast + good patterns)
Complex algorithm design → o1/o3 reasoning models
Large codebase analysis → Gemini 2.5 Pro (1M context)
Simple autocomplete → cursor-small (automatic via Tab)
"I don't know" → Auto mode
How to Switch Models
Per conversation: Click the model name in the top-right of Chat or Composer panel.
Default model: Cursor Settings > Models > set default for Chat and Composer separately.
Auto mode: Select "Auto" as the model. Cursor picks the best model per query based on complexity and current server load.
Bring Your Own Key (BYOK)
Use your own API keys to bypass Cursor's quota system. You pay the provider directly at their rates.
Configuration
Cursor Settings > Models > enable Use own API key:
OpenAI:
API Key: sk-proj-xxxxxxxxxxxxxxxxxxxx
Anthropic:
API Key: sk-ant-xxxxxxxxxxxxxxxxxxxx
Google (Gemini):
API Key: AIzaSyxxxxxxxxxxxxxxxxxxxxxxxxx
Azure OpenAI
For enterprise Azure deployments:
Cursor Settings > Models > Azure:
API Key: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
Endpoint: https://my-instance.openai.azure.com
Deployment: gpt-4o-deployment-name
API Version: 2024-10-21
Adding Custom Models
For OpenAI-compatible providers (Ollama, LM Studio, Together AI):
Cursor Settings>Models>Add Model- Enter model name (e.g.,
llama-3.1-70b) - Enable
Override OpenAI Base URL - Enter base URL:
http://localhost:11434/v1(Ollama) or provider URL - Enter API key if required
BYOK Limitations
| Feature | Uses BYOK Key? | Uses Cursor Model? |
|---|---|---|
| Chat | Yes | -- |
| Composer | Yes | -- |
| Agent mode | Yes | -- |
| Tab Completion | No | Always Cursor model |
| Apply from Chat | No | Always Cursor model |
Tab Completion always uses Cursor's proprietary model regardless of BYOK configuration.
Cost Optimization Strategies
Tiered Model Usage
Tier 1 (Fast + Cheap): cursor-small, GPT-4o-mini, Claude Haiku
Use for: simple questions, syntax help, boilerplate
Tier 2 (Balanced): GPT-4o, Claude Sonnet
Use for: most coding tasks, debugging, refactoring
Tier 3 (Premium): Claude Opus, GPT-5, o1/o3
Use for: architecture decisions, critical bugs, complex logic
Quota Management
Cursor subscription includes a monthly quota of "fast requests" (premium model uses). When exceeded, requests queue behind other users ("slow requests").
- Check remaining quota:
cursor.com/settings> Usage - Pro plan: ~500 fast requests/month
- Business plan: ~500 fast requests/month per seat
Tips to Reduce Usage
- Use Auto mode -- it picks cheaper models when they suffice
- Start with Sonnet/GPT-4o, escalate to Opus/o1 only if needed
- Write detailed prompts to avoid back-and-forth (fewer requests)
- Use BYOK for heavy usage -- pay per token instead of per request
Model Behavior Differences
Code Generation Style
# Claude models: Verbose, well-documented, defensive
def process_order(order: Order) -> Result[ProcessedOrder, OrderError]:
"""Process an order through the payment and fulfillment pipeline.
Args:
order: The order to process.
Returns:
Result containing the processed order or an error.
Raises:
Never raises -- errors returned as Result.Err.
"""
if not order.items:
return Err(OrderError.EMPTY_ORDER)
...
# GPT models: Concise, pragmatic, fewer comments
def process_order(order: Order) -> ProcessedOrder:
if not order.items:
raise ValueError("Order has no items")
...
Reasoning Models (o1, o3)
These models "think" before responding. They are slower but significantly better at:
- Multi-step logic problems
- Finding subtle bugs in complex code
- Mathematical or algorithmic optimization
- Understanding implicit requirements
They are overkill for simple tasks. Use them deliberately for hard problems.
Enterprise Considerations
- Model access control: Admins can restrict which models team members access via the admin dashboard
- Spending limits: Set per-user or per-team spending caps when using BYOK
- Compliance: Some models route through different providers -- verify data handling per model
- Azure preference: Enterprise teams on Azure can route all requests through their own Azure OpenAI deployments
- Audit: Model selection per request is visible in usage analytics (Business/Enterprise plans)
Resources
> related_skills --same-repo
> fathom-cost-tuning
Optimize Fathom API usage and plan selection. Trigger with phrases like "fathom cost", "fathom pricing", "fathom plan".
> fathom-core-workflow-b
Sync Fathom meeting data to CRM and build automated follow-up workflows. Use when integrating Fathom with Salesforce, HubSpot, or custom CRMs, or creating automated post-meeting email summaries. Trigger with phrases like "fathom crm sync", "fathom salesforce", "fathom follow-up", "fathom post-meeting workflow".
> fathom-core-workflow-a
Build a meeting analytics pipeline with Fathom transcripts and summaries. Use when extracting insights from meetings, building CRM sync, or creating automated meeting follow-up workflows. Trigger with phrases like "fathom analytics", "fathom meeting pipeline", "fathom transcript analysis", "fathom action items sync".
> fathom-common-errors
Diagnose and fix Fathom API errors including auth failures and missing data. Use when API calls fail, transcripts are empty, or webhooks are not firing. Trigger with phrases like "fathom error", "fathom not working", "fathom api failure", "fix fathom".