@@ -42,29 +42,66 @@ batch_size = 64
4242# Set to false for maximum speed if using an external agent
4343enabled = false
4444
45- # LLM provider: "ollama" or "lmstudio"
46- # "lmstudio" recommended for MLX + Flash Attention 2 (macOS)
45+ # LLM provider: "ollama", "lmstudio", "anthropic", "openai", or "openai-compatible"
46+ # - "lmstudio": Local LLMs via LM Studio (recommended for MLX + Flash Attention 2 on macOS)
47+ # - "ollama": Local LLMs via Ollama
48+ # - "anthropic": Anthropic Claude API (requires API key)
49+ # - "openai": OpenAI GPT API (requires API key)
50+ # - "openai-compatible": Any OpenAI-compatible API endpoint
4751provider = " lmstudio"
4852
4953# LLM model identifier
50- # For LM Studio: lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF/DeepSeek-Coder-V2-Lite-Instruct-Q4_K_M.gguf
54+ # For LM Studio: lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF
5155# For Ollama: Model name (e.g., "qwen2.5-coder:14b", "codellama:13b")
52- # Recommended: DeepSeek Coder v2 Lite Instruct Q4_K_M (superior performance)
56+ # For Anthropic: Model name (e.g., "claude-3-5-sonnet-20241022", "claude-3-5-haiku-20241022")
57+ # For OpenAI: Model name (e.g., "gpt-4o", "gpt-4o-mini", "gpt-4-turbo")
58+ # For OpenAI-compatible: Custom model name
59+ # Recommended: DeepSeek Coder v2 Lite Instruct Q4_K_M (local), or Claude 3.5 Sonnet (cloud)
5360model = " lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF"
5461
55- # LM Studio URL (default port 1234 )
62+ # LM Studio URL (only used if provider is "lmstudio" )
5663lmstudio_url = " http://localhost:1234"
5764
58- # Ollama URL
65+ # Ollama URL (only used if provider is "ollama")
5966ollama_url = " http://localhost:11434"
6067
68+ # OpenAI-compatible base URL (only used if provider is "openai-compatible")
69+ # Example: "http://localhost:1234/v1" for LM Studio OpenAI endpoint
70+ # openai_compatible_url = "http://localhost:1234/v1"
71+
72+ # Anthropic API key (only used if provider is "anthropic")
73+ # Can also be set via ANTHROPIC_API_KEY environment variable
74+ # anthropic_api_key = "sk-ant-..."
75+
76+ # OpenAI API key (only used if provider is "openai" or some "openai-compatible" endpoints)
77+ # Can also be set via OPENAI_API_KEY environment variable
78+ # openai_api_key = "sk-..."
79+
6180# Context window size (tokens)
6281# DeepSeek Coder v2 Lite: 32768 tokens
82+ # Claude 3.5 Sonnet: 200000 tokens
83+ # GPT-4o: 128000 tokens
6384context_window = 32000
6485
65- # Temperature for generation (0.0 = deterministic, 1 .0 = creative)
86+ # Temperature for generation (0.0 = deterministic, 2 .0 = very creative)
6687temperature = 0.1
6788
89+ # Maximum tokens to generate in responses (legacy parameter, use max_output_tokens for Responses API)
90+ max_tokens = 4096
91+
92+ # Maximum output tokens for Responses API and reasoning models
93+ # If not set, falls back to max_tokens
94+ # max_output_tokens = 4096
95+
96+ # Reasoning effort for reasoning models (o1, o3, o4-mini, GPT-5)
97+ # Options: "minimal", "low", "medium", "high"
98+ # Higher effort = more reasoning tokens = better quality but slower and more expensive
99+ # Only applies to reasoning models, ignored by standard models
100+ # reasoning_effort = "medium"
101+
102+ # Request timeout in seconds
103+ timeout_secs = 120
104+
68105# Insights mode: "context-only", "balanced", or "deep"
69106# - context-only: Return context only (fastest, for agents)
70107# - balanced: Process top 10 files with LLM (good speed/quality)
0 commit comments