Skip to content

Commit ad730ca

Browse files
authored
Merge pull request #44 from Jakedismo/claude/review-current-implementation-011CUr7KL9X815xqg5KbY9PX
Review and Refactor Current Implementation
2 parents 4eb8f58 + cc0ade3 commit ad730ca

24 files changed

+5814
-228
lines changed

.codegraph.toml.example

Lines changed: 44 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -42,29 +42,66 @@ batch_size = 64
4242
# Set to false for maximum speed if using an external agent
4343
enabled = false
4444

45-
# LLM provider: "ollama" or "lmstudio"
46-
# "lmstudio" recommended for MLX + Flash Attention 2 (macOS)
45+
# LLM provider: "ollama", "lmstudio", "anthropic", "openai", or "openai-compatible"
46+
# - "lmstudio": Local LLMs via LM Studio (recommended for MLX + Flash Attention 2 on macOS)
47+
# - "ollama": Local LLMs via Ollama
48+
# - "anthropic": Anthropic Claude API (requires API key)
49+
# - "openai": OpenAI GPT API (requires API key)
50+
# - "openai-compatible": Any OpenAI-compatible API endpoint
4751
provider = "lmstudio"
4852

4953
# LLM model identifier
50-
# For LM Studio: lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF/DeepSeek-Coder-V2-Lite-Instruct-Q4_K_M.gguf
54+
# For LM Studio: lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF
5155
# For Ollama: Model name (e.g., "qwen2.5-coder:14b", "codellama:13b")
52-
# Recommended: DeepSeek Coder v2 Lite Instruct Q4_K_M (superior performance)
56+
# For Anthropic: Model name (e.g., "claude-3-5-sonnet-20241022", "claude-3-5-haiku-20241022")
57+
# For OpenAI: Model name (e.g., "gpt-4o", "gpt-4o-mini", "gpt-4-turbo")
58+
# For OpenAI-compatible: Custom model name
59+
# Recommended: DeepSeek Coder v2 Lite Instruct Q4_K_M (local), or Claude 3.5 Sonnet (cloud)
5360
model = "lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF"
5461

55-
# LM Studio URL (default port 1234)
62+
# LM Studio URL (only used if provider is "lmstudio")
5663
lmstudio_url = "http://localhost:1234"
5764

58-
# Ollama URL
65+
# Ollama URL (only used if provider is "ollama")
5966
ollama_url = "http://localhost:11434"
6067

68+
# OpenAI-compatible base URL (only used if provider is "openai-compatible")
69+
# Example: "http://localhost:1234/v1" for LM Studio OpenAI endpoint
70+
# openai_compatible_url = "http://localhost:1234/v1"
71+
72+
# Anthropic API key (only used if provider is "anthropic")
73+
# Can also be set via ANTHROPIC_API_KEY environment variable
74+
# anthropic_api_key = "sk-ant-..."
75+
76+
# OpenAI API key (only used if provider is "openai" or some "openai-compatible" endpoints)
77+
# Can also be set via OPENAI_API_KEY environment variable
78+
# openai_api_key = "sk-..."
79+
6180
# Context window size (tokens)
6281
# DeepSeek Coder v2 Lite: 32768 tokens
82+
# Claude 3.5 Sonnet: 200000 tokens
83+
# GPT-4o: 128000 tokens
6384
context_window = 32000
6485

65-
# Temperature for generation (0.0 = deterministic, 1.0 = creative)
86+
# Temperature for generation (0.0 = deterministic, 2.0 = very creative)
6687
temperature = 0.1
6788

89+
# Maximum tokens to generate in responses (legacy parameter, use max_output_tokens for Responses API)
90+
max_tokens = 4096
91+
92+
# Maximum output tokens for Responses API and reasoning models
93+
# If not set, falls back to max_tokens
94+
# max_output_tokens = 4096
95+
96+
# Reasoning effort for reasoning models (o1, o3, o4-mini, GPT-5)
97+
# Options: "minimal", "low", "medium", "high"
98+
# Higher effort = more reasoning tokens = better quality but slower and more expensive
99+
# Only applies to reasoning models, ignored by standard models
100+
# reasoning_effort = "medium"
101+
102+
# Request timeout in seconds
103+
timeout_secs = 120
104+
68105
# Insights mode: "context-only", "balanced", or "deep"
69106
# - context-only: Return context only (fastest, for agents)
70107
# - balanced: Process top 10 files with LLM (good speed/quality)

Cargo.toml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@ members = [
1313
"crates/codegraph-lb",
1414
"crates/codegraph-ai",
1515
"crates/codegraph-api",
16+
"crates/codegraph-cli",
1617
"crates/core-rag-mcp-server",
1718
"scripts",
1819
"tests/integration"

0 commit comments

Comments
 (0)