Quick Start
How It Works
Choosing a Model
| Model | Best For | Speed |
|---|---|---|
gpt-4o-mini | Fast responses, cost effective | ⚡ Fast |
gpt-4o | High quality, complex tasks | 🔵 Medium |
claude-3-opus | Long documents, analysis | 🔵 Medium |
ollama/llama3 | Privacy, offline use | 🟢 Local |
Configuration
| Option | Type | Default | Description |
|---|---|---|---|
model | String | gpt-4o-mini | Model name |
api_key | String | From ENV | API key |
base_url | String | OpenAI default | API endpoint |
temperature | f32 | 0.7 | Randomness (0-1) |
max_tokens | u32 | None | Max response length |
Best Practices
Use environment variables for API keys
Use environment variables for API keys
Set
OPENAI_API_KEY instead of hardcoding keys.Start with gpt-4o-mini
Start with gpt-4o-mini
Fast and cheap - upgrade to gpt-4o only when needed.
Use local models for privacy
Use local models for privacy
Ollama runs models locally with no data leaving your machine.

