# Option 1: One-line install (Linux/macOS)
wget -qO- https://raw.githubusercontent.com/alchemiststudiosDOTai/tunacode/master/scripts/install_linux.sh | bash
# Option 2: UV install (recommended)
uv tool install tunacode-cli
# Option 3: pip install
pip install tunacode-cli
For detailed installation and configuration instructions, see the Getting Started Guide.
# 1) Install (choose one)
uv tool install tunacode-cli # recommended
# or: pip install tunacode-cli
# 2) Launch the CLI
tunacode --wizard # guided setup (enter an API key, pick a model)
# 3) Try common commands in the REPL
/help # see commands
/model # explore models and set a default
/plan # enter read-only Plan Mode
Tip: You can also skip the wizard and set everything via flags:
tunacode --model openai:gpt-4.1 --key sk-your-key
For contributors and developers who want to work on TunaCode:
# Clone the repository
git clone https://github.com/alchemiststudiosDOTai/tunacode.git
cd tunacode
# Quick setup (recommended) - uses UV automatically if available
./scripts/setup_dev_env.sh
# Or manual setup with UV (recommended)
uv venv
source .venv/bin/activate # On Windows: .venv\Scripts\activate
uv pip install -e ".[dev]"
# Alternative: traditional setup
python3 -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate
pip install -e ".[dev]"
# Verify installation
tunacode --version
See the Hatch Build System Guide for detailed instructions on the development environment.
Choose your AI provider and set your API key. For more details, see the Configuration Section in the Getting Started Guide. For local models (LM Studio, Ollama, etc.), see the Local Models Setup Guide.
TunaCode now automatically saves your model selection for future sessions. When you choose a model using /model <provider:name>
, it will be remembered across restarts.
If you encounter API key errors, you can manually create a configuration file that matches the current schema:
# Create the config file
cat > ~/.config/tunacode.json << 'EOF'
{
"default_model": "openai:gpt-4.1",
"env": {
"OPENAI_API_KEY": "your-openai-api-key-here",
"ANTHROPIC_API_KEY": "",
"GEMINI_API_KEY": "",
"OPENROUTER_API_KEY": ""
},
"settings": {
"enable_streaming": true,
"max_iterations": 40,
"context_window_size": 200000
},
"mcpServers": {}
}
EOF
Replace the model and API key with your preferred provider and credentials. Examples:
openai:gpt-4.1
(requires OPENAI_API_KEY)anthropic:claude-4-sonnet-20250522
(requires ANTHROPIC_API_KEY)google:gemini-2.5-pro
(requires GEMINI_API_KEY)
I apologize for any recent issues with model selection and configuration. I'm actively working to fix these problems and improve the overall stability of TunaCode. Your patience and feedback are greatly appreciated as I work to make the tool more reliable.
Based on extensive testing, these models provide the best performance:
google/gemini-2.5-pro
- Excellent for complex reasoningopenai/gpt-4.1
- Strong general-purpose modeldeepseek/deepseek-r1-0528
- Great for code generationopenai/gpt-4.1-mini
- Fast and cost-effectiveanthropic/claude-4-sonnet-20250522
- Superior context handling
Note: Formal evaluations coming soon. Any model can work, but these have shown the best results in practice.
tunacode
Command | Description |
---|---|
/help |
Show all commands |
/model <provider:name> |
Switch model |
/clear |
Clear message history |
/compact |
Summarize conversation |
/branch <name> |
Create Git branch |
/yolo |
Skip confirmations |
!<command> |
Run shell command |
exit |
Exit TunaCode |
TunaCode leverages parallel execution for read-only operations, achieving 3x faster file operations:
Multiple file reads, directory listings, and searches execute concurrently using async I/O, making code exploration significantly faster.
- Bug Fixes: Actively addressing issues - please report any bugs you encounter!
Note: While the tool is fully functional, we're focusing on stability and core features before optimizing for speed.
- Use Git branches before making changes
- Review file modifications before confirming
- Keep backups of important work
For a complete overview of the documentation, see the Documentation Hub.
- Getting Started - How to install, configure, and use TunaCode.
- Commands - A complete list of all available commands.
- Architecture (planned) - The overall architecture of the TunaCode application.
- Contributing (planned) - Guidelines for contributing to the project.
- Tools (planned) - How to create and use custom tools.
- Testing (planned) - Information on the testing philosophy and how to run tests.
- Advanced Configuration - An example of an advanced configuration file.
- Changelog (planned) - A history of changes to the application.
- Roadmap (planned) - The future direction of the project.
- Security (planned) - Information about the security of the application.
MIT License - see LICENSE file