ManyLLM v1.0.0
Local-first AI platform
Introduction
Welcome to ManyLLM documentation. Learn how to run, manage, and integrate multiple local LLMs in one unified workspace.
Install & Launch
Download and run ManyLLM on your system.
Add Models
Configure your first local model via the UI.
Start Chatting
Use the chat interface or API endpoints.
# 1. Start ManyLLM
./ManyLLM
# 2. Configure your first model (via UI or CLI)
manyllm model add llama3 --provider ollama
# 3. Test the API
curl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer your-local-key" \
-d '{
"model": "llama3",
"messages": [{"role": "user", "content": "Hello!"}],
"temperature": 0.7
}'
Key Features
Drop-in replacement for OpenAI API with local models. Use your existing tools and scripts.
View API ReferenceOrganize conversations with file context and local RAG capabilities for enhanced productivity.
Learn About WorkspacesAutomatic detection and management of Ollama, llama.cpp, and MLX runtimes.
Model ConfigurationLocal-first architecture ensures your data never leaves your device by default.
Privacy PolicyFor Developers
- • API Integration Guide
- • Custom Model Configuration
- • Workspace Automation
For Teams
- • Shared Workspace Setup
- • Model Performance Optimization
- • Privacy & Security Best Practices
Documentation In Progress
Comprehensive documentation is being written alongside the desktop application. Check back soon for detailed guides, API references, and tutorials.