Installation¶
OpenJarvis runs entirely on your hardware. Choose the interface that fits your workflow.
Browser App¶
Run the full chat UI in your browser. Everything stays local — the backend runs on
your machine and the frontend connects via localhost.
One-command setup¶
The script handles everything:
- Checks for Python 3.10+ and Node.js 18+
- Installs Ollama if not present and pulls a starter model
- Installs Python and frontend dependencies
- Starts the backend API server and frontend dev server
- Opens
http://localhost:5173in your browser
Manual setup¶
If you prefer to run each step yourself:
Then open http://localhost:5173.
Desktop App¶
The desktop app is a native window for the OpenJarvis chat UI. All inference and backend processing happens on your local machine — the app connects to the backend you start locally.
Setup¶
Step 1. Start the backend (same as Browser App):
Step 2. Download and open the desktop app:
| Platform | Download |
|---|---|
| macOS (Apple Silicon) | OpenJarvis.dmg |
| Windows (64-bit) | OpenJarvis-setup.exe |
| Linux (DEB) | OpenJarvis.deb |
| Linux (RPM) | OpenJarvis.rpm |
| Linux (AppImage) | OpenJarvis.AppImage |
The app connects to http://localhost:8000 automatically.
macOS: \"app is damaged\"
If macOS says the app is damaged, clear the Gatekeeper quarantine flag:
This is normal for open-source apps distributed outside the App Store.All releases
Browse all versions on the GitHub Releases page.
Build from source¶
git clone https://github.com/open-jarvis/OpenJarvis.git
cd OpenJarvis/desktop
npm install
npm run tauri build
The built installer will be in desktop/src-tauri/target/release/bundle/.
CLI¶
The command-line interface is the fastest way to interact with OpenJarvis programmatically. Every feature is accessible from the terminal.
Install¶
Verify¶
First commands¶
jarvis ask "What is the capital of France?"
jarvis ask --agent orchestrator --tools calculator "What is 137 * 42?"
jarvis serve --port 8000
jarvis doctor
jarvis model list
jarvis chat
Inference backend required
The CLI requires a running inference backend (e.g., Ollama). See Setting up an inference backend below.
Python SDK¶
For programmatic access, the Jarvis class provides a high-level sync API.
Install¶
Quick example¶
from openjarvis import Jarvis
j = Jarvis()
print(j.ask("Explain quicksort in two sentences."))
j.close()
With agents and tools¶
result = j.ask_full(
"What is the square root of 144?",
agent="orchestrator",
tools=["calculator", "think"],
)
print(result["content"]) # "12"
print(result["tool_results"]) # tool invocations
print(result["turns"]) # number of agent turns
Composition layer¶
For full control, use the SystemBuilder:
from openjarvis import SystemBuilder
system = (
SystemBuilder()
.engine("ollama")
.model("qwen3:8b")
.agent("orchestrator")
.tools(["calculator", "web_search", "file_read"])
.enable_telemetry()
.enable_traces()
.build()
)
result = system.ask("Summarize the latest AI news.")
system.close()
See the Python SDK guide for the full API reference.
Requirements¶
| Requirement | Version | Notes |
|---|---|---|
| Python | 3.10+ | Required |
| Inference backend | Any | At least one of Ollama, vLLM, llama.cpp, SGLang, or a cloud API |
| Node.js | 18+ | Required for the browser UI; 22+ for the WhatsApp Baileys channel bridge |
Optional Extras¶
OpenJarvis uses optional extras to keep the base installation lightweight.
Inference Backends¶
| Extra | Install Command | Description |
|---|---|---|
inference-cloud |
uv sync --extra inference-cloud |
OpenAI and Anthropic APIs |
inference-google |
uv sync --extra inference-google |
Google Gemini API |
Ollama, vLLM, and llama.cpp are HTTP-based
These engines have no additional Python dependencies — OpenJarvis communicates over HTTP. You still need the engine software running on your machine.
Memory Backends¶
| Extra | Install Command | Description |
|---|---|---|
memory-faiss |
uv sync --extra memory-faiss |
FAISS vector store |
memory-colbert |
uv sync --extra memory-colbert |
ColBERTv2 late-interaction retrieval |
memory-bm25 |
uv sync --extra memory-bm25 |
BM25 sparse retrieval |
SQLite memory is always available
The default SQLite/FTS5 memory backend requires no additional dependencies.
Server & Other¶
| Extra | Install Command | Description |
|---|---|---|
server |
uv sync --extra server |
OpenAI-compatible API server (jarvis serve) |
dev |
uv sync --extra dev |
Development and testing tools |
docs |
uv sync --extra docs |
Documentation build tools |
Combine extras:
Setting Up an Inference Backend¶
OpenJarvis requires at least one inference backend. Choose the one that matches your hardware.
Ollama (Recommended)¶
The easiest way to get started. Handles model downloading and serving automatically.
- Install from ollama.com
-
Start the server and pull a model:
-
Verify:
jarvis model list
Best for: Apple Silicon Macs, consumer NVIDIA GPUs, CPU-only systems
vLLM¶
High-throughput serving optimized for datacenter GPUs.
- Install following the official guide
- Start:
vllm serve Qwen/Qwen2.5-7B-Instruct - Auto-detected at
http://localhost:8000
Best for: NVIDIA datacenter GPUs (A100, H100), AMD GPUs
llama.cpp¶
Efficient CPU and GPU inference with GGUF quantized models.
- Build from github.com/ggerganov/llama.cpp
- Start:
llama-server -m /path/to/model.gguf --port 8080 - Auto-detected at
http://localhost:8080
Cloud APIs¶
uv sync --extra inference-cloud --extra inference-google
export OPENAI_API_KEY="sk-..."
export ANTHROPIC_API_KEY="sk-ant-..."
Next Steps¶
- Quick Start — Run your first query
- Configuration — Customize engine hosts, model routing, memory, and more