The easiest way to run LLMs locally. One-command installation for Llama, Qwen, Mistral, and more.

Supported Models

Llama 3.3Qwen 3.6MistralGemmaDeepSeekPhi-4

Key Features

  • One-command install
  • Model library
  • API server
  • Docker support

Pros

  • Completely free
  • Unlimited usage
  • Privacy
  • Easy setup

Cons

  • Requires GPU/RAM
  • Setup complexity
  • Slower than APIs

Best Use Cases

Privacy-focusedOffline workCost controlExperimentation