Skip to content

Welcome to Backend.AI GO

Backend.AI GO is a powerful, user-friendly desktop application designed to let you run Large Language Models (LLMs) like Gemma 3, Qwen3, gpt-oss-120B, Qwen3-235b-a22b, Solar-Open-100B, and Mistral directly on your local machine.

Whether you are a developer looking for a local AI sandbox, a privacy-conscious professional, or an AI enthusiast, Backend.AI GO provides the tools you need to download, manage, and interact with the latest AI models without relying on cloud services.

Key Highlights

  • Privacy First: Your data stays on your machine. No cloud uploads required for local inference.
  • Hardware Optimized: High performance on macOS (Metal/MLX), Windows/Linux (NVIDIA CUDA, AMD ROCm), and standard CPUs.
  • Agentic Capabilities: Use built-in agents that can perform tasks using tools like web search, file operations, and more.
  • Easy Management: Search and download models directly from Hugging Face within the app.
  • Open Standard API: Provides an OpenAI-compatible API so you can use Backend.AI GO as a backend for your favorite AI tools.

How to use this documentation

If you are new to Backend.AI GO, we recommend starting with the Getting Started section.

  • Quickstart: Get up and running in less than 5 minutes.
  • Cowork: Learn how to use autonomous agents.
  • CLI Reference: For those who prefer the terminal.

Developed and maintained as part of the Backend.AI project.