Skip to content

Welcome to Backend.AI GO

Backend.AI GO is a powerful, user-friendly desktop application designed to let you run Large Language Models (LLMs) like Gemma 3, Qwen3, gpt-oss-120B, Qwen3-235b-a22b, Solar-Open-100B, and Mistral directly on your local machine.

Whether you are a developer looking for a local AI sandbox, a privacy-conscious professional, or an AI enthusiast, Backend.AI GO provides the tools you need to download, manage, and interact with the latest AI models without relying on cloud services.

Key Highlights

  • Privacy First: Your data stays on your machine. No cloud uploads required for local inference.
  • Hardware Optimized: High performance on macOS (Metal/MLX), Windows/Linux (NVIDIA CUDA, AMD ROCm), and standard CPUs.
  • Agentic Capabilities: Use built-in agents that can perform tasks using tools like web search, file operations, and more.
  • Easy Management: Search and download models directly from Hugging Face within the app.
  • Open Standard API: Provides an OpenAI-compatible API so you can use Backend.AI GO as a backend for your favorite AI tools.

How to use this documentation

If you are new to Backend.AI GO, we recommend starting with the Getting Started section.


Developed and maintained as part of the Backend.AI project.