Continuum Router & API¶
Backend.AI GO isn't just a chat app; it's a powerful AI API Gateway for your local network.
What is Continuum Router?¶
The Continuum Router is an internal component that acts as a traffic controller. It provides a single, unified entry point for all your AI models—whether they are running locally on your machine or in the cloud.
For more detailed technical documentation, please refer to docs.continuum.lablup.ai.
OpenAI-Compatible API¶
Backend.AI GO exposes an API that mimics the OpenAI API standard. This means you can use Backend.AI GO as a drop-in replacement for any application that supports OpenAI.
- Endpoint:
http://localhost:8000/v1(Default) - Authentication: Optional (configurable in Settings)
Example: Using with curl¶
curl http://localhost:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d
"model": "llama-3",
"messages": [{"role": "user", "content": "Hello!"}]
External Access¶
By default, the API is only accessible from your own computer (127.0.0.1). If you want to access your models from other devices on your local network (like a tablet or another laptop):
- Go to Settings > Network.
- Enable Allow External Access.
- Restart the router.
- Use your computer's local IP address (e.g.,
http://192.168.1.10:8000/v1) in your other apps.
Security Tip
Only enable external access if you trust all devices on your local network.