SaaS • Private Deployment • Docs • Discord
Deploy AI agents in your own data center or VPC and retain complete data security & control.
HelixML is an enterprise-grade platform for building and deploying AI agents with support for RAG (Retrieval-Augmented Generation), API calling, vision, and multi-provider LLM support. Build and deploy LLM applications by writing a simple helix.yaml
configuration file.
Our intelligent GPU scheduler packs models efficiently into available GPU memory and dynamically loads and unloads models based on demand, optimizing resource utilization.
- Easy-to-use Web UI for agent interaction and management
- Session-based architecture with pause/resume capabilities
- Multi-step reasoning with tool orchestration
- Memory management for context-aware interactions
- Support for multiple LLM providers (OpenAI, Anthropic, and local models)
- REST API integration with OpenAPI schema support
- MCP (Model Context Protocol) server compatibility
- GPTScript integration for advanced scripting
- OAuth token management for secure third-party access
- Custom tool development with flexible SDK
- Built-in document ingestion (PDFs, Word, text files)
- Web scraper for automatic content extraction
- Multiple RAG backends: Typesense, Haystack, PGVector, LlamaIndex
- Vector embeddings with PGVector for semantic search
- Vision RAG support for multimodal content
Main use cases:
- Upload and analyze corporate documents
- Add website documentation URLs to create instant customer support agents
- Build knowledge bases from multiple sources
Context is everything. Agents can process tens of thousands of tokens per step—Helix provides complete visibility under the hood:
Tracing features:
- View all agent execution steps
- Inspect requests and responses to LLM providers, third-party APIs, and MCP servers
- Real-time token usage tracking
- Pricing and cost analysis
- Performance metrics and debugging
- Multi-tenancy with organization, team, and role-based access control
- Scheduled tasks and cron jobs
- Webhook triggers for event-driven workflows
- Evaluation framework for testing and quality assurance
- Payment integration with Stripe support
- Notifications via Slack, Discord, and email
- Keycloak authentication with OAuth and OIDC support
HelixML uses a microservices architecture with the following components:
┌─────────────────────────────────────────────────────────┐
│ Frontend (React) │
│ vite + TypeScript │
└────────────────────┬────────────────────────────────────┘
│
┌────────────────────▼────────────────────────────────────┐
│ API / Control Plane (Go) │
│ ┌──────────────┬──────────────┬──────────────────────┐ │
│ │ Agents │ Knowledge │ Auth & Sessions │ │
│ │ Skills │ RAG Pipeline│ Organizations │ │
│ │ Tools │ Vector DB │ Usage Tracking │ │
│ └──────────────┴──────────────┴──────────────────────┘ │
└─────────┬──────────────────────────────────┬───────────┘
│ │
┌─────────▼──────────┐ ┌─────────▼──────────┐
│ PostgreSQL │ │ GPU Runners │
│ + PGVector │ │ Model Scheduler │
└────────────────────┘ └────────────────────┘
│
┌─────────▼──────────────────────────────────────────────┐
│ Supporting Services: Keycloak, Typesense, Haystack, │
│ GPTScript Runner, Chrome/Rod, Tika, SearXNG │
└────────────────────────────────────────────────────────┘
Three-layer agent hierarchy:
- Session: Manages agent lifecycle and state
- Agent: Coordinates skills and handles LLM interactions
- Skills: Group related tools for specific capabilities
- Tools: Individual actions (API calls, functions, scripts)
- Go 1.24.0 - Main backend language
- PostgreSQL + PGVector - Data storage and vector embeddings
- GORM - ORM for database operations
- Gorilla Mux - HTTP routing
- Keycloak - Identity and access management
- NATS - Message queue
- Zerolog - Structured logging
- React 18.3.1 - UI framework
- TypeScript - Type-safe JavaScript
- Material-UI (MUI) - Component library
- MobX - State management
- Vite - Build tool
- Monaco Editor - Code editing
- OpenAI SDK - GPT models integration
- Anthropic SDK - Claude models integration
- LangChain Go - LLM orchestration
- GPTScript - Scripting capabilities
- Typesense / Haystack / LlamaIndex - RAG backends
- Docker & Docker Compose - Containerization
- Kubernetes + Helm - Orchestration
- Flux - GitOps operator
Use our quickstart installer:
curl -sL -O https://get.helixml.tech/install.sh
chmod +x install.sh
sudo ./install.sh
The installer will prompt you before making changes to your system. By default, the dashboard will be available on http://localhost:8080
.
For setting up a deployment with a DNS name, see ./install.sh --help
or read the detailed docs. We've documented easy TLS termination for you.
Next steps:
- Attach your own GPU runners per runners docs
- Use any external OpenAI-compatible LLM
Use our Helm charts for production deployments:
All server configuration is done via environment variables. You can find the complete list of configuration options in api/pkg/config/config.go
.
Key environment variables:
OPENAI_API_KEY
- OpenAI API credentialsANTHROPIC_API_KEY
- Anthropic API credentialsPOSTGRES_*
- Database connection settingsKEYCLOAK_*
- Authentication settingsSERVER_URL
- Public URL for the deploymentRUNNER_*
- GPU runner configuration
See the configuration documentation for detailed setup instructions.
For local development, refer to the Helix local development guide.
Prerequisites:
- Docker Desktop (or Docker + Docker Compose)
- Go 1.24.0+
- Node.js 18+
- Make
Quick development setup:
# Clone the repository
git clone https://github.com/helixml/helix.git
cd helix
# Start supporting services
docker-compose up -d postgres keycloak
# Run the backend
cd api
go run . serve
# Run the frontend (in a new terminal)
cd frontend
npm install
npm run dev
See local-development.md
for comprehensive setup instructions.
- Overview - Platform introduction
- Getting Started - Build your first agent
- Control Plane Deployment - Production deployment guide
- Runner Deployment - GPU runner setup
- Agent Architecture - Technical specification
- API Reference - REST API documentation
- Contributing Guide - How to contribute
- Upgrading Guide - Migration instructions
We welcome contributions! Please see our Contributing Guide for details.
By contributing, you confirm that:
- Your changes will fall under the same license
- Your changes will be owned by HelixML, Inc.
Helix is licensed under a similar license to Docker Desktop. You can run the source code (in this repo) for free for:
- Personal Use: Individuals or people personally experimenting
- Educational Use: Schools and universities
- Small Business Use: Companies with under $10M annual revenue and less than 250 employees
If you fall outside of these terms, please use the Launchpad to purchase a license for large commercial use. Trial licenses are available for experimentation.
You are not allowed to use our code to build a product that competes with us.
- We generate revenue to support the development of Helix. We are an independent software company.
- We don't want cloud providers to take our open source code and build a rebranded service on top of it.
If you would like to use some part of this code under a more permissive license, please get in touch.
- Discord Community - Join our community for help and discussions
- GitHub Issues - Report bugs or request features
- Documentation - Comprehensive guides and references
- Email - Contact us for commercial inquiries
If you find Helix useful, please consider giving us a star on GitHub!
Built with ❤️ by HelixML, Inc.