diff --git a/.devcontainer/.zshrc b/.devcontainer/.zshrc deleted file mode 100644 index 7df48cf..0000000 --- a/.devcontainer/.zshrc +++ /dev/null @@ -1,111 +0,0 @@ -# If you come from bash you might have to change your $PATH. -export PATH=$HOME/bin:$HOME/.local/bin:/usr/local/bin:$PATH - -# Path to your Oh My Zsh installation. -export ZSH=$HOME/.oh-my-zsh - -# Set name of the theme to load --- if set to "random", it will -# load a random theme each time Oh My Zsh is loaded, in which case, -# to know which specific one was loaded, run: echo $RANDOM_THEME -# See https://github.com/ohmyzsh/ohmyzsh/wiki/Themes -ZSH_THEME="powerlevel10k/powerlevel10k" - -# Set list of themes to pick from when loading at random -# Setting this variable when ZSH_THEME="powerlevel10k/powerlevel10k" -# a theme from this variable instead of looking in $ZSH/themes/ -# If set to an empty array, this variable will have no effect. -# ZSH_THEME_RANDOM_CANDIDATES=( "robbyrussell" "agnoster" ) - -# Uncomment the following line to use case-sensitive completion. -# CASE_SENSITIVE="true" - -# Uncomment the following line to use hyphen-insensitive completion. -# Case-sensitive completion must be off. _ and - will be interchangeable. -# HYPHEN_INSENSITIVE="true" - -# Uncomment one of the following lines to change the auto-update behavior -# zstyle ':omz:update' mode disabled # disable automatic updates -# zstyle ':omz:update' mode auto # update automatically without asking -# zstyle ':omz:update' mode reminder # just remind me to update when it's time - -# Uncomment the following line to change how often to auto-update (in days). -# zstyle ':omz:update' frequency 13 - -# Uncomment the following line if pasting URLs and other text is messed up. -# DISABLE_MAGIC_FUNCTIONS="true" - -# Uncomment the following line to disable colors in ls. -# DISABLE_LS_COLORS="true" - -# Uncomment the following line to disable auto-setting terminal title. -# DISABLE_AUTO_TITLE="true" - -# Uncomment the following line to enable command auto-correction. -# ENABLE_CORRECTION="true" - -# Uncomment the following line to display red dots whilst waiting for completion. -# You can also set it to another string to have that shown instead of the default red dots. -# e.g. COMPLETION_WAITING_DOTS="%F{yellow}waiting...%f" -# Caution: this setting can cause issues with multiline prompts in zsh < 5.7.1 (see #5765) -# COMPLETION_WAITING_DOTS="true" - -# Uncomment the following line if you want to disable marking untracked files -# under VCS as dirty. This makes repository status check for large repositories -# much, much faster. -# DISABLE_UNTRACKED_FILES_DIRTY="true" - -# Uncomment the following line if you want to change the command execution time -# stamp shown in the history command output. -# You can set one of the optional three formats: -# "mm/dd/yyyy"|"dd.mm.yyyy"|"yyyy-mm-dd" -# or set a custom format using the strftime function format specifications, -# see 'man strftime' for details. -# HIST_STAMPS="mm/dd/yyyy" - -# Would you like to use another custom folder than $ZSH/custom? -# ZSH_CUSTOM=/path/to/new-custom-folder - -# Which plugins would you like to load? -# Standard plugins can be found in $ZSH/plugins/ -# Custom plugins may be added to $ZSH_CUSTOM/plugins/ -# Example format: plugins=(rails git textmate ruby lighthouse) -# Add wisely, as too many plugins slow down shell startup. -plugins=(git) - -source $ZSH/oh-my-zsh.sh - -# User configuration - -# export MANPATH="/usr/local/man:$MANPATH" - -# You may need to manually set your language environment -# export LANG=en_US.UTF-8 - -# Preferred editor for local and remote sessions -# if [[ -n $SSH_CONNECTION ]]; then -# export EDITOR='vim' -# else -# export EDITOR='mvim' -# fi - -# Compilation flags -# export ARCHFLAGS="-arch $(uname -m)" - -# Set personal aliases, overriding those provided by Oh My Zsh libs, -# plugins, and themes. Aliases can be placed here, though Oh My Zsh -# users are encouraged to define aliases within a top-level file in -# the $ZSH_CUSTOM folder, with .zsh extension. Examples: -# - $ZSH_CUSTOM/aliases.zsh -# - $ZSH_CUSTOM/macos.zsh -# For a full list of active aliases, run `alias`. -# -# Example aliases -# alias zshconfig="mate ~/.zshrc" -# alias ohmyzsh="mate ~/.oh-my-zsh" -DISABLE_AUTO_UPDATE=true -DISABLE_UPDATE_PROMPT=true -[[ ! -f ~/.p10k.zsh ]] || source ~/.p10k.zsh - -POWERLEVEL9K_DISABLE_CONFIGURATION_WIZARD=true - -eval "$(task --completion zsh)" diff --git a/.devcontainer/Dockerfile b/.devcontainer/Dockerfile index 86ce818..56dc7f5 100644 --- a/.devcontainer/Dockerfile +++ b/.devcontainer/Dockerfile @@ -24,6 +24,18 @@ RUN npm install -g \ @semantic-release/github \ conventional-changelog-conventionalcommits -RUN git clone --depth=1 https://github.com/romkatv/powerlevel10k.git ${ZSH_CUSTOM}/themes/powerlevel10k - USER node + +# Use Powerlevel10k theme +RUN git clone --depth=1 https://github.com/romkatv/powerlevel10k.git /home/node/.powerlevel10k +RUN echo 'source /home/node/.powerlevel10k/powerlevel10k.zsh-theme' >> /home/node/.zshrc +RUN echo 'POWERLEVEL9K_DISABLE_CONFIGURATION_WIZARD=true' >> /home/node/.zshrc + +# Configure Powerlevel10k +RUN echo 'POWERLEVEL9K_LEFT_PROMPT_ELEMENTS=(dir vcs)' >> /home/node/.zshrc && \ + echo 'POWERLEVEL9K_RIGHT_PROMPT_ELEMENTS=(command_execution_time status)' >> /home/node/.zshrc && \ + echo 'POWERLEVEL9K_COMMAND_EXECUTION_TIME_THRESHOLD=0' >> /home/node/.zshrc && \ + echo 'POWERLEVEL9K_COMMAND_EXECUTION_TIME_PRECISION=2' >> /home/node/.zshrc && \ + echo 'POWERLEVEL9K_COMMAND_EXECUTION_TIME_FORMAT="duration"' >> /home/node/.zshrc + + diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json index ef7bea8..6bd155b 100644 --- a/.devcontainer/devcontainer.json +++ b/.devcontainer/devcontainer.json @@ -1,6 +1,9 @@ { "name": "TypeScript Development", "dockerFile": "Dockerfile", + "features": { + "ghcr.io/devcontainers/features/docker-in-docker:latest": {} + }, "customizations": { "vscode": { "extensions": [ @@ -37,19 +40,48 @@ "git.enableCommitSigning": true, "dev.containers.copyGitConfig": true, "files.insertFinalNewline": true, + "github.copilot.enable": { + "*": true + }, + "github.copilot.advanced": { + "authProvider": "github" + }, "github.copilot.chat.codeGeneration.useInstructionFiles": true, + "github.copilot.chat.codeGeneration.instructions": [ + { + "file": ".github/copilot-instructions.md" + }, + { + "file": "../README.md" + } + ], "github.copilot.chat.commitMessageGeneration.instructions": [ { "text": "Always use conventional commit message format." } - ] + ], + "mcp": { + "servers": { + "context7": { + "command": "docker", + "args": ["run", "-i", "--rm", "node:lts", "npx", "-y", "@upstash/context7-mcp@latest"] + }, + "github": { + "command": "docker", + "args": [ + "run", + "-i", + "--rm", + "-e", + "GITHUB_TOKEN=${localEnv:GITHUB_TOKEN}", + "mcp/github" + ] + } + } + } } } }, - "mounts": [ - "source=${localWorkspaceFolder}/.devcontainer/.zshrc,target=/home/node/.zshrc,type=bind,consistency=cached", - "source=${localWorkspaceFolder}/.devcontainer/launch.json,target=/home/node/.vscode/launch.json,type=bind,consistency=cached" - ], "remoteEnv": { "GITHUB_TOKEN": "${localEnv:GITHUB_TOKEN}" } diff --git a/.devcontainer/launch.json b/.devcontainer/launch.json deleted file mode 100644 index e913111..0000000 --- a/.devcontainer/launch.json +++ /dev/null @@ -1,4 +0,0 @@ -{ - "version": "0.2.0", - "configurations": [] -} diff --git a/.github/copilot-instructions.md b/.github/copilot-instructions.md index 14a6974..f910b7a 100644 --- a/.github/copilot-instructions.md +++ b/.github/copilot-instructions.md @@ -1,9 +1,15 @@ # Custom Instructions for Copilot -Today is May 22, 2025. +Today is June 8, 2025. - Always use Context7 to check for the latest updates, features, or best practices of a library relevant to the task at hand. -## Useful Links +## Related Repositories -- [Inference Gateway](https://github.com/inference-gateway/inference-gateway) +- [Inference Gateway](https://github.com/inference-gateway) + - [Inference Gateway UI](https://github.com/inference-gateway/ui) + - [Go SDK](https://github.com/inference-gateway/go-sdk) + - [Rust SDK](https://github.com/inference-gateway/rust-sdk) + - [TypeScript SDK](https://github.com/inference-gateway/typescript-sdk) + - [Python SDK](https://github.com/inference-gateway/python-sdk) + - [Google Calendar Agent](https://github.com/inference-gateway/google-calendar-agent) diff --git a/CHANGELOG.md b/CHANGELOG.md index 2c896f9..33616ef 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -2,6 +2,18 @@ All notable changes to this project will be documented in this file. +## [0.4.7](https://github.com/inference-gateway/docs/compare/0.4.6...0.4.7) (2025-06-08) + +### 📚 Documentation + +* Add A2A section ([83920ff](https://github.com/inference-gateway/docs/commit/83920ffb70537114ca5b40325911259b9ebf5c9c)) + +### 🔧 Miscellaneous + +* **devcontainer:** Add Docker-in-Docker feature to devcontainer configuration ([835041f](https://github.com/inference-gateway/docs/commit/835041f149c36f319ed02c6c522cd70506d15d02)) +* **devcontainer:** Add MCP server configurations for Context7 and GitHub ([5369733](https://github.com/inference-gateway/docs/commit/536973309218aab8447559ff082fd1ec35668f41)) +* **devcontainer:** Update Powerlevel10k configuration and remove unused files ([c027820](https://github.com/inference-gateway/docs/commit/c0278206c0455084122ac687c8fe695f4bee79db)) + ## [0.4.6](https://github.com/inference-gateway/docs/compare/0.4.5...0.4.6) (2025-05-25) ### 📚 Documentation diff --git a/app/a2a/A2AContent.tsx b/app/a2a/A2AContent.tsx new file mode 100644 index 0000000..1e9cf99 --- /dev/null +++ b/app/a2a/A2AContent.tsx @@ -0,0 +1,7 @@ +'use client'; + +import A2A from '@/markdown/a2a.mdx'; + +export default function A2AContent() { + return ; +} diff --git a/app/a2a/page.tsx b/app/a2a/page.tsx new file mode 100644 index 0000000..c174272 --- /dev/null +++ b/app/a2a/page.tsx @@ -0,0 +1,12 @@ +import type { Metadata } from 'next'; +import A2AContent from './A2AContent'; + +export const metadata: Metadata = { + title: 'Agent-To-Agent (A2A) Integration - Inference Gateway', + description: + 'Learn how to use Agent-To-Agent (A2A) integration in Inference Gateway to coordinate multiple specialized agents and extend LLM capabilities.', +}; + +export default function A2APage() { + return ; +} diff --git a/components/Sidebar.tsx b/components/Sidebar.tsx index ff9a2c1..363bde7 100644 --- a/components/Sidebar.tsx +++ b/components/Sidebar.tsx @@ -29,6 +29,10 @@ const sections = [ title: 'Model Context Protocol', items: [{ title: 'MCP Integration', href: '/mcp' }], }, + { + title: 'Agent-To-Agent (A2A)', + items: [{ title: 'A2A Integration', href: '/a2a' }], + }, { title: 'API Reference', items: [ diff --git a/markdown/a2a.mdx b/markdown/a2a.mdx new file mode 100644 index 0000000..0f0cbf6 --- /dev/null +++ b/markdown/a2a.mdx @@ -0,0 +1,520 @@ +# Agent-To-Agent (A2A) Integration + +The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces. + +## What is Agent-To-Agent (A2A)? + +Agent-To-Agent (A2A) is a protocol that enables LLMs to discover, communicate with, and coordinate multiple specialized agents simultaneously. Each agent can provide specific capabilities (called "skills") that the LLM can automatically discover and use to fulfill user requests. + +## Key Features + +- **🤖 Automatic Agent Discovery**: The LLM automatically discovers available agents and their capabilities +- **🔄 Multi-Agent Coordination**: Coordinate multiple agents in a single conversation +- **🛠️ Specialized Skills**: Each agent provides specialized tools for specific domains +- **🌐 Distributed Architecture**: Agents can run as separate services and scale independently +- **💬 Natural Language Integration**: Users interact naturally while the LLM handles agent coordination +- **🔌 Protocol Standardization**: Based on standardized A2A protocol for interoperability + +## How A2A Works + +```mermaid +%%{init: {'theme': 'base', 'themeVariables': { 'primaryColor': '#326CE5', 'primaryTextColor': '#fff', 'lineColor': '#5D8AA8', 'secondaryColor': '#006100' }, 'flowchart': { 'nodeSpacing': 50, 'rankSpacing': 80, 'curve': 'linear' }}}%% + +graph TD + A[👤 User Request] --> B[🧠 LLM] + B --> C[🔍 Discover
Available Agents] + C --> D[📋 Agent Registry] + D --> E[🤖 Agent Skills] + B --> F[🎯 Coordinate
Multiple Agents] + F --> G[📧 Agent 1:
Email] + F --> H[📅 Agent 2:
Calendar] + F --> I[🧮 Agent 3:
Calculator] + F --> J[🌤️ Agent 4:
Weather] + G --> K[📤 Results] + H --> K + I --> K + J --> K + K --> B + B --> L[✨ Unified Response] + + classDef user fill:#9370DB,stroke:#333,stroke-width:2px,color:white; + classDef llm fill:#326CE5,stroke:#fff,stroke-width:2px,color:white; + classDef agent fill:#32CD32,stroke:#333,stroke-width:2px,color:white; + classDef system fill:#F5A800,stroke:#333,stroke-width:2px,color:black; + + class A user; + class B,L llm; + class G,H,I,J agent; + class C,D,E,F,K system; +``` + +When a user makes a request: + +1. **Request Analysis**: The LLM analyzes the user's request +2. **Agent Discovery**: Available agents and their skills are discovered +3. **Task Decomposition**: The request is broken down into tasks for specific agents +4. **Agent Coordination**: Multiple agents are called simultaneously or sequentially +5. **Result Integration**: Results from all agents are integrated into a coherent response + +## Configuration + +### Environment Variables + +Enable A2A integration by setting these environment variables: + +```bash +# Enable A2A protocol support +A2A_ENABLE=true + +# Expose A2A endpoints for debugging and agent discovery +A2A_EXPOSE=true + +# Comma-separated list of A2A agent URLs +# Note: Replace with actual agent URLs in production +A2A_AGENTS=http://google-calendar-agent:3003 + +# Example with multiple agents (replace with real implementations): +# A2A_AGENTS=http://google-calendar-agent:3003,http://your-custom-agent:3004 + +# A2A client timeout (default: 30s) +A2A_CLIENT_TIMEOUT=30s +``` + +### Docker Compose Example + +> **Note**: This example shows how to configure multiple agents. Replace the example agents (helloworld-agent, calculator-agent, weather-agent) with your actual agent implementations. Only the Google Calendar Agent is currently available as a real implementation. + +```yaml +version: '3.8' +services: + inference-gateway: + image: ghcr.io/inference-gateway/inference-gateway:latest + ports: + - '8080:8080' + environment: + - A2A_ENABLE=true + - A2A_EXPOSE=true + # Update with your actual agent URLs + - A2A_AGENTS=http://google-calendar-agent:8084 + # Example with multiple agents: + # - A2A_AGENTS=http://google-calendar-agent:8084,http://your-custom-agent:8085 + - DEEPSEEK_API_KEY=${DEEPSEEK_API_KEY} + depends_on: + - google-calendar-agent + + # Google Calendar Agent - Real implementation + google-calendar-agent: + image: ghcr.io/inference-gateway/google-calendar-agent:latest + ports: + - '8084:8084' + environment: + - GOOGLE_CALENDAR_SA_JSON=${GOOGLE_CALENDAR_SA_JSON} + + # Example agents for demonstration (not real implementations) + # Replace these with your actual custom agents + + # helloworld-agent: + # image: your-registry/your-helloworld-agent:latest + # ports: + # - '8081:8081' + + # calculator-agent: + # image: your-registry/your-calculator-agent:latest + # ports: + # - '8082:8082' + + # weather-agent: + # image: your-registry/your-weather-agent:latest + # ports: + # - '8083:8083' +``` + +## Available Agents + +> **Note**: The following agents are provided as examples to demonstrate A2A capabilities. Currently, only the **Google Calendar Agent** is a fully implemented and available agent. The other agents (Calculator, Weather, Hello World) serve as examples for developers to understand the A2A protocol and can be used as templates for creating custom agents. + +### Google Calendar Agent + +The Google Calendar Agent provides comprehensive calendar management capabilities and is the **only fully implemented agent** available: + +**Skills:** + +- `list-events`: List calendar events for specified time periods +- `create-event`: Create new calendar events with natural language parsing +- `update-event`: Update existing calendar events +- `delete-event`: Delete calendar events + +**Features:** + +- Natural language date/time parsing ("tomorrow at 2pm", "next Monday") +- Smart attendee extraction ("meeting with John and Sarah") +- Location detection and parsing +- Google Calendar API integration with fallback to mock service + +### Calculator Agent (Example) + +_This is an example agent for demonstration purposes._ + +Provides mathematical computation capabilities: + +**Skills:** + +- `add`: Add two numbers together +- `subtract`: Subtract one number from another +- `multiply`: Multiply two numbers together +- `divide`: Divide one number by another + +### Weather Agent (Example) + +_This is an example agent for demonstration purposes._ + +Provides weather information services: + +**Skills:** + +- `get_weather`: Get current weather information for a location +- `get_forecast`: Get weather forecast for a location + +### Hello World Agent (Example) + +_This is an example agent for demonstration purposes._ + +A simple demonstration agent: + +**Skills:** + +- `hello_world`: Returns a simple greeting message + +## API Endpoints + +### List Available Agents + +Get information about all configured A2A agents and their capabilities: + +```bash +curl -X GET http://localhost:8080/v1/a2a/agents +``` + +**Response:** + +```json +{ + "agents": [ + { + "url": "http://calculator-agent:8082", + "capabilities": { + "skills": [ + { + "id": "add", + "name": "Add Numbers", + "description": "Add two numbers together" + }, + { + "id": "multiply", + "name": "Multiply Numbers", + "description": "Multiply two numbers together" + } + ] + } + }, + { + "url": "http://google-calendar-agent:8084", + "capabilities": { + "skills": [ + { + "id": "list-events", + "name": "List Calendar Events", + "description": "List calendar events for a specified time period" + }, + { + "id": "create-event", + "name": "Create Calendar Event", + "description": "Create new calendar events with natural language parsing" + } + ] + } + } + ] +} +``` + +## Usage Examples + +> **Note**: The following examples demonstrate A2A interactions. Replace the example agents with your actual implementations. Only interactions with the Google Calendar Agent will work with real functionality. + +### Single Agent Interaction + +Request a simple greeting from the Hello World agent (example): + +```bash +curl -X POST http://localhost:8080/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "deepseek/deepseek-chat", + "messages": [ + { + "role": "system", + "content": "You are a helpful assistant." + }, + { + "role": "user", + "content": "Say hello using the hello world agent." + } + ] +}' +``` + +### Mathematical Calculations (Example) + +Use the calculator agent for mathematical operations (example agent): + +```bash +curl -X POST http://localhost:8080/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "deepseek/deepseek-chat", + "messages": [ + { + "role": "system", + "content": "You are a helpful assistant with access to mathematical tools." + }, + { + "role": "user", + "content": "Calculate the result of 15 + 27 * 3." + } + ] +}' +``` + +### Weather Information (Example) + +Get current weather information using the weather agent (example agent): + +```bash +curl -X POST http://localhost:8080/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "deepseek/deepseek-chat", + "messages": [ + { + "role": "system", + "content": "You are a helpful assistant with access to weather information." + }, + { + "role": "user", + "content": "What is the current weather in New York?" + } + ] +}' +``` + +### Calendar Management (Real Implementation) + +Manage calendar events using the Google Calendar agent (this works with the actual implementation): + +```bash +curl -X POST http://localhost:8080/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "deepseek/deepseek-chat", + "messages": [ + { + "role": "system", + "content": "You are a helpful assistant with access to calendar management tools." + }, + { + "role": "user", + "content": "Schedule a team meeting tomorrow at 2 PM for 1 hour." + } + ] +}' +``` + +### Multi-Agent Coordination (Mixed Example) + +The real power of A2A comes from coordinating multiple agents in a single request. **Note**: This example mixes real (Google Calendar) and example agents: + +```bash +curl -X POST http://localhost:8080/v1/chat/completions \ + -H "Content-Type: application/json" \ + -d '{ + "model": "deepseek/deepseek-chat", + "messages": [ + { + "role": "system", + "content": "You are a helpful assistant with access to multiple specialized agents." + }, + { + "role": "user", + "content": "Check my calendar for today and schedule a meeting tomorrow at 2 PM." + } + ] +}' +``` + +In this example, the LLM will: + +1. Use the Google Calendar agent to check today's events (real functionality) +2. Use the Google Calendar agent to schedule the meeting (real functionality) + +For a fully functional multi-agent setup, you would need to implement additional custom agents based on your requirements. + +## Creating Custom Agents + +To create your own A2A-compatible agent, implement these endpoints: + +### Required Endpoints + +- `/.well-known/agent.json` - Agent capabilities and metadata +- `/a2a` - Main A2A protocol endpoint +- `/health` - Health check endpoint + +### Agent Capabilities Schema + +Your agent must expose its capabilities via the `/.well-known/agent.json` endpoint: + +```json +{ + "capabilities": { + "skills": [ + { + "id": "your-skill-id", + "name": "Your Skill Name", + "description": "Description of what your skill does" + } + ] + } +} +``` + +### A2A Protocol Implementation + +Implement the A2A protocol at the `/a2a` endpoint to handle: + +- `message/send` - Send a message and receive response +- `message/stream` - Send a streaming message +- `task/get` - Get task status (optional) +- `task/cancel` - Cancel a running task (optional) + +## Kubernetes Deployment + +Deploy A2A agents in Kubernetes environments: + +```yaml +apiVersion: apps/v1 +kind: Deployment +metadata: + name: calendar-agent + namespace: inference-gateway +spec: + replicas: 2 + selector: + matchLabels: + app: calendar-agent + template: + metadata: + labels: + app: calendar-agent + spec: + containers: + - name: calendar-agent + image: ghcr.io/inference-gateway/google-calendar-agent:latest + ports: + - containerPort: 8080 + env: + - name: GOOGLE_CALENDAR_SA_JSON + valueFrom: + secretKeyRef: + name: google-credentials + key: service-account.json +--- +apiVersion: v1 +kind: Service +metadata: + name: calendar-agent + namespace: inference-gateway +spec: + selector: + app: calendar-agent + ports: + - port: 8080 + targetPort: 8080 +``` + +## Best Practices + +### Agent Design + +- **Single Responsibility**: Each agent should focus on a specific domain +- **Stateless Operations**: Design agents to be stateless for better scalability +- **Error Handling**: Implement robust error handling and fallback mechanisms +- **Documentation**: Clearly document agent capabilities and expected inputs + +### Security + +- **Authentication**: Implement proper authentication between gateway and agents +- **Input Validation**: Validate all inputs to prevent injection attacks +- **Rate Limiting**: Implement rate limiting to prevent abuse +- **Network Security**: Use secure communication channels (HTTPS/TLS) + +### Performance + +- **Caching**: Implement caching for frequently accessed data +- **Timeouts**: Set appropriate timeouts for agent communications +- **Load Balancing**: Use load balancing for high-availability deployments +- **Monitoring**: Monitor agent performance and availability + +## Troubleshooting + +### Common Issues + +**Agents Not Discovered** + +- Verify `A2A_ENABLE=true` is set +- Check that agent URLs in `A2A_AGENTS` are accessible +- Ensure agents expose the `/.well-known/agent.json` endpoint + +**Connection Timeouts** + +- Adjust `A2A_CLIENT_TIMEOUT` setting +- Check network connectivity between gateway and agents +- Verify agent health endpoints are responding + +**Agent Communication Failures** + +- Check agent logs for errors +- Verify A2A protocol implementation in custom agents +- Test agent endpoints directly with curl + +### Debugging + +Enable debug logging for A2A operations: + +```bash +A2A_EXPOSE=true +LOG_LEVEL=debug +``` + +Then check agent discovery: + +```bash +curl -X GET http://localhost:8080/v1/a2a/agents +``` + +## Future Enhancements + +The A2A integration is continuously evolving. Planned enhancements include: + +- **Agent Marketplace**: Discover and install agents from a central marketplace +- **Advanced Routing**: Intelligent routing based on agent capabilities and load +- **Agent Composition**: Combine multiple agents into composite workflows +- **Enhanced Security**: Advanced authentication and authorization mechanisms +- **Monitoring Dashboard**: Visual monitoring and management of agent ecosystems + +## Related Resources + +- [Awesome A2A](https://github.com/inference-gateway/awesome-a2a) + - [Google Calendar Agent Repository](https://github.com/inference-gateway/google-calendar-agent) +- [A2A Protocol Specification](https://github.com/inference-gateway/inference-gateway/tree/main/a2a) +- [Example Agents](https://github.com/inference-gateway/inference-gateway/tree/main/examples/docker-compose/a2a) +- [Kubernetes Examples](https://github.com/inference-gateway/inference-gateway/tree/main/examples/kubernetes) diff --git a/markdown/main.mdx b/markdown/main.mdx index 37c7581..94015c6 100644 --- a/markdown/main.mdx +++ b/markdown/main.mdx @@ -15,6 +15,7 @@ of Mixture of Experts. - ⚙️ **Environment Configuration**: Easily configure API keys and URLs through environment variables. - 🔧 **Tool-use Support**: Enable function calling capabilities across supported providers with a unified API. - 🔌 **MCP Integration**: Seamlessly connect to Model Context Protocol servers for automatic tool discovery and execution. +- 🤖 **Agent-To-Agent (A2A)**: Coordinate multiple specialized agents to extend LLM capabilities with external services. - 🌊 **Streaming Responses**: Stream tokens in real-time as they're generated from language models. - 🐳 **Docker Support**: Use Docker and Docker Compose for easy setup and deployment. - ☸️ **Kubernetes Support**: Ready for deployment in Kubernetes environments. @@ -65,6 +66,27 @@ curl -X POST http://localhost:8080/v1/chat/completions \ Learn more about MCP Integration and explore our comprehensive examples. +## Agent-To-Agent (A2A) Integration + +Inference Gateway supports **Agent-To-Agent (A2A)** integration, enabling LLMs to coordinate with multiple specialized agents simultaneously. This powerful feature allows LLMs to: + +- **Coordinate multiple agents** in a single conversation +- **Access specialized services** like calendars, calculators, and weather APIs +- **Discover agent capabilities** automatically +- **Scale agent ecosystems** with distributed architecture + +```bash +# Enable A2A with multiple agents +export A2A_ENABLE=true +export A2A_AGENTS="http://calendar-agent:3001,http://calculator-agent:3002,http://weather-agent:3003" + +# LLMs can coordinate all agents in one request +curl -X POST http://localhost:8080/v1/chat/completions \ + -d '{"model": "deepseek-chat", "messages": [{"role": "user", "content": "Schedule a meeting, calculate costs, and check the weather"}]}' +``` + +Learn more about A2A Integration and see how to build your own agents. + ## Community Inference Gateway is an open-source project maintained by a growing community. diff --git a/public/search-index.json b/public/search-index.json index c34d5ca..ad3ab1c 100644 --- a/public/search-index.json +++ b/public/search-index.json @@ -1 +1 @@ -{"10":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"15":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"17":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"20":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"30":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"40":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"50":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"70":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"80":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"200":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"333":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"401":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"1000":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"8080":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8081":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8082":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"1741879542":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"api":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"reference":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"base":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"url":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"authentication":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"endpoints":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"list":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"all":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"models":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"provider":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"chat":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"completions":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"request":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"body":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"streaming":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"response":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"proxy":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"requests":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"example":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"openai":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"completion":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"health":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"check":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"error":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"responses":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"advanced":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"features":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"tool":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"use":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"direct":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"openapi":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"specification":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"inference":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"gateway":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"provides":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"restful":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"for":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interacting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"with":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"language":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"from":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"various":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"providers":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"this":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"documents":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"available":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"formats":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"and":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"structures":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"are":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"relative":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"to":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"your":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"installation":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"by":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"default":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"is":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"http":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"localhost":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"when":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"running":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"locally":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"if":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"enabled":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_authtrue":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"must":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"include":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"bearer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"token":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"authorization":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"your_jwt_token":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"the":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"jwt":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"issued":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"configured":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"identity":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"idp":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"as":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"specified":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"in":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"openid":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connect":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"settings":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"oidc_issuer_url":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"oidc_client_id":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"etc":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"validates":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"these":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"against":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"authenticate":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"get":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"of":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"across":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"v1models":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"status":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"ok":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"contenttype":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"applicationjson":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"object":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"data":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"id":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"gpt4":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"model":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"created":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"owned_by":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"served_by":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"claude3opus20240229":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"anthropic":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"meta":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"groq":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"specific":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"v1modelsprovider":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"where":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"one":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"cohere":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"cloudflare":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"ollama":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deepseek":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gpt3":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"5turbo":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"using":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"post":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"json":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"ollamadeepseekr1":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"5b":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"name":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"messages":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"array":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"conversation":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"role":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"system":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"user":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"assistant":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"message":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sender":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"content":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"hi":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"how":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"you":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"doing":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"today":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"stream":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"false":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"optional":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"they":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"re":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"generated":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"max_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"maximum":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"number":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"generate":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"tools":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"can":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"type":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"function":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"string":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"description":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"parameters":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"schema":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"defining":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"chatcmpl753":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"deepseekr1":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"choices":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"index":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"thinknokay":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"so":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"greeted":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"me":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"said":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"just":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"starting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"say":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"that":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"should":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"respond":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"friendly":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"way":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"nnmaybe":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"acknowledge":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"their":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"greeting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"offer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"my":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"help":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"something":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"since":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"mentioned":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"working":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"on":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"math":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"problems":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"or":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"solving":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"puzzles":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"ll":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"stick":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"nni":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"want":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"make":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"sure":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"approaching":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"it":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"right":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"not":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"question":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"yet":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"but":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"see":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"more":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"them":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"maybe":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"assistance":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"responding":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"an":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"emoji":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"like":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"would":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"be":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"nice":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"nthinknnhello":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"finish_reason":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"length":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"usage":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"prompt_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"completion_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"total_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"true":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"streamed":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"server":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"sent":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"events":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"sse":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"https":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"developer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"mozilla":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"objects":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"curl":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"helpful":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"architecture":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"overview":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"general":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"kubernetes":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"setup":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"document":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"highlevel":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"inferencegateway":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"designed":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"modular":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"extensible":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"allowing":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"easy":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"integration":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"new":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mermaid":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"init":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"theme":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"themevariables":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"primarycolor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"326ce5":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"primarytextcolor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"fff":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"linecolor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"5d8aa8":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"secondarycolor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"006100":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"fontfamily":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"arial":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"flowchart":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"nodespacing":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"rankspacing":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"padding":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"graph":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"td":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"client":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"nodes":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"clients":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"agents":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"v1chatcompletions":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"auth":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"node":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"oidc":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"ig1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ig2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ig3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"define":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"styles":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"classdef":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"fill":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"9370db":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"stroke":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"strokewidth":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"1px":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"color":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"white":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"f5a800":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"black":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"32cd32":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"apply":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"class":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"run":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"following":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"diagram":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"shows":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"external":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ingress":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"border":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"subgraph":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"cluster":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"ingressapi":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"service":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"internal":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalclients":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalagents":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"discovery":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"pod1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pod2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pod3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pods":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pod":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pg1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"end":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pg2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pg3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"monitoring":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"stack":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"servicemonitor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"metrics":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"sm":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"scrapes":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"prometheus":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"grafana":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"connection":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"externalproviders":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"placed":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"inside":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"k8s":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"visually":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"separate":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ext1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext4":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext5":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext6":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext7":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ffffff":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"externalsvc":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"monitor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"84a392":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalclient":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"flow":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"configuration":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"keycloak":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"prerequisites":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"setting":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"up":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"option":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"manual":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"configure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"environment":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"variables":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"configmap":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"secret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"helm":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"obtaining":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"access":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"password":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"grant":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"testing":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"only":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"credentials":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"servicetoservice":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"making":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"authenticated":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"selfsigned":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"certificates":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"create":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"keycloaks":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"ca":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"certificate":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"mount":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"best":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"practices":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"troubleshooting":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"common":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"issues":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"next":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"steps":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"supports":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"through":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"secure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"guide":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"focuses":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"popular":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"opensource":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"management":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"solution":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"valid":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"header":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"validated":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"users":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"applications":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"processed":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"otherwise":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"unauthorized":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"returned":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"enable":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"set":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"bash":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"oidc_issuer_urlhttps":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"section":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"detailed":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"integrating":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"www":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"org":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"v24":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"later":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"recommended":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"github":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"v0":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"kubectl":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"iodocstaskstools":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"deployment":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"shdocsintroinstall":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"chart":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"task":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"taskfile":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"devinstallation":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"tasks":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"complete":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"clone":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"repository":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"git":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"cd":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"deploy":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"infrastructure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"postgresql":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"deployinfrastructure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"admin":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"console":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"local":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"username":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"tempadmin":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"output":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"previous":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"command":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"test":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"fetchaccesstoken":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"localv1models":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"manually":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"follow":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"install":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"official":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"realm":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"log":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"click":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"enter":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"go":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"save":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"page":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"confidential":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"account":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"changes":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"tab":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"copy":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"add":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"email":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"userexample":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"com":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"verified":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"temporary":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"off":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"update":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"yaml":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"apiversion":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"v1":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"kind":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"metadata":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"namespace":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_auth":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"stringdata":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"oidc_client_secret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"yourclientsecret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"opaque":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"upgrade":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"createnamespace":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"envfrom":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"oci":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"ghcr":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"methods":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"variable":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"file":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"env":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"configmaps":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"secrets":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"import":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"configtable":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"flexible":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"options":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"adapt":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"needs":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"facilitate":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"apis":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"proper":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"essential":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"optimal":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"performance":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"security":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"multiple":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"suit":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"different":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"scenarios":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"most":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"deployments":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"kubernetesbased":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"files":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"development":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"primary":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"method":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"configuring":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"control":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"everything":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"basic":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"providerspecific":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"configurations":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"rows":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"defaultvalue":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"production":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_telemetry":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"opentelemetry":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tracing":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"exposes":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"endpoint":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"scraping":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"generates":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"distributed":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"traces":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"collected":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"collectors":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"issuer":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"core":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"behavior":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"server_host":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"host":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"server_port":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"port":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"server_read_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"read":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"30s":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"server_write_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"write":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"server_idle_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"idle":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"120s":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"server_tls_cert_path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"tls":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"server_tls_key_path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"strongly":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"pem":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connects":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"thirdparty":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"client_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connections":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"per":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"minimum":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"version":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"tls12":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"highthroughput":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"consider":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"increasing":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"pool":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"llm":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"at":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"plan":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"openai_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"comv1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"openai_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"anthropic_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"anthropic_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cohere_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cohere_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"groq_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"comopenaiv1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"groq_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ollama_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"8080v1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ollama_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cloudflare_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"ui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"required":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"doesnt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"exist":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"keys":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"creation":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"certmanager":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"horizontal":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"autoscaling":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"accessing":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"upgrading":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"uninstalling":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"remove":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"checking":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"services":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sources":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"explains":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"charts":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"installed":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"v3":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"there":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"two":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deploys":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"both":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"directly":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"registry":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"enabledtrue":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"before":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"deploying":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"may":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"need":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"doesn":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"dryrunclient":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"eof":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"annotations":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"shreleasename":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"inferencegatewayui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"shreleasenamespace":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"labels":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"app":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"iomanagedby":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"deepseek_api_key":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"yourdeepseekapikey":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"classnamenginx":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hosts":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hostui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"paths":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pathtypeprefix":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"certpathtotls":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"crt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"keypathtotls":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"we":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"recommend":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"io":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"automate":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"approach":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"automates":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"issuance":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"renewal":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"already":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"repo":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"jetstack":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"jetstackcertmanager":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"crds":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"clusterissuer":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"let":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"encrypt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"iov1":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"letsencryptprod":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"spec":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"acme":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"acmev02":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"letsencrypt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"orgdirectory":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"youremailexample":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"privatekeysecretref":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"solvers":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"http01":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"nginx":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hostyourdomain":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"yourdomain":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"will":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"automatically":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"manage":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"examples":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"docker":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"compose":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"environments":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"simple":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"minimal":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"image":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"latest":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ports":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"appsv1":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"replicas":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"selector":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"matchlabels":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"template":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"containers":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"containerport":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"valuefrom":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"secretkeyref":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"llmsecrets":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"openaiapikey":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"including":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"visit":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"openaigpt4":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"explain":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"works":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"compare":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"contrast":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"cases":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"out":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"full":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"directory":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"getting":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"started":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"learn":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"pull":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"rm":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"checkout":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"send":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"hello":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"world":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"integrated":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ides":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"vscode":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"cursor":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"fully":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"openaicompatible":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"favorite":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"extensions":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"standard":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sections":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"provide":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"instructions":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"continue":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"dev":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"extension":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"follows":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"marketplace":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"visualstudio":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"continuedev":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"open":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"title":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"yourmodelname":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"apibase":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"apikey":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"yourapikey":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"restart":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"sh":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"aifirst":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"code":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"editor":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"clicking":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"gear":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"icon":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"bottom":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"left":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"corner":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"keyboard":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"shortcut":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"cmd":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"mac":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"ctrl":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"windowslinux":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"sidebar":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"select":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"ai":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"scroll":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"down":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"find":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"custom":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"toggle":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"now":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"operations":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"other":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"aipowered":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"documentation":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"context":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"protocol":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"servers":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"llms":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"community":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"main":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"link":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"nextlink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"allows":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interact":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"unified":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"interface":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"simplifying":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"process":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"sending":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"receiving":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"enabling":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mixture":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"experts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"source":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"under":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mit":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"license":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"easily":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"urls":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"tooluse":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"support":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"calling":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"capabilities":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"supported":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"seamlessly":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"automatic":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"execution":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"realtime":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ready":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"analyze":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"built":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mind":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"configurable":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"timeouts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"lightweight":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"includes":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"libraries":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"runtime":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"resulting":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"smaller":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"size":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"binary":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"8mb":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"resource":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"consumption":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"consume":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"resources":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"have":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"lower":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"footprint":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"well":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"documented":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"guides":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"tested":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"extensively":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"unit":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"tests":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"maintained":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"actively":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"developed":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"scalable":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"used":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"hpa":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"compliance":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"privacy":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"project":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"does":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"collect":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"analytics":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"ensuring":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"selfhosted":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"over":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"try":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"our":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"href":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"gettingstarted":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"guidelink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"own":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"instance":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"minutes":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"acts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"intermediary":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"between":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"standardizing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"interactions":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"single":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"switch":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"without":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"changing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"application":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"implement":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"sophisticated":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"routing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"fallback":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mechanisms":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"centralize":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"policies":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"native":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"discover":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"connected":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"execute":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"calls":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"clientside":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"filesystems":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"databases":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"extend":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"integrations":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"export":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mcp_enabletrue":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_servers":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"filesystemserver":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8081mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"searchserver":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8082mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"search":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"recent":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"news":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"about":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"integrationlink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"explore":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"comprehensive":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"growing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"contributions":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"welcome":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"what":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"middleware":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"expose":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"debugging":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"commaseparated":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"types":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"filesystem":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"time":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"database":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"handling":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"considerations":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"inspector":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"included":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"logging":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tutorials":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"python":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"structure":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"failed":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"verify":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"network":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"connectivity":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"appearing":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"checks":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcpspecific":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_exposetrue":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"seamless":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"large":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"powerful":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"feature":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"discovers":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"requiring":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"individually":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"modelcontextprotocol":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"enables":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"securely":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"systems":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"query":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"manipulate":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"engines":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"retrieve":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"information":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"web":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"analysis":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"discovered":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"made":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"multiserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"simultaneously":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"dynamic":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"injection":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"injected":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"into":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"based":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"executed":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"transparently":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"results":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"zero":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"don":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"know":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"individual":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"builtin":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"observability":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tb":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"assembly":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"processing":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sends":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"added":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"decides":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"which":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"executes":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"via":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"result":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"delivery":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"timeserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8083mcp":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_dial_timeout5s":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_servershttp":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcptimeserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcpsearchserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"depends_on":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"values":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_enable":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_expose":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_client_timeout":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"10s":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_request_timeout":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"once":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"called":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"txt":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"collection":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"components":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"implementation":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"prebuilt":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"dashboards":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"robust":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"troubleshoot":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"integrate":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"logs":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"insights":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"telemetry":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_telemetrytrue":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"scraped":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"count":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"duration":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"rates":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"utilization":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"cpu":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"memory":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"exposing":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"timeseries":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"storing":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"visualization":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"platform":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"scrape":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"coreos":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"interval":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"15s":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"namespaceselector":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"matchnames":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"visualize":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"volume":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"latency":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"imported":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"outputs":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"structured":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"format":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"analyzed":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"elasticsearch":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"loki":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"software":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"kits":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"sdks":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"sdk":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"programming":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"languages":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"simplify":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"typed":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"interfaces":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"convenience":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"currently":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"div":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"classname":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"grid":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gridcols1":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"md":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gridcols2":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gap4":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"mb8":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"p4":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"roundedmd":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3pythonh3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"stronginstallation":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strong":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pip":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"stronggithub":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongpypi":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"pypi":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3typescripth3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"typescript":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"npm":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"inferencegatewaysdk":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongnpm":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"npmjs":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3goh3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3rusth3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"rust":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"cargo":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongcrate":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"crates":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"details":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"each":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3openaih3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"paccess":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gpt":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"tokenp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pstrongdefault":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comv1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3deepseekh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"puse":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"natural":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3anthropich3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pconnect":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"claude":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"highquality":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"conversational":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"xheaderp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3cohereh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3groqh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"highperformance":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"lpuaccelerated":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comopenaiv1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3cloudflareh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"workers":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comclientv4accountsp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"textsm":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"account_id":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"aip":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3ollamah3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"prun":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"none":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"8080v1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"requires":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"provider_api_url":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"provider_api_key":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"replace":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"uppercase":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"offers":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"approaches":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"consistent":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"model_name":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"also":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"8080v1models":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deepseekreasoner":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"capital":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"france":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"quantum":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"computing":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"terms":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"selection":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"screenshots":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"webbased":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interactively":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"exploring":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"userfriendly":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"experiment":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"view":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"writing":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"any":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"uses":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"backend":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ve":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"selected":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"choose":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"offered":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"filters":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"familiar":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"experience":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"prompt":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"input":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"area":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"markdown":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"formatting":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interactive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"reasoning":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"thought":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"statistics":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"clean":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"modern":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"design":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"light":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"dark":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mode":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"layout":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"responsive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"desktop":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mobile":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"devices":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"imageschatui":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"png":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"technologies":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"nextjs":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"react":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"component":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"safety":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"tailwind":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"css":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"styling":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"deployed":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"standalone":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"container":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"could":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"orchestration":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"such":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"self":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"building":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"provided":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"approx":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"300mb":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"js":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"extensive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"customize":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}]} \ No newline at end of file +{"10":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"15":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"17":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"20":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"30":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"40":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"50":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"70":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"80":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"200":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"333":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"401":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"1000":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"3003":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"3004":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"8080":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8081":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8082":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8084":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"8085":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"1741879542":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"agenttoagent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"a2a":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"integration":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"what":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"is":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"key":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"features":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"how":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"works":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"configuration":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"environment":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"variables":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"enable":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"protocol":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"support":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"expose":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"endpoints":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"for":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"debugging":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"and":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"agent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"discovery":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"commaseparated":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"list":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"of":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"urls":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"note":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"replace":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"with":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"actual":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"in":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"production":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"example":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"multiple":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"agents":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"real":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"implementations":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"client":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"timeout":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"default":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"30s":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"docker":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"compose":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"available":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"google":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"calendar":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"calculator":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"weather":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"hello":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"world":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"api":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"usage":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"examples":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"single":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"interaction":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"mathematical":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"calculations":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"information":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"management":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"implementation":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"multiagent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"coordination":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"mixed":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"creating":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"custom":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"required":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"capabilities":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"schema":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"kubernetes":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"deployment":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"best":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"practices":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"design":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"security":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"performance":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"troubleshooting":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"common":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"issues":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"future":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"enhancements":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"related":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"resources":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"the":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"inference":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"gateway":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"now":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"supports":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"enabling":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"large":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"language":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"models":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"llms":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"to":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"seamlessly":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"coordinate":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"external":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"specialized":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"this":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"powerful":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"feature":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"allows":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"access":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"utilize":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"wide":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"range":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"tools":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"services":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"through":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"standardized":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"interfaces":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"that":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"enables":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"discover":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"communicate":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"simultaneously":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"each":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"can":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"provide":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"specific":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"called":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"skills":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"llm":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"automatically":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"use":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"fulfill":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"user":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"requests":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"automatic":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"discovers":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"their":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"conversation":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"provides":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"domains":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"distributed":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"architecture":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"run":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"as":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"separate":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"scale":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"independently":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"natural":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"users":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interact":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"naturally":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"while":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"handles":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"standardization":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"based":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"on":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interoperability":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"mermaid":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"init":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"theme":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"base":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"themevariables":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"primarycolor":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"326ce5":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"primarytextcolor":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"fff":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"linecolor":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"5d8aa8":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"secondarycolor":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"006100":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"flowchart":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"nodespacing":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"rankspacing":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"curve":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"linear":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"graph":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"td":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"request":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"discoverbravailable":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"registry":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"coordinatebrmultiple":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"bremail":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"brcalendar":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"brcalculator":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"brweather":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"results":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"unified":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"response":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"classdef":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"fill":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"9370db":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"stroke":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"strokewidth":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"2px":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"color":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"white":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"32cd32":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"system":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"f5a800":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"black":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"class":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"when":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"makes":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"analysis":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"analyzes":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"are":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"discovered":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"task":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"decomposition":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"broken":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"down":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"into":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tasks":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"or":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"sequentially":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"result":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"from":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"all":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"integrated":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"coherent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"by":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"setting":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"these":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"bash":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"a2a_enabletrue":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"a2a_exposetrue":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"a2a_agentshttp":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"googlecalendaragent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"http":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"yourcustomagent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"shows":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"configure":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"helloworldagent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"calculatoragent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"weatheragent":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"your":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"only":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"currently":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"yaml":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"version":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"inferencegateway":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"image":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ghcr":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"latest":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ports":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"update":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"deepseek_api_key":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"depends_on":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"demonstration":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"}],"not":[{"title":"Agent-To-Agent (A2A) Integration","excerpt":"The Inference Gateway now supports **Agent-To-Agent (A2A)** integration, enabling Large Language Models (LLMs) to seamlessly coordinate with external specialized agents. This powerful feature allows LLMs to access and utilize a wide range of external tools and services through standardized agent interfaces.","url":"/a2a"},{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"reference":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"url":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"authentication":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"provider":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"chat":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"completions":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"body":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"streaming":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"proxy":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"openai":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"completion":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"health":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"check":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"error":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"responses":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"advanced":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"tool":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"direct":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"openapi":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"specification":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"restful":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"interacting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"various":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"providers":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"documents":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"formats":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"structures":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"relative":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"installation":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"localhost":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"running":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"locally":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"if":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"enabled":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_authtrue":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"must":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"include":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"bearer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"token":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"authorization":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"your_jwt_token":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"jwt":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"issued":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"configured":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"identity":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"idp":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"specified":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"openid":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connect":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"settings":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"oidc_issuer_url":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"oidc_client_id":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"etc":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"validates":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"against":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"authenticate":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"get":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"across":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"v1models":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"status":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"ok":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"contenttype":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"applicationjson":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"object":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"data":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"id":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"gpt4":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"model":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"created":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"owned_by":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"served_by":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"claude3opus20240229":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"anthropic":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"meta":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"groq":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"v1modelsprovider":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"where":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"one":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"cohere":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"cloudflare":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"ollama":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deepseek":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gpt3":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"5turbo":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"using":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"post":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"json":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"ollamadeepseekr1":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"5b":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"name":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"messages":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"array":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"role":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"assistant":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"message":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sender":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"content":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"hi":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"you":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"doing":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"today":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"stream":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"false":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"optional":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"they":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"re":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"generated":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"max_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"maximum":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"number":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"generate":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"type":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"function":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"string":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"description":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"parameters":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"defining":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"chatcmpl753":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"deepseekr1":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"choices":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"index":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"thinknokay":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"so":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"greeted":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"me":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"said":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"just":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"starting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"say":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"should":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"respond":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"friendly":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"way":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"nnmaybe":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"acknowledge":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"greeting":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"offer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"my":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"help":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"something":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"since":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"mentioned":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"working":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"math":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"problems":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"solving":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"puzzles":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"ll":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"stick":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"nni":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"want":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"make":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"sure":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"approaching":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"it":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"right":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"question":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"yet":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"but":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"see":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"more":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"them":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"maybe":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"assistance":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"responding":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"an":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"emoji":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"like":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"would":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"be":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"nice":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"nthinknnhello":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"finish_reason":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"length":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"prompt_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"completion_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"total_tokens":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"true":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"streamed":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"server":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"sent":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"events":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"sse":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"https":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"developer":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"mozilla":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"objects":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"}],"curl":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"helpful":[{"title":"API Reference","excerpt":"Inference Gateway provides a RESTful API for interacting with language models from various providers.","url":"/api-reference"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"overview":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"general":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"setup":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"document":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"highlevel":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"designed":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"modular":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"extensible":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"allowing":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"easy":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"new":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"fontfamily":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"arial":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"padding":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"nodes":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"clients":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"v1chatcompletions":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"auth":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"node":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"oidc":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"ig1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ig2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ig3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"h3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"define":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"styles":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"1px":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"apply":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"following":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"diagram":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ingress":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"border":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"subgraph":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"cluster":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"ingressapi":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"service":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"internal":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalclients":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalagents":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pod1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pod2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pod3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pods":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pod":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pg1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"end":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pg2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"pg3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"monitoring":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"stack":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"servicemonitor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"metrics":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"sm":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"scrapes":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"prometheus":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"grafana":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"connection":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"externalproviders":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"placed":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"inside":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"k8s":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"visually":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext1":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext2":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext3":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext4":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext5":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext6":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ext7":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"ffffff":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"externalsvc":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"monitor":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"84a392":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"internalclient":[{"title":"Architecture Overview","excerpt":"This document provides a high-level overview of the architecture of the Inference Gateway. The Inference-Gateway is designed to be modular and extensible, allowing easy integration of new models and providers.","url":"/architecture-overview"}],"flow":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"keycloak":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"prerequisites":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"up":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"option":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"manual":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"configmap":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"secret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"helm":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"obtaining":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"password":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"grant":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"testing":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"credentials":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"servicetoservice":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"making":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"authenticated":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"selfsigned":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"certificates":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"create":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"keycloaks":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"ca":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"certificate":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"mount":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"next":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"steps":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"secure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"guide":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"focuses":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"popular":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"opensource":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"solution":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"valid":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"header":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"validated":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"applications":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"processed":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"otherwise":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"unauthorized":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"returned":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"set":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"oidc_issuer_urlhttps":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"section":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"detailed":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"integrating":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"www":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"org":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"v24":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"later":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"recommended":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"github":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"v0":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"kubectl":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"iodocstaskstools":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"shdocsintroinstall":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"chart":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"taskfile":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"devinstallation":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"complete":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"clone":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"repository":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"git":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"cd":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"deploy":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"infrastructure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"postgresql":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"deployinfrastructure":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"admin":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"console":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"local":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"username":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"tempadmin":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"output":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"previous":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"command":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"test":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"fetchaccesstoken":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"localv1models":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"manually":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"follow":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"install":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"official":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"realm":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"log":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"click":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"enter":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"go":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"save":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"page":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"confidential":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"account":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"changes":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"tab":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"copy":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"add":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"email":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"userexample":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"com":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"verified":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"temporary":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"off":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"apiversion":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"v1":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"kind":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"metadata":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"namespace":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_auth":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"stringdata":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"oidc_client_secret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"yourclientsecret":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"}],"opaque":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"upgrade":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"createnamespace":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"envfrom":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"oci":[{"title":"Authentication","excerpt":"Inference Gateway supports authentication through OpenID Connect (OIDC), allowing you to secure your API with various identity providers. This guide focuses on setting up authentication with Keycloak, a popular open-source identity and access management solution.","url":"/authentication"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"methods":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"variable":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"file":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"env":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"configmaps":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"secrets":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"import":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"configtable":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"flexible":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"options":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"adapt":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"needs":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"facilitate":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"apis":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"proper":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"essential":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"optimal":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"suit":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"different":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"scenarios":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"most":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"deployments":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"kubernetesbased":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"files":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"development":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"primary":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"method":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"configuring":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"control":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"everything":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"basic":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"providerspecific":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"configurations":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"rows":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"defaultvalue":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"enable_telemetry":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"opentelemetry":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tracing":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"exposes":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"endpoint":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"scraping":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"generates":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"traces":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"collected":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"collectors":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"issuer":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"core":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"behavior":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"server_host":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"host":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"server_port":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"port":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"server_read_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"read":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"server_write_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"write":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"server_idle_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"idle":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"120s":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"server_tls_cert_path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"tls":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"server_tls_key_path":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"strongly":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"pem":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connects":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"thirdparty":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"client_timeout":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"connections":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"per":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"minimum":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"tls12":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"highthroughput":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"consider":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"increasing":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"pool":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"at":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"plan":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"openai_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"comv1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"openai_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"anthropic_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"anthropic_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cohere_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cohere_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"groq_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"comopenaiv1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"groq_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"ollama_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"8080v1":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ollama_api_key":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"cloudflare_api_url":[{"title":"Configuration","excerpt":"import ConfigTable from '../components/ConfigTable';","url":"/configuration"}],"ui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"doesnt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"exist":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"keys":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"creation":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"certmanager":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"horizontal":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"autoscaling":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"accessing":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"upgrading":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"uninstalling":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"remove":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"checking":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"sources":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"explains":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"charts":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"installed":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"v3":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"there":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"two":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deploys":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"both":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"directly":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"enabledtrue":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"before":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"deploying":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"may":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"need":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"doesn":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"dryrunclient":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"eof":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"annotations":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"shreleasename":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"inferencegatewayui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"shreleasenamespace":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"labels":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"app":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"iomanagedby":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"yourdeepseekapikey":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"classnamenginx":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hosts":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hostui":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"paths":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"pathtypeprefix":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"certpathtotls":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"crt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"keypathtotls":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"we":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"recommend":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"io":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"automate":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"approach":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"automates":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"issuance":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"renewal":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"already":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"repo":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"jetstack":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"jetstackcertmanager":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"crds":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"clusterissuer":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"let":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"encrypt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"iov1":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"letsencryptprod":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"spec":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"acme":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"acmev02":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"letsencrypt":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"orgdirectory":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"youremailexample":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"privatekeysecretref":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"solvers":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"http01":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"nginx":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"hostyourdomain":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"yourdomain":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"}],"will":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"manage":[{"title":"Deployment with Helm","excerpt":"This guide explains how to deploy the Inference Gateway using the official Helm charts.","url":"/deployment"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"environments":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"simple":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"minimal":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"appsv1":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"replicas":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"selector":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"matchlabels":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"template":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"containers":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"containerport":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"valuefrom":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"secretkeyref":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"llmsecrets":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"openaiapikey":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"including":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"visit":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"openaigpt4":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"explain":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"compare":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"contrast":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"cases":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"out":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"full":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"directory":[{"title":"Examples","excerpt":"This page provides examples of how to use Inference Gateway in various scenarios and environments.","url":"/examples"}],"getting":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"started":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"learn":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"pull":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"rm":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"checkout":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"send":[{"title":"Getting Started","excerpt":"Learn how to install and set up Inference Gateway.","url":"/getting-started"}],"ides":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"vscode":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"cursor":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"fully":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"openaicompatible":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"favorite":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"extensions":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"standard":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sections":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"instructions":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"continue":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"dev":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"extension":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"follows":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"marketplace":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"visualstudio":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"continuedev":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"open":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"title":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"yourmodelname":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"apibase":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"apikey":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"yourapikey":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"restart":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"sh":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"aifirst":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"code":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"editor":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"clicking":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"gear":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"icon":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"bottom":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"left":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"corner":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"keyboard":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"shortcut":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"cmd":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"mac":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"ctrl":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"windowslinux":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"sidebar":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"select":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"ai":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"scroll":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"find":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"toggle":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"operations":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"other":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"aipowered":[{"title":"Integrated Development Environments (IDEs)","excerpt":"Inference Gateway is fully OpenAI-compatible, allowing you to configure it with your favorite IDEs and extensions using standard OpenAI integration settings. The following sections provide instructions for configuring popular IDEs and extensions.","url":"/ides"}],"documentation":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"context":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"servers":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"community":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"main":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"link":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"nextlink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"interface":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"simplifying":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"process":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"sending":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"receiving":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mixture":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"experts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"source":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"under":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mit":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"license":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"easily":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"tooluse":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"calling":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"supported":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"execution":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"extend":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"realtime":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ready":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"analyze":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"built":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mind":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"configurable":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"timeouts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"lightweight":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"includes":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"libraries":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"runtime":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"resulting":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"smaller":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"size":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"binary":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"8mb":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"resource":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"consumption":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"consume":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"have":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"lower":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"footprint":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"well":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"documented":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"guides":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"tested":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"extensively":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"unit":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"tests":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"maintained":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"actively":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"developed":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"scalable":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"used":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"hpa":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"compliance":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"privacy":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"project":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"does":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"collect":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"analytics":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"ensuring":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"selfhosted":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"over":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"try":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"our":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"href":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"gettingstarted":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"guidelink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"own":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"instance":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"minutes":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"acts":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"intermediary":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"between":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"standardizing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"interactions":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"switch":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"without":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"changing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"application":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"implement":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"sophisticated":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"routing":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"fallback":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mechanisms":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"centralize":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"policies":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"native":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"connected":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"execute":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"calls":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"clientside":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"filesystems":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"databases":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"integrations":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"export":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"mcp_enabletrue":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_servers":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"filesystemserver":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8081mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"searchserver":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8082mcp":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"search":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"recent":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"news":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"about":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"integrationlink":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"explore":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"}],"comprehensive":[{"title":"Inference Gateway Documentation","excerpt":"Inference Gateway is a proxy server designed to facilitate access to various","url":"/"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"middleware":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"types":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"filesystem":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"time":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"database":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"handling":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"considerations":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"inspector":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"included":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"logging":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tutorials":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"python":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"structure":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"failed":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"verify":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"network":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"connectivity":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"appearing":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"checks":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcpspecific":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_exposetrue":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"seamless":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"requiring":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"individually":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"modelcontextprotocol":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"securely":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"systems":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"query":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"manipulate":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"engines":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"retrieve":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"web":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"made":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"multiserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"dynamic":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"injection":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"injected":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"executed":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"transparently":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"zero":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"don":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"know":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"individual":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"builtin":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"observability":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"tb":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"assembly":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"processing":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"sends":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"added":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"decides":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"which":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"executes":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"via":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"delivery":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"timeserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"8083mcp":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_dial_timeout5s":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_servershttp":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcptimeserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcpsearchserver":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"values":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_enable":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_expose":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_client_timeout":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"10s":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"mcp_request_timeout":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"once":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"txt":[{"title":"Model Context Protocol (MCP) Integration","excerpt":"The Inference Gateway supports **Model Context Protocol (MCP)** integration, enabling seamless access to external tools and data sources for Large Language Models (LLMs). This powerful feature automatically discovers and provides tools to LLMs without requiring clients to manage them individually.","url":"/mcp"}],"collection":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"components":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"prebuilt":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"dashboards":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"robust":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"troubleshoot":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"integrate":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"logs":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"insights":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"telemetry":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"enable_telemetrytrue":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"scraped":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"count":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"duration":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"rates":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"utilization":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"cpu":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"memory":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"exposing":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"timeseries":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"storing":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"visualization":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"platform":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"scrape":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"coreos":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"interval":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"15s":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"namespaceselector":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"matchnames":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"visualize":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"volume":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"latency":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"imported":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"outputs":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"structured":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"format":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"analyzed":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"elasticsearch":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"loki":[{"title":"Observability","excerpt":"Inference Gateway provides robust observability features to help monitor and troubleshoot your deployment. These features include metrics collection, tracing, and logging capabilities that integrate with popular monitoring tools.","url":"/observability"}],"software":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"kits":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"sdks":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"sdk":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"programming":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"languages":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"simplify":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"typed":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"convenience":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"div":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"classname":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"grid":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gridcols1":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"md":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gridcols2":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gap4":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"mb8":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"p4":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"roundedmd":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3pythonh3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"stronginstallation":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strong":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pip":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"stronggithub":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongpypi":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"pypi":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3typescripth3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"typescript":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"},{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"npm":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"inferencegatewaysdk":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongnpm":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"npmjs":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3goh3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"h3rusth3":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"rust":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"cargo":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"strongcrate":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"crates":[{"title":"Software Development Kits (SDK's)","excerpt":"Inference Gateway provides official SDKs for various programming languages to simplify integration with your applications. These SDKs offer typed interfaces, error handling, and convenience methods for interacting with the Inference Gateway API.","url":"/sdks"}],"details":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3openaih3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"paccess":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"gpt":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"tokenp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pstrongdefault":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comv1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3deepseekh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"puse":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3anthropich3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"pconnect":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"claude":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"highquality":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"conversational":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"xheaderp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3cohereh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3groqh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"highperformance":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"lpuaccelerated":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comopenaiv1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3cloudflareh3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"workers":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"comclientv4accountsp":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"textsm":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"account_id":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"aip":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"h3ollamah3":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"prun":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"none":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"8080v1p":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"requires":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"provider_api_url":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"provider_api_key":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"uppercase":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"offers":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"approaches":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"consistent":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"model_name":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"also":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"8080v1models":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"deepseekreasoner":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"capital":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"france":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"quantum":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"computing":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"terms":[{"title":"Supported Providers","excerpt":"Inference Gateway provides a unified interface to interact with multiple LLM providers.","url":"/supported-providers"}],"selection":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"screenshots":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"webbased":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interactively":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"exploring":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"userfriendly":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"experiment":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"view":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"writing":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"any":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"uses":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"backend":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"ve":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"selected":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"choose":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"offered":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"filters":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"familiar":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"experience":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"prompt":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"input":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"area":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"markdown":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"formatting":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"interactive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"reasoning":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"thought":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"statistics":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"clean":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"modern":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"light":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"dark":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mode":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"layout":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"responsive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"desktop":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"mobile":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"devices":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"imageschatui":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"png":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"technologies":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"nextjs":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"react":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"component":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"safety":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"tailwind":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"css":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"styling":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"deployed":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"standalone":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"container":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"could":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"orchestration":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"such":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"self":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"building":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"provided":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"approx":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"300mb":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"js":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"extensive":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}],"customize":[{"title":"User Interface (UI)","excerpt":"Inference Gateway includes an optional web-based user interface for interactively testing and exploring the various language model providers and capabilities.","url":"/ui"}]} \ No newline at end of file diff --git a/public/sitemap.xml b/public/sitemap.xml index 2cb71b8..d183b99 100644 --- a/public/sitemap.xml +++ b/public/sitemap.xml @@ -2,7 +2,7 @@ https://docs.inference-gateway.com - 2025-05-25T23:05:57.605Z + 2025-06-08T20:51:32.328Z weekly 1.0 @@ -12,6 +12,12 @@ weekly 1.0 + + https://docs.inference-gateway.com/a2a + 2025-06-08T21:07:28.428Z + monthly + 0.8 + https://docs.inference-gateway.com/api-reference 2025-04-15T17:06:58.869Z