Thanks to visit codestin.com
Credit goes to github.com

Skip to content

NilayYadav/mcpify

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

mcpify

Automatically generate MCP (Model Context Protocol) servers by observing HTTP API traffic.

Demo

Demo

What it does

mcpify watches HTTP requests to your server and automatically creates MCP tools for each discovered endpoint. AI assistants can then call these tools to interact with your APIs.

Installation

brew tap nilayyadav/mcpify
brew install mcpify

Quick Start

  1. Start your API server (e.g., on localhost:3000)
  2. Run mcpify to observe traffic:
sudo mcpify --target http://localhost:3000
  1. Make API calls to your server (using your app, curl, Postman, etc.)
  2. Each unique endpoint becomes available as an MCP tool at http://localhost:8081/mcp

Persistent Configuration

mcpify automatically saves discovered tools and configuration:

  • Linux: ~/.config/mcpify/config.json
  • macOS: ~/Library/Application Support/mcpify/config.json

Discovered tools persist across restarts. If you run mcpify without --target, it will use the last observed server.

# First run - discovers and saves tools
sudo mcpify --target http://localhost:3000

# Later runs - automatically uses saved target and loads existing tools
sudo mcpify

Grouping Feature

mcpify can now automatically group related API endpoints into logical tool groups. This makes it easier for AI assistants to understand and interact with your API by organizing endpoints by resource or functionality (e.g., all /users endpoints are grouped together).

How Grouping Works

  • Endpoints are analyzed and grouped based on URL patterns and HTTP methods.
  • Each group is exposed as a collection of related tools in the MCP server.
  • Grouping improves discoverability and usability for large APIs.

Grouping is enabled by default. You can control grouping behavior with the following command line flag:

sudo mcpify --target http://localhost:3000 --grouping

Grouped tools are available at http://localhost:8081/mcp as usual, but now organized by group.

Configuration

Environment Variables

export LLM="openai/gpt-oss-120b:together"
export LLM_ENDPOINT="https://router.huggingface.co/v1"
export LLM_API_KEY="HF_TOKEN"

Command Line Options

sudo -E mcpify --target http://localhost:3000 \
       --max-tools 100 \
       --use-llm \
       --verbose
Flag Description Default
--target Target server URL to observe (uses saved target if omitted) -
--mcp-port MCP server port 8081
--mcp-name Name of the MCP server mcpify
--max-tools Maximum number of tools to capture 100
--use-llm Enable LLM for tool name generation false
--verbose Enable verbose logging false
--grouping Enable grouping of related API endpoints true

Requirements

  • macOS or Linux
  • Root/sudo privileges (for packet capture)
  • Target server running on HTTP (not HTTPS)

MCP Integration

Connect AI assistants to http://localhost:8081/mcp to access auto-generated tools.

License

MIT

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages