OmniLLM: Universal LLM Bridge for Claude

OmniLLM: Universal LLM Bridge for Claude

By sabpap GitHub

OmniLLM: A Model Context Protocol (MCP) server that enables Claude to access and integrate responses from multiple LLMs including ChatGPT, Azure OpenAI, and Google Gemini, creating a unified AI knowledge hub.

omnillm llm-bridge
Overview

what is OmniLLM?

OmniLLM is a Model Context Protocol (MCP) server that enables Claude to access and integrate responses from multiple large language models (LLMs) including ChatGPT, Azure OpenAI, and Google Gemini, creating a unified AI knowledge hub.

how to use OmniLLM?

To use OmniLLM, set up the server by installing the necessary dependencies, configuring your API keys, and integrating it with the Claude Desktop application. Once set up, you can query different LLMs through Claude.

key features of OmniLLM?

  • Query OpenAI's ChatGPT models
  • Query Azure OpenAI services
  • Query Google's Gemini models
  • Get responses from all LLMs for comparison
  • Check which LLM services are configured and available

use cases of OmniLLM?

  1. Comparing responses from different LLMs for better insights.
  2. Enhancing Claude's responses by leveraging multiple AI models.
  3. Accessing diverse AI knowledge for various queries in one place.

FAQ from OmniLLM?

  • What LLMs can I query with OmniLLM?

You can query ChatGPT, Azure OpenAI, and Google Gemini models.

  • Do I need API keys for all LLMs?

You only need API keys for the services you want to use.

  • Is OmniLLM free to use?

The usage of OmniLLM is free, but you may incur costs based on the LLM services you access.

Content

OmniLLM: Universal LLM Bridge for Claude

OmniLLM is an MCP server that allows Claude to query and integrate responses from other large language models (LLMs) like ChatGPT, Azure OpenAI, and Google Gemini, creating a unified access point for all your AI needs.

Features

  • Query OpenAI's ChatGPT models
  • Query Azure OpenAI services
  • Query Google's Gemini models
  • Get responses from all LLMs for comparison
  • Check which LLM services are configured and available

Setup Instructions

1. Prerequisites

  • Python 3.10+
  • Claude Desktop application
  • API keys for the LLMs you want to use

2. Installation

# Clone or download this repository
git clone https://github.com/yourusername/omnillm-mcp.git
cd omnillm-mcp

# Create virtual environment
python -m venv venv
source venv/bin/activate  # On Windows: venv\Scripts\activate

# Install dependencies
pip install mcp[cli] httpx python-dotenv

3. Configuration

Create a .env file in the project root with your API keys:

OPENAI_API_KEY=your_openai_key_here
AZURE_OPENAI_API_KEY=your_azure_key_here
AZURE_OPENAI_ENDPOINT=your_azure_endpoint_here
GOOGLE_API_KEY=your_google_api_key_here

You only need to add the keys for the services you want to use.

4. Integrating with Claude Desktop

  1. Open Claude Desktop
  2. Navigate to Settings > Developer > Edit Config
  3. Add the server to your claude_desktop_config.json file:
{
  "mcpServers": {
    "omnillm": {
      "command": "python",
      "args": [
        "path/to/server.py"
      ],
      "env": {
        "PYTHONPATH": "path/to/omnillm-mcp"
      }
    }
  }
}

Replace "path/to/server.py" with the actual path to your server.py file.

  1. Save the config file and restart Claude Desktop

Usage Examples

Once connected to Claude Desktop, you can use phrases like:

  • "What would be the top places to visit if you're looking for an adventurous hiking trip? Consult ChatGPT"
  • "What's the best way to learn programming? Ask Gemini for their opinion."
  • "Compare different frameworks for building web applications, and then get input from both ChatGPT and Azure OpenAI"

Claude will automatically detect when to use the Multi-LLM Proxy tools to enhance its responses.

Available Tools

  1. query_chatgpt - Query OpenAI's ChatGPT with a custom prompt
  2. query_azure_chatgpt - Query Azure OpenAI's ChatGPT with a custom prompt
  3. query_gemini - Query Google's Gemini with a custom prompt
  4. query_all_llms - Query all available LLMs and get all responses together
  5. check_available_models - Check which LLM APIs are properly configured

Troubleshooting

  • Check that your API keys are correctly set in the .env file
  • Ensure Claude Desktop is properly configured with the server path
  • Verify that all dependencies are installed in your virtual environment
  • Check Claude's logs for any connection or execution errors

License

MIT License

No tools information available.
No content found.