
Agentic MCP Client
A standalone agent runner that executes tasks using MCP (Model Context Protocol) tools via Anthropic Claude, AWS BedRock and OpenAI APIs. It enables AI agents to run autonomously in cloud environments and interact with various systems securely.
What is Agentic MCP Client?
Agentic MCP Client is a standalone agent runner that executes tasks using Model Context Protocol (MCP) tools via Anthropic Claude, AWS BedRock, and OpenAI APIs. It allows AI agents to operate autonomously in cloud environments and interact securely with various systems.
How to use Agentic MCP Client?
To use the Agentic MCP Client, clone the repository, set up dependencies, create a JSON configuration file for the agent's tasks, and run the agent using the provided command.
Key features of Agentic MCP Client?
- Run standalone agents with tasks defined in JSON configuration files.
- Support for both Anthropic Claude and OpenAI models.
- Automatic tool discovery and execution.
- Session logging for tracking agent progress.
- Configurable system prompts and model parameters.
- Support for MCP sampling and maximum iteration control to prevent infinite loops.
Use cases of Agentic MCP Client?
- Automating tasks in cloud environments.
- Interacting with various systems securely using AI agents.
- Executing complex workflows through the integration of multiple tools.
FAQ from Agentic MCP Client?
- Can I use Agentic MCP Client with any cloud provider?
Yes! It supports AWS BedRock and can be configured for other providers as well.
- Is there a limit to the number of tasks I can run?
The limit is primarily based on the configuration and resources available in your cloud environment.
- How do I contribute to the project?
You can fork the repository, create a new branch, make your changes, and submit a pull request.
Agentic MCP Client
A standalone agent runner that executes tasks using MCP (Model Context Protocol) tools via Anthropic Claude, AWS BedRock and OpenAI APIs. It enables AI agents to run autonomously in cloud environments and interact with various systems securely.
Current Features
- Run standalone agents with tasks defined in JSON configuration files
- Support for both Anthropic Claude and OpenAI models
- Automatic tool discovery and execution
- Session logging for tracking agent progress
- Configurable system prompts and model parameters
- Support for MCP sampling
- Maximum iteration control to prevent infinite loops
Installation
-
Clone the repository
-
Set up dependencies:
uv sync
- Create an agent_worker_task.json file
Here is an example configuration file:
{
"task": "Find all image files in the current directory and tell me their sizes",
"model": "claude-3-7-sonnet-20250219",
"system_prompt": "You are a helpful assistant that completes tasks using available tools.",
"verbose": true,
"max_iterations": 10
}
- Run the agent:
uv run agentic_mcp_client/agent_worker/run.py
Configuration
The project requires a config.json
file in the root directory to define the inference server settings and available MCP tools. Here's an example configuration:
{
"inference_server": {
"base_url": "https://api.anthropic.com/v1/",
"api_key": "YOUR_API_KEY_HERE",
"use_bedrock": true,
"aws_region": "us-east-1",
"aws_access_key_id": "YOUR_AWS_ACCESS_KEY",
"aws_secret_access_key": "YOUR_AWS_SECRET_KEY"
},
"mcp_servers": {
"mcp-remote-macos-use": {
"command": "docker",
"args": [
"run",
"-i",
"-e",
"MACOS_USERNAME=your_username",
"-e",
"MACOS_PASSWORD=your_password",
"-e",
"MACOS_HOST=your_host_ip",
"--rm",
"buryhuang/mcp-remote-macos-use:latest"
]
},
"mcp-my-apple-remembers": {
"command": "docker",
"args": [
"run",
"-i",
"-e",
"MACOS_USERNAME=your_username",
"-e",
"MACOS_PASSWORD=your_password",
"-e",
"MACOS_HOST=your_host_ip",
"--rm",
"buryhuang/mcp-my-apple-remembers:latest"
]
}
}
}
Configuration Sections
Inference Server
The inference_server
section configures the connection to your language model provider:
base_url
: The API endpoint for your chosen LLM providerapi_key
: Your authentication key for the LLM serviceuse_bedrock
: Set to true to use Amazon Bedrock for model inference- AWS credentials (when using Bedrock)
MCP Servers
The mcp_servers
section defines available MCP tools. Each tool has:
- A unique identifier (e.g., "mcp-remote-macos-use")
command
: The command to execute (typically Docker for containerized tools)args
: Configuration parameters for the tool
This example shows MCP tools for remotely controlling a macOS system through Docker containers.
How MCP Works
The Model Context Protocol provides a standardized way for applications to:
- Share contextual information with language models
- Expose tools and capabilities to AI systems
- Build composable integrations and workflows
The protocol uses JSON-RPC 2.0 messages to establish communication between hosts (LLM applications), clients (connectors within applications), and servers (services providing context and capabilities).
Our agent worker implements this workflow:
- Initialize MCP clients for all available tools
- Send the initial task message to the selected model
- Process model responses (either tool calls or text)
- If a tool call is made, execute the tool and send the result back to the model
- Repeat until the task is completed or maximum iterations reached
- Shut down all MCP clients
sequenceDiagram
participant User
participant AgentWorker
participant LLM as Language Model
participant MCP as MCP Tools
User->>AgentWorker: Task + Configuration
AgentWorker->>MCP: Initialize Tools
AgentWorker->>LLM: Send Task
loop Until completion
LLM->>AgentWorker: Request Tool Use
AgentWorker->>MCP: Execute Tool
MCP->>AgentWorker: Tool Result
AgentWorker->>LLM: Send Tool Result
LLM->>AgentWorker: Response
end
AgentWorker->>User: Final Result
Contribution Guidelines
Contributions to Agentic MCP Client are welcome! To contribute, please follow these steps:
- Fork the repository.
- Create a new branch for your feature or bug fix.
- Make your changes and commit them.
- Push your changes to your fork.
- Create a pull request to the main repository.
Acknowledgments
This project was inspired by and builds upon the work the excellent open-source projects in the MCP ecosystem:
- MCP-Bridge - A middleware that provides an OpenAI-compatible endpoint for calling MCP tools, which helped inform our approach to tool integration and standardization.
We are grateful to the contributors of these projects for their pioneering work in the MCP space, which has helped make autonomous agent development more accessible and powerful.
License
Agentic MCP Client is licensed under the Apache 2.0 License. See the LICENSE file for more information.