🪐 MCP IPFS Server (storacha.network) 🛰️

🪐 MCP IPFS Server (storacha.network) 🛰️

By alexbakers GitHub

🪐 MCP IPFS Server

ai storage
Overview

What is MCP IPFS Server?

MCP IPFS Server is a Node.js server that implements the Model Context Protocol (MCP) for interacting with the storacha.network platform, allowing users to manage spaces, upload/download data, and perform various tasks using the w3 command-line interface.

How to use MCP IPFS Server?

To use the MCP IPFS Server, install the @web3-storage/w3cli globally, configure your environment, and run the server using Node.js or Docker. You can interact with it through MCP-compatible clients.

Key features of MCP IPFS Server?

  • Native integration with storacha.network via w3 CLI.
  • Comprehensive MCP tools for authentication, space management, data management, sharing, delegations, and account billing.
  • Supports advanced storage functionalities.

Use cases of MCP IPFS Server?

  1. Managing data storage on the storacha.network.
  2. Automating interactions with the w3 CLI for various tasks.
  3. Facilitating language model interactions with storage services.

FAQ from MCP IPFS Server?

  • What are the prerequisites for using MCP IPFS Server?
    You need Node.js version 22.0.0 or higher and the @web3-storage/w3cli installed globally.

  • Can I run MCP IPFS Server in Docker?
    Yes, you can build and run the server using Docker with the provided commands.

  • Is there a license for MCP IPFS Server?
    Yes, it is licensed under the MIT License.

Content

🪐 MCP IPFS Server (storacha.network) 🛰️

Screenshot

Publish Docker Publish NPM npm version smithery badge

A Node.js server implementing the Model Context Protocol (MCP) for interacting with the storacha.network platform via the w3 command-line interface (@web3-storage/w3cli).

This server empowers language models 🤖 and other MCP clients to manage storacha.network spaces, upload/download data, manage delegations, and perform various other tasks by seamlessly wrapping w3 commands.

✨ Features

  • Wraps the w3 CLI for native integration with storacha.network.
  • Provides MCP tools covering a wide range of w3 functionality:
    • 🔑 Authentication & Agent: w3_login, w3_reset, w3_account_ls (for checking authorization)
    • 📦 Space Management: w3_space_ls, w3_space_use, w3_space_info, w3_space_add, w3_space_provision (Note: w3_space_create must be run manually due to interactive prompts)
    • 💾 Data Management: w3_up, w3_ls, w3_rm
    • 🔗 Sharing: w3_open (generates w3s.link URL)
    • 🤝 Delegations & Proofs: w3_delegation_create, w3_delegation_ls, w3_delegation_revoke, w3_proof_add, w3_proof_ls
    • 🔐 Keys & Tokens: w3_key_create, w3_bridge_generate_tokens
    • ⚙️ Advanced Storage (w3 can ...): Blob, CAR, Upload, Index, Access Claim, Filecoin Info management
    • 💳 Account & Billing: w3_plan_get, w3_coupon_create, w3_usage_report

🛠️ Prerequisites

  • Node.js: Version 22.0.0 or higher (node -v).
  • w3 CLI: The server executes w3 commands directly. Ensure @web3-storage/w3cli is installed globally and configured:
    npm install -g @web3-storage/w3cli
    w3 login <your-email@example.com>
    # Follow email verification steps
    
  • Environment Variable: The w3_login tool requires the W3_LOGIN_EMAIL environment variable to be set to the same email used for w3 login.

🏗️ Project Structure

The codebase is organized as follows:

src/
├── index.ts          # Main server entry point, MCP setup, request routing
├── schemas.ts        # Zod schemas defining input arguments for each tool
├── tool_handlers.ts  # Implementation logic for each MCP tool
├── utils.ts          # Helper functions (e.g., running w3 commands, parsing JSON)
└── utils/
    └── logger.ts     # Basic logger configuration

🚀 Usage with MCP Clients

This server can be used with any MCP-compatible client. You need to configure your client to connect to this server.

This assumes npm and the prerequisites are met.

{
  "mcpServers": {
    "ipfs": {
      "command": "npx",
      "args": ["-y", "mcp-ipfs"],
      "env": {
        "W3_LOGIN_EMAIL": "your-email@example.com"
      }
    }
  }
}

Example: Docker

Build the image first (see Build section) or use the pre-built image alexbakers/mcp-ipfs.

{
  "mcpServers": {
    "mcp-ipfs": {
      "command": "docker",
      "args": [
        "run",
        "-i",
        "--rm",
        "-v",
        "/path/to/your/project:/path/to/your/project",
        "-e",
        "W3_LOGIN_EMAIL",
        "alexbakers/mcp-ipfs"
      ],
      "env": {
        "W3_LOGIN_EMAIL": "your-email@example.com"
      }
    }
  }
}

📝 Note on Paths:

Several w3 commands require absolute filesystem paths (e.g., w3_up, w3_delegation_create --output, w3_proof_add, w3_can_blob_add, w3_can_store_add).

  • NPX: Provide absolute paths from your host machine.
  • Docker: Provide absolute paths inside the container. If interacting with files from your host (e.g., uploading), you must mount the relevant host directory into the container using the -v flag (e.g., -v /Users/me/project:/Users/me/project) and then use the container path (e.g., /Users/me/project/my_file.txt) in the tool arguments.

📦 Build

Clone the repository and install dependencies:

git clone https://github.com/alexbakers/mcp-ipfs.git
cd mcp-ipfs
npm install

Build the TypeScript code:

npm run build

You can then run the server directly:

# Ensure W3_LOGIN_EMAIL is set in your environment
export W3_LOGIN_EMAIL="your-email@example.com"
node dist/index.js

Or publish it (if you have the rights):

npm publish

🐳 Docker Build

Build the Docker image:

# Build locally (replace with your username/repo and desired tag)
docker build -t alexbakers/mcp-ipfs .

📜 License

This MCP server is licensed under the MIT License. See the LICENSE file for details.

No tools information available.
FridayAI
FridayAI by VedantRGosavi

AI-gaming companion to help with quests.

ai gaming
View Details

mcp-use is the easiest way to interact with mcp servers with custom agents

Nerve
Nerve by evilsocket

The Simple Agent Development Kit.

Deep-Co
Deep-Co by succlz123

A Chat Client for LLMs, written in Compose Multiplatform.

MCP server that allows AI tools to interact with the CheerLights API.

LibraryAI
LibraryAI by YELANDAOKONG

[C#/AI] (LibraryAI/ConsoleAI) AI 知识库与数据处理程序