Sponsored by Deepsite.site

mcp-server-deepseek

Created By
tizee9 months ago
A MCP server provides access to DeepSeek-R1's reasoning capabilities for LLMs
Content

mcp-server-deepseek

A Model Context Protocol (MCP) server that provides access to DeepSeek-R1's reasoning capabilities, allowing non-reasoning models to generate better responses with enhanced thinking.

Overview

This server acts as a bridge between LLM applications and DeepSeek's reasoning capabilities. It exposes DeepSeek-R1's reasoning content through an MCP tool, which can be used by any MCP-compatible client.

The server is particularly useful for:

  • Enhancing responses from models without native reasoning capabilities
  • Accessing DeepSeek-R1's thinking process for complex problem solving
  • Adding structured reasoning to Claude or other LLMs that support MCP

Features

  • Access to DeepSeek-R1: Connects to DeepSeek's API to leverage their reasoning model
  • Structured Thinking: Returns reasoning in a structured <thinking> format
  • Integration with MCP: Fully compatible with the Model Context Protocol
  • Error Handling: Robust error handling with detailed logging

Installation

Prerequisites

  • Python 3.13 or higher
  • An API key for DeepSeek

Setup

  1. Clone the repository:

    git clone https://github.com/yourusername/mcp-server-deepseek.git
    cd mcp-server-deepseek
    
  2. Create a virtual environment:

    python -m venv venv
    source venv/bin/activate  # On Windows: venv\Scripts\activate
    
  3. Install the package:

    pip install -e .
    
  4. Create a .env file with your DeepSeek API credentials:

    cp .env.example .env
    
  5. Edit the .env file with your API key and model details:

    MCP_SERVER_DEEPSEEK_MODEL_NAME=deepseek-reasoner
    MCP_SERVER_DEEPSEEK_API_KEY=your_api_key_here
    MCP_SERVER_DEEPSEEK_API_BASE_URL=https://api.deepseek.com
    

Usage

Running the Server

You can run the server directly:

mcp-server-deepseek

Or use the development mode with the MCP Inspector:

make dev

MCP Tool

The server exposes a single tool:

think_with_deepseek_r1

This tool sends a prompt to DeepSeek-R1 and returns its reasoning content.

Arguments:

  • prompt (string): The full user prompt to process

Returns:

  • String containing DeepSeek-R1's reasoning wrapped in <thinking> tags

Example Usage

When used with Claude or another LLM that supports MCP, you can trigger the thinking process by calling the tool:

Please use the think_with_deepseek_r1 tool with the following prompt:
"How can I optimize a neural network for time series forecasting?"

Development

Testing

For development and testing, use the MCP Inspector:

npx @modelcontextprotocol/inspector uv run mcp-server-deepseek

Logging

Logs are stored in ~/.cache/mcp-server-deepseek/server.log

The log level can be configured using the LOG_LEVEL environment variable (defaults to DEBUG).

Troubleshooting

Common Issues

  • API Key Issues: Ensure your DeepSeek API key is correctly set in the .env file
  • Timeout Errors: Complex prompts may cause timeouts. Try simplifying your prompt
  • Missing Reasoning: Some queries might not generate reasoning content. Try rephrasing

Error Logs

Check the logs for detailed error messages:

cat ~/.cache/mcp-server-deepseek/server.log

License

MIT

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

Acknowledgements

  • Thanks to the DeepSeek team for their powerful reasoning model
  • Built with the Model Context Protocol framework
Recommend Servers
TraeBuild with Free GPT-4.1 & Claude 3.7. Fully MCP-Ready.
Zhipu Web SearchZhipu Web Search MCP Server is a search engine specifically designed for large models. It integrates four search engines, allowing users to flexibly compare and switch between them. Building upon the web crawling and ranking capabilities of traditional search engines, it enhances intent recognition capabilities, returning results more suitable for large model processing (such as webpage titles, URLs, summaries, site names, site icons, etc.). This helps AI applications achieve "dynamic knowledge acquisition" and "precise scenario adaptation" capabilities.
AiimagemultistyleA Model Context Protocol (MCP) server for image generation and manipulation using fal.ai's Stable Diffusion model.
MiniMax MCPOfficial MiniMax Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech, image generation and video generation APIs.
ChatWiseThe second fastest AI chatbot™
Serper MCP ServerA Serper MCP Server
Context7Context7 MCP Server -- Up-to-date code documentation for LLMs and AI code editors
Visual Studio Code - Open Source ("Code - OSS")Visual Studio Code
DeepChatYour AI Partner on Desktop
BlenderBlenderMCP connects Blender to Claude AI through the Model Context Protocol (MCP), allowing Claude to directly interact with and control Blender. This integration enables prompt assisted 3D modeling, scene creation, and manipulation.
TimeA Model Context Protocol server that provides time and timezone conversion capabilities. This server enables LLMs to get current time information and perform timezone conversions using IANA timezone names, with automatic system timezone detection.
MCP AdvisorMCP Advisor & Installation - Use the right MCP server for your needs
Amap Maps高德地图官方 MCP Server
Baidu Map百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
CursorThe AI Code Editor
Howtocook Mcp基于Anduin2017 / HowToCook (程序员在家做饭指南)的mcp server,帮你推荐菜谱、规划膳食,解决“今天吃什么“的世纪难题; Based on Anduin2017/HowToCook (Programmer's Guide to Cooking at Home), MCP Server helps you recommend recipes, plan meals, and solve the century old problem of "what to eat today"
WindsurfThe new purpose-built IDE to harness magic
EdgeOne Pages MCPAn MCP service designed for deploying HTML content to EdgeOne Pages and obtaining an accessible public URL.
Tavily Mcp
Playwright McpPlaywright MCP server
Jina AI MCP ToolsA Model Context Protocol (MCP) server that integrates with Jina AI Search Foundation APIs.