- Transcription MCP Sever
Transcription MCP Sever
Content
Transcription MCP Sever
An MCP (Model-Context-Protocol) server for transcribing MP3 audio files using the AssemblyAI API. This server provides command-line and Docker-based tools to transcribe audio. this application can be use by Podcasters, Content Creator, Educators and Business Teams making thier task more easier.
Features
- Accepts .mp3 files via command-line interface using Fire
- Accepts .mp3 file path provided via an AI assistant such as Claude or Cursor.
- Transcribe MP3 audio files into JSON transcripts using AssemblyAI API
- Speaker diarization (speaker labels) enabled
- Save transcripts in a local "transcripts/" directory
- Easy-to-use command-line interface
- ready to run in a Docker container for portability and deployment
System Architecture Overview
This application is a microservice designed to transcribe audio files using AssemblyAI.
Architecture Breakdown
- Uvicorn as MCP Server Runtime: Uvicorn runs the server process, acting as the entry point for executing transcription jobs.
- AssemblyAI Client: This component handles all communication with the AssemblyAI API.
- CLI via Google Fire: The application includes a robust command-line interface built with Google Fire. Users can transcribe audio files directly from the terminal
- Dockerized Environment: While the server can be run directly on a local machine, it also includes a lightweight Docker configuration for users who prefer containerized deployment.
Getting started
Prerequisites
- Python 3.10+ (its Stable, popular, well-supported)
- Uvicorn (Lightweight ASGI server )
- Google Fire ( CLI Framework, Makes CLI creation simple and powerful)
- AssemblyAI API key (API Interface, sign up at https://www.assemblyai.com/)
- Docker (optional, for containerized usage. Ensures consistent, portable deployment)
Installation
- Clone the repository:
git clone <your-repo-url>
cd <repo-folder>
- Install dependencies:
pip install -r requirements.txt
- Set your AssemblyAI API key as an environment variable:
export API_KEY="your_assemblyai_api_key" # Linux / macOS
set API_KEY="your_assemblyai_api_key" # Windows CMD
Usage
Inspec With MCP Dev
Run This Command To Inspect And Test The Fuctionality Of Your Tool on a Web Ui.
uv run mcp dev mcp\server_transcription.py
connect and test server_transcription
Run the command to connect with claude
uv run mcp install mcp\server_transcription.py
Json Format
when connecting with an AI assitant like claude or cursor, the config should be in this jason format.
{
"mcpServers": {
"Audio Transcription Service": {
"command": "C:\\Users\\HomePC\\Desktop\\mcp_task2\\.venv\\Scripts\\uv.EXE",
"args": [
"run",
"--with",
"mcp[cli]",
"mcp",
"run",
"C:\\Users\\HomePC\\Desktop\\mcp_task2\\mcp\\server_transcription.py"
],
"env": {
"API_KEY": "<your-api-key-here>"
}
}
}
}
test client_transcription
Run the command
python mcp\client_transcription.py "path/to/audio/file"
Docker Usage
Build Docker Image
docker build -t transcription-service -f run_with_docker/Dockerfile .
Run Docker Container
set API_KEY="your-api-key-here"
docker run -d -p 8050:8050 -e API_KEY=%API_KEY% -v C:\Users:/mnt/users -v %CD%\transcripts:/app/transcripts -v %CD%\logs:/app/logs --dns 8.8.8.8 --name transcription-server transcription-service
run client
python run_with_docker\client.py "path/to/audio/file"
Recommend Servers
TraeBuild with Free GPT-4.1 & Claude 3.7. Fully MCP-Ready.
AiimagemultistyleA Model Context Protocol (MCP) server for image generation and manipulation using fal.ai's Stable Diffusion model.
Jina AI MCP ToolsA Model Context Protocol (MCP) server that integrates with Jina AI Search Foundation APIs.
Playwright McpPlaywright MCP server
DeepChatYour AI Partner on Desktop
WindsurfThe new purpose-built IDE to harness magic
MiniMax MCPOfficial MiniMax Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech, image generation and video generation APIs.
Visual Studio Code - Open Source ("Code - OSS")Visual Studio Code
CursorThe AI Code Editor
Zhipu Web SearchZhipu Web Search MCP Server is a search engine specifically designed for large models. It integrates four search engines, allowing users to flexibly compare and switch between them. Building upon the web crawling and ranking capabilities of traditional search engines, it enhances intent recognition capabilities, returning results more suitable for large model processing (such as webpage titles, URLs, summaries, site names, site icons, etc.). This helps AI applications achieve "dynamic knowledge acquisition" and "precise scenario adaptation" capabilities.
Baidu Map百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Context7Context7 MCP Server -- Up-to-date code documentation for LLMs and AI code editors
TimeA Model Context Protocol server that provides time and timezone conversion capabilities. This server enables LLMs to get current time information and perform timezone conversions using IANA timezone names, with automatic system timezone detection.
Howtocook Mcp基于Anduin2017 / HowToCook (程序员在家做饭指南)的mcp server,帮你推荐菜谱、规划膳食,解决“今天吃什么“的世纪难题;
Based on Anduin2017/HowToCook (Programmer's Guide to Cooking at Home), MCP Server helps you recommend recipes, plan meals, and solve the century old problem of "what to eat today"
EdgeOne Pages MCPAn MCP service designed for deploying HTML content to EdgeOne Pages and obtaining an accessible public URL.
ChatWiseThe second fastest AI chatbot™
Serper MCP ServerA Serper MCP Server
MCP AdvisorMCP Advisor & Installation - Use the right MCP server for your needs
BlenderBlenderMCP connects Blender to Claude AI through the Model Context Protocol (MCP), allowing Claude to directly interact with and control Blender. This integration enables prompt assisted 3D modeling, scene creation, and manipulation.
Amap Maps高德地图官方 MCP Server
Tavily Mcp