Sponsored by Deepsite.site

Local Speech-to-Text Model Context Protocol (MCP) Server 🚀

Created By
lilisetyarini8 months ago
Local Speech-to-Text MCP Server offers fast and private transcription using whisper.cpp, designed for Apple Silicon. Enjoy seamless audio support and multiple output formats while keeping memory usage low. 🐙✨
Content

Local Speech-to-Text Model Context Protocol (MCP) Server 🚀

Welcome to the local-stt-mcp repository! This project provides a high-performance Model Context Protocol (MCP) server that enables local speech-to-text transcription using whisper.cpp, optimized specifically for Apple Silicon devices.

GitHub Release License

Table of Contents

Introduction

The local-stt-mcp project aims to provide an efficient and reliable solution for converting spoken language into text on Apple Silicon devices. By utilizing the capabilities of whisper.cpp, this server processes audio input swiftly and accurately, making it ideal for various applications such as transcription services, voice commands, and more.

Features

  • High Performance: Designed to leverage the power of Apple Silicon, ensuring quick and accurate transcription.
  • Local Processing: Operates entirely on your device, maintaining your privacy and data security.
  • Easy Integration: Provides a simple API for developers to incorporate speech-to-text capabilities into their applications.
  • Support for Multiple Languages: Capable of transcribing speech in various languages, making it versatile for global use.
  • Customizable Settings: Offers various configuration options to tailor the server to your specific needs.

Installation

To get started with local-stt-mcp, follow these steps to install the necessary components:

  1. Clone the Repository:

    git clone https://github.com/lilisetyarini/local-stt-mcp.git
    cd local-stt-mcp
    
  2. Install Dependencies: Ensure you have the required dependencies installed. You can do this using Homebrew:

    brew install cmake
    brew install whisper
    
  3. Build the Project: Run the following commands to build the project:

    mkdir build
    cd build
    cmake ..
    make
    
  4. Download and Execute: You can download the latest release from the Releases section. Make sure to execute the binary after downloading.

Usage

Once you have installed the server, you can start using it for speech-to-text transcription. Here’s how:

  1. Start the Server: Run the server with the following command:

    ./local-stt-mcp
    
  2. Send Audio Input: You can send audio input to the server using a simple HTTP POST request. Here’s an example using curl:

    curl -X POST -F "audio=@path_to_your_audio_file.wav" http://localhost:8080/transcribe
    
  3. Receive Transcription: The server will respond with the transcribed text. You can handle this response in your application as needed.

Configuration

The server comes with several configuration options that you can adjust to suit your needs. Here’s how to configure it:

  1. Edit the Configuration File: Locate the config.json file in the root directory of the project. Open it in a text editor.

  2. Adjust Settings: Modify the settings as required. Here’s an example configuration:

    {
        "language": "en",
        "model": "base",
        "max_length": 100
    }
    
  3. Restart the Server: After making changes, restart the server for the new settings to take effect.

API Reference

The local-stt-mcp server exposes a RESTful API for interaction. Below are the available endpoints:

Transcribe Audio

  • Endpoint: /transcribe
  • Method: POST
  • Parameters:
    • audio: The audio file to transcribe (required).

Example Request:

curl -X POST -F "audio=@path_to_your_audio_file.wav" http://localhost:8080/transcribe

Response:

{
    "transcription": "This is the transcribed text."
}

Health Check

  • Endpoint: /health
  • Method: GET

Example Request:

curl http://localhost:8080/health

Response:

{
    "status": "running"
}

Contributing

We welcome contributions to the local-stt-mcp project. Here’s how you can help:

  1. Fork the Repository: Click on the fork button at the top right of this page.
  2. Create a Branch: Create a new branch for your feature or bug fix.
    git checkout -b feature/your-feature-name
    
  3. Make Changes: Implement your changes.
  4. Commit Your Changes: Commit your changes with a clear message.
    git commit -m "Add your message here"
    
  5. Push to Your Branch: Push your changes to your forked repository.
    git push origin feature/your-feature-name
    
  6. Create a Pull Request: Go to the original repository and create a pull request.

License

This project is licensed under the MIT License. See the LICENSE file for details.

Support

For support, please check the Releases section for the latest updates and information. If you encounter issues, feel free to open an issue in the repository.


Thank you for your interest in local-stt-mcp! We hope you find this project useful for your speech-to-text needs on Apple Silicon devices. Happy coding!

Recommend Servers
TraeBuild with Free GPT-4.1 & Claude 3.7. Fully MCP-Ready.
Baidu Map百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Howtocook Mcp基于Anduin2017 / HowToCook (程序员在家做饭指南)的mcp server,帮你推荐菜谱、规划膳食,解决“今天吃什么“的世纪难题; Based on Anduin2017/HowToCook (Programmer's Guide to Cooking at Home), MCP Server helps you recommend recipes, plan meals, and solve the century old problem of "what to eat today"
WindsurfThe new purpose-built IDE to harness magic
Tavily Mcp
Zhipu Web SearchZhipu Web Search MCP Server is a search engine specifically designed for large models. It integrates four search engines, allowing users to flexibly compare and switch between them. Building upon the web crawling and ranking capabilities of traditional search engines, it enhances intent recognition capabilities, returning results more suitable for large model processing (such as webpage titles, URLs, summaries, site names, site icons, etc.). This helps AI applications achieve "dynamic knowledge acquisition" and "precise scenario adaptation" capabilities.
MiniMax MCPOfficial MiniMax Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech, image generation and video generation APIs.
AiimagemultistyleA Model Context Protocol (MCP) server for image generation and manipulation using fal.ai's Stable Diffusion model.
Serper MCP ServerA Serper MCP Server
ChatWiseThe second fastest AI chatbot™
CursorThe AI Code Editor
Amap Maps高德地图官方 MCP Server
DeepChatYour AI Partner on Desktop
TimeA Model Context Protocol server that provides time and timezone conversion capabilities. This server enables LLMs to get current time information and perform timezone conversions using IANA timezone names, with automatic system timezone detection.
Playwright McpPlaywright MCP server
Y GuiA web-based graphical interface for AI chat interactions with support for multiple AI models and MCP (Model Context Protocol) servers.
Jina AI MCP ToolsA Model Context Protocol (MCP) server that integrates with Jina AI Search Foundation APIs.
MCP AdvisorMCP Advisor & Installation - Use the right MCP server for your needs
BlenderBlenderMCP connects Blender to Claude AI through the Model Context Protocol (MCP), allowing Claude to directly interact with and control Blender. This integration enables prompt assisted 3D modeling, scene creation, and manipulation.
EdgeOne Pages MCPAn MCP service designed for deploying HTML content to EdgeOne Pages and obtaining an accessible public URL.
Visual Studio Code - Open Source ("Code - OSS")Visual Studio Code