Sponsored by Deepsite.site

ObsidianFetch

Created By
soukouki8 months ago
MCP servers focused on fetching and presenting information from Obsidian vaults.
Content

ObsidianFetch

MCP servers focused on fetching and presenting information from Obsidian vaults.

The existing MCP server has the following drawbacks:

  • It supports many commands, which can cause slow prompt loading when computational resources are limited.
  • When reading a note labeled "Sample Note", it is necessary to search for its path first before loading it, but the LLM may not always follow this procedure.
  • Some tools include unnecessary options, leading the LLM to sometimes fail to invoke them correctly.

These issues become particularly noticeable when running an LLM on a local GPU.
To address this, we developed a new MCP server that simply retrieves and loads lists of notes.

The new server also provides the following additional features:

  • When the LLM attempts to retrieve link information by searching with brackets like [[link name]], the server automatically removes any characters that cannot be used in links.
  • In addition to loading the note contents, it also displays backlinks—notes that link to the currently opened note.
    • This allows the LLM to load and understand the connections between related notes via backlinks.

Installation

gem install obsidian_fetch

Usage

obsidian_fetch /path/to/your/vault

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/soukouki/obsidian_fetch.

License

The gem is available as open source under the terms of the MIT License.

Recommend Servers
TraeBuild with Free GPT-4.1 & Claude 3.7. Fully MCP-Ready.
Jina AI MCP ToolsA Model Context Protocol (MCP) server that integrates with Jina AI Search Foundation APIs.
Amap Maps高德地图官方 MCP Server
WindsurfThe new purpose-built IDE to harness magic
CursorThe AI Code Editor
MiniMax MCPOfficial MiniMax Model Context Protocol (MCP) server that enables interaction with powerful Text to Speech, image generation and video generation APIs.
TimeA Model Context Protocol server that provides time and timezone conversion capabilities. This server enables LLMs to get current time information and perform timezone conversions using IANA timezone names, with automatic system timezone detection.
Zhipu Web SearchZhipu Web Search MCP Server is a search engine specifically designed for large models. It integrates four search engines, allowing users to flexibly compare and switch between them. Building upon the web crawling and ranking capabilities of traditional search engines, it enhances intent recognition capabilities, returning results more suitable for large model processing (such as webpage titles, URLs, summaries, site names, site icons, etc.). This helps AI applications achieve "dynamic knowledge acquisition" and "precise scenario adaptation" capabilities.
ChatWiseThe second fastest AI chatbot™
Baidu Map百度地图核心API现已全面兼容MCP协议,是国内首家兼容MCP协议的地图服务商。
Howtocook Mcp基于Anduin2017 / HowToCook (程序员在家做饭指南)的mcp server,帮你推荐菜谱、规划膳食,解决“今天吃什么“的世纪难题; Based on Anduin2017/HowToCook (Programmer's Guide to Cooking at Home), MCP Server helps you recommend recipes, plan meals, and solve the century old problem of "what to eat today"
DeepChatYour AI Partner on Desktop
BlenderBlenderMCP connects Blender to Claude AI through the Model Context Protocol (MCP), allowing Claude to directly interact with and control Blender. This integration enables prompt assisted 3D modeling, scene creation, and manipulation.
EdgeOne Pages MCPAn MCP service designed for deploying HTML content to EdgeOne Pages and obtaining an accessible public URL.
Tavily Mcp
Serper MCP ServerA Serper MCP Server
Visual Studio Code - Open Source ("Code - OSS")Visual Studio Code
Playwright McpPlaywright MCP server
Context7Context7 MCP Server -- Up-to-date code documentation for LLMs and AI code editors
MCP AdvisorMCP Advisor & Installation - Use the right MCP server for your needs
AiimagemultistyleA Model Context Protocol (MCP) server for image generation and manipulation using fal.ai's Stable Diffusion model.