
mcp-local-rag MCP Server
The mcp-local-rag MCP Server enables privacy-respecting, local Retrieval-Augmented Generation (RAG) web search for LLMs. It allows AI assistants to access, embe...
FlowHunt provides an additional security layer between your internal systems and AI tools, giving you granular control over which tools are accessible from your MCP servers. MCP servers hosted in our infrastructure can be seamlessly integrated with FlowHunt's chatbot as well as popular AI platforms like ChatGPT, Claude, and various AI editors.
The mcp-rag-local MCP Server is designed as a memory server that allows AI assistants to store and retrieve text passages based on their semantic meaning, not just keywords. Leveraging Ollama for generating text embeddings and ChromaDB for vector storage and similarity search, it enables seamless storage (“memorization”) and retrieval of relevant texts for a given query. This empowers AI-driven workflows such as knowledge management, contextual recall, and semantic search. Developers can interact with the server to store individual texts, multiple texts, or even contents of PDF files, and later retrieve the most contextually relevant information, enhancing productivity and contextual awareness in applications.
memorize_text
Allows the server to store a single text passage for future semantic retrieval.
memorize_multiple_texts
Enables batch storage of several texts at once, facilitating bulk knowledge ingestion.
memorize_pdf_file
Reads and extracts up to 20 pages at a time from a PDF file, chunks the content, and memorizes it for semantic retrieval.
retrieve_similar_texts
Retrieves the most relevant stored text passages based on a user’s query, using semantic similarity.
(Tool names inferred from documented usage patterns; exact names may vary in code.)
Personal Knowledge Base
Developers and users can build a persistent, searchable knowledge base by memorizing articles, notes, or research papers for semantic recall.
Document and PDF Summarization
By memorizing entire PDF documents, users can later query and retrieve relevant sections or summaries, streamlining research and review.
Conversational Memory for Chatbots
Enhance AI assistants or chatbots with long-term, context-aware memory to provide more coherent and contextually relevant responses over time.
Semantic Search Engine
Implement a semantic search feature in applications, allowing users to find relevant information based on meaning, not just keywords.
Research and Data Exploration
Store and query technical documents, code snippets, or scientific literature for rapid, meaning-based retrieval during investigation or development.
git clone <repository-url>cd mcp-rag-localdocker-compose up to start ChromaDB and Ollama.docker exec -it ollama ollama pull all-minilm:l6-v2mcpServers):"mcp-rag-local": {
"command": "uv",
"args": [
"--directory",
"path\\to\\mcp-rag-local",
"run",
"main.py"
],
"env": {
"CHROMADB_PORT": "8321",
"OLLAMA_PORT": "11434"
}
}
"mcpServers": {
"mcp-rag-local": {
"command": "uv",
"args": [
"--directory",
"path\\to\\mcp-rag-local",
"run",
"main.py"
],
"env": {
"CHROMADB_PORT": "8321",
"OLLAMA_PORT": "11434"
}
}
}
"mcpServers": {
"mcp-rag-local": {
"command": "uv",
"args": [
"--directory",
"path\\to\\mcp-rag-local",
"run",
"main.py"
],
"env": {
"CHROMADB_PORT": "8321",
"OLLAMA_PORT": "11434"
}
}
}
"mcpServers": {
"mcp-rag-local": {
"command": "uv",
"args": [
"--directory",
"path\\to\\mcp-rag-local",
"run",
"main.py"
],
"env": {
"CHROMADB_PORT": "8321",
"OLLAMA_PORT": "11434"
}
}
}
env section of your configuration."env": {
"CHROMADB_PORT": "8321",
"OLLAMA_PORT": "11434",
"MY_API_KEY": "${MY_API_KEY}"
}
Using MCP in FlowHunt
To integrate MCP servers into your FlowHunt workflow, start by adding the MCP component to your flow and connecting it to your AI agent:

Click on the MCP component to open the configuration panel. In the system MCP configuration section, insert your MCP server details using this JSON format:
{
"mcp-rag-local": {
"transport": "streamable_http",
"url": "https://yourmcpserver.example/pathtothemcp/url"
}
}
Once configured, the AI agent is now able to use this MCP as a tool with access to all its functions and capabilities. Remember to change “mcp-rag-local” to whatever the actual name of your MCP server is and replace the URL with your own MCP server URL.
| Section | Availability | Details/Notes |
|---|---|---|
| Overview | ✅ | |
| List of Prompts | ⛔ | No prompts/templates documented |
| List of Resources | ⛔ | No resources documented |
| List of Tools | ✅ | memorize_text, memorize_multiple_texts, etc. |
| Securing API Keys | ✅ | via env in config, example shown |
| Sampling Support (less important in evaluation) | ⛔ | Not mentioned |
This MCP is straightforward and well-focused on semantic memory, but lacks advanced features like prompt templates, explicit resources, or sampling/roots support. Tooling and setup are clear. Best for simple RAG/local knowledge workflows.
| Has a LICENSE | ✅ (MIT) |
|---|---|
| Has at least one tool | ✅ |
| Number of Forks | 1 |
| Number of Stars | 5 |
Supercharge your AI workflows with semantic memory and local document search using mcp-rag-local. Set up in minutes and transform how your agents recall and reason over knowledge.

The mcp-local-rag MCP Server enables privacy-respecting, local Retrieval-Augmented Generation (RAG) web search for LLMs. It allows AI assistants to access, embe...

The Rememberizer MCP Server bridges AI assistants and knowledge management, enabling semantic search, unified document retrieval, and team collaboration across ...

Integrate Rember’s spaced repetition flashcard system with AI assistants using the Rember MCP Server. Automate flashcard creation from chats, documents, and use...
Cookie Consent
We use cookies to enhance your browsing experience and analyze our traffic. See our privacy policy.