by chatmcp
Summarize chat messages by querying a local chat database and returning concise overviews.
Mcp Server Chatsum provides a tool to query stored chat messages and generate summaries based on custom prompts. It works with a local SQLite chat database created by the companion chatbot project.
CHAT_DB_PATH
to point to your chat.db
file.pnpm install
and build the server (pnpm build
).npx
command.query_chat_messages
tool with desired query parameters and a summarization prompt.pnpm watch
) and debugging via MCP Inspector.chatbot
project.query_chat_messages
tool accepts any prompt you provide.pnpm inspector
) for interactive debugging.This MCP Server is used to summarize your chat messages.
Before you start
move to chatbot directory, follow the README to setup the chat database.
start chatbot to save your chat messages.
query_chat_messages
- Query chat messages
create .env
file in the root directory, and set your chat database path.
CHAT_DB_PATH=path-to/chatbot/data/chat.db
pnpm install
Build the server:
pnpm build
For development with auto-rebuild:
pnpm watch
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"mcp-server-chatsum": {
"command": "path-to/bin/node",
"args": ["path-to/mcp-server-chatsum/build/index.js"],
"env": {
"CHAT_DB_PATH": "path-to/mcp-server-chatsum/chatbot/data/chat.db"
}
}
}
}
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
pnpm inspector
The Inspector will provide a URL to access debugging tools in your browser.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "mcp-server-chatsum": { "command": "npx", "args": [ "-y", "mcp-server-chatsum" ], "env": { "CHAT_DB_PATH": "<PATH_TO_CHAT_DB>" } } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by modelcontextprotocol
A basic implementation of persistent memory using a local knowledge graph. This lets Claude remember information about the user across chats.
by topoteretes
Provides dynamic memory for AI agents through modular ECL (Extract, Cognify, Load) pipelines, enabling seamless integration with graph and vector stores using minimal code.
by basicmachines-co
Enables persistent, local‑first knowledge management by allowing LLMs to read and write Markdown files during natural conversations, building a traversable knowledge graph that stays under the user’s control.
by smithery-ai
Provides read and search capabilities for Markdown notes in an Obsidian vault for Claude Desktop and other MCP clients.
by dmayboroda
Provides on‑premises conversational retrieval‑augmented generation (RAG) with configurable Docker containers, supporting fully local execution, ChatGPT‑based custom GPTs, and Anthropic Claude integration.
by GreatScottyMac
Provides a project‑specific memory bank that stores decisions, progress, architecture, and custom data, exposing a structured knowledge graph via MCP for AI assistants and IDE tools.
by andrea9293
Provides document management and AI-powered semantic search for storing, retrieving, and querying text, markdown, and PDF files locally without external databases.
by scorzeth
Provides a local MCP server that interfaces with a running Anki instance to retrieve, create, and update flashcards through standard MCP calls.
by sirmews
Read and write records in a Pinecone vector index via Model Context Protocol, enabling semantic search and document management for Claude Desktop.