by mamertofabian
Generate speech audio from text via ElevenLabs API and manage voice generation tasks through a Model Context Protocol server with a companion SvelteKit web client.
Provides a Model Context Protocol (MCP) server that converts text to speech using the ElevenLabs API, stores job history in SQLite, and offers a sample SvelteKit client for interacting with the server.
uvx
approach: no manual install needed, just configure the MCP settings JSON with the command uvx
and the appropriate environment variables.uvx elevenlabs-mcp-server
based on the configuration.clients/web-ui
, install dependencies with pnpm
, copy .env.example
to .env
, then start the UI via pnpm dev
and open http://localhost:5174
.generate_audio_simple
, generate_audio_script
, delete_job
, list_voices
, etc., through MCP.Q: Which API key is required?
A: An ElevenLabs API key (ELEVENLABS_API_KEY
).
Q: How to choose a voice?
A: Set ELEVENLABS_VOICE_ID
in the environment or use the list_voices
tool to retrieve available IDs.
Q: Can I change the synthesis model?
A: Yes, adjust ELEVENLABS_MODEL_ID
(default eleven_flash_v2
).
Q: Where are audio files saved?
A: To the directory defined by ELEVENLABS_OUTPUT_DIR
(default output
).
Q: How is job history accessed?
A: Via the get_voiceover_history
tool or the resource URI voiceover://history/{job_id}
.
A Model Context Protocol (MCP) server that integrates with ElevenLabs text-to-speech API, featuring both a server component and a sample web-based MCP Client (SvelteKit) for managing voice generation tasks.
To install ElevenLabs MCP Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install elevenlabs-mcp-server --client claude
When using uvx
, no specific installation is needed.
Add the following configuration to your MCP settings file (e.g., cline_mcp_settings.json
for Claude Desktop):
{
"mcpServers": {
"elevenlabs": {
"command": "uvx",
"args": ["elevenlabs-mcp-server"],
"env": {
"ELEVENLABS_API_KEY": "your-api-key",
"ELEVENLABS_VOICE_ID": "your-voice-id",
"ELEVENLABS_MODEL_ID": "eleven_flash_v2",
"ELEVENLABS_STABILITY": "0.5",
"ELEVENLABS_SIMILARITY_BOOST": "0.75",
"ELEVENLABS_STYLE": "0.1",
"ELEVENLABS_OUTPUT_DIR": "output"
}
}
}
}
uv venv
.env.example
to .env
and fill in your ElevenLabs credentials{
"mcpServers": {
"elevenlabs": {
"command": "uv",
"args": [
"--directory",
"path/to/elevenlabs-mcp-server",
"run",
"elevenlabs-mcp-server"
],
"env": {
"ELEVENLABS_API_KEY": "your-api-key",
"ELEVENLABS_VOICE_ID": "your-voice-id",
"ELEVENLABS_MODEL_ID": "eleven_flash_v2",
"ELEVENLABS_STABILITY": "0.5",
"ELEVENLABS_SIMILARITY_BOOST": "0.75",
"ELEVENLABS_STYLE": "0.1",
"ELEVENLABS_OUTPUT_DIR": "output"
}
}
}
}
cd clients/web-ui
pnpm install
.env.example
to .env
and configure as neededpnpm dev
generate_audio_simple
: Generate audio from plain text using default voice settingsgenerate_audio_script
: Generate audio from a structured script with multiple voices and actorsdelete_job
: Delete a job by its IDget_audio_file
: Get the audio file by its IDlist_voices
: List all available voicesget_voiceover_history
: Get voiceover job history. Optionally specify a job ID for a specific job.voiceover://history/{job_id}
: Get the audio file by its IDvoiceover://voices
: List all available voicesThis project is licensed under the MIT License - see the LICENSE file for details.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "elevenlabs": { "command": "uvx", "args": [ "elevenlabs-mcp-server" ], "env": { "ELEVENLABS_API_KEY": "<YOUR_API_KEY>", "ELEVENLABS_VOICE_ID": "<YOUR_VOICE_ID>", "ELEVENLABS_MODEL_ID": "eleven_flash_v2", "ELEVENLABS_STABILITY": "0.5", "ELEVENLABS_SIMILARITY_BOOST": "0.75", "ELEVENLABS_STYLE": "0.1", "ELEVENLABS_OUTPUT_DIR": "output" } } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by burningion
Upload, edit, search, and generate videos by leveraging LLM capabilities together with Video Jungle's media library.
by Flyworks-AI
Create fast, free lip‑sync videos for digital avatars by providing audio or text, with optional avatar generation from images or videos.
by mberg
Generates spoken audio from text, outputting MP3 files locally and optionally uploading them to Amazon S3.
by allvoicelab
Generate natural speech, translate and dub videos, clone voices, remove hardcoded subtitles, and extract subtitles using powerful AI APIs.
by nabid-pf
Extracts YouTube video captions, subtitles, and metadata to supply structured information for AI assistants to generate concise video summaries.
by omergocmen
Provides video generation and status checking via the json2video API for seamless integration with LLMs, agents, and other MCP‑compatible clients.
by cartesia-ai
Provides clients such as Cursor, Claude Desktop, and OpenAI agents with capabilities to localize speech, convert text to audio, and infill voice clips via Cartesia's API.
by TSavo
Provides an enterprise‑grade MCP server that exposes 12 AI video generation tools, enabling AI assistants to create avatar videos, URL‑to‑video conversions, short videos, scripts, custom avatars, advanced lip‑sync, and more through natural language interactions.
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.