- Explore MCP Servers
- groq-mcp-server
Groq Mcp Server
What is Groq Mcp Server
The groq-mcp-server is a server that facilitates lightning-fast inference for query models hosted on Groq, enabling communication with various MCP clients through the Model Context Protocol (MCP).
Use cases
Use cases for the groq-mcp-server include describing images, extracting information from images, converting text to speech in multiple languages, transcribing audio files, and processing batches of prompts for large-scale tasks.
How to use
To use the groq-mcp-server, obtain a Groq API key, install the ‘uv’ package manager, and configure the Claude Desktop settings to include the groq MCP server with the appropriate command and environment variables.
Key features
Key features of the groq-mcp-server include access to vision models for image analysis, text-to-speech conversion, audio transcription, and batch processing capabilities for handling multiple requests efficiently.
Where to use
The groq-mcp-server can be utilized in fields such as artificial intelligence, machine learning, multimedia processing, and application development that require fast and efficient data processing.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Overview
What is Groq Mcp Server
The groq-mcp-server is a server that facilitates lightning-fast inference for query models hosted on Groq, enabling communication with various MCP clients through the Model Context Protocol (MCP).
Use cases
Use cases for the groq-mcp-server include describing images, extracting information from images, converting text to speech in multiple languages, transcribing audio files, and processing batches of prompts for large-scale tasks.
How to use
To use the groq-mcp-server, obtain a Groq API key, install the ‘uv’ package manager, and configure the Claude Desktop settings to include the groq MCP server with the appropriate command and environment variables.
Key features
Key features of the groq-mcp-server include access to vision models for image analysis, text-to-speech conversion, audio transcription, and batch processing capabilities for handling multiple requests efficiently.
Where to use
The groq-mcp-server can be utilized in fields such as artificial intelligence, machine learning, multimedia processing, and application development that require fast and efficient data processing.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Content
Groq MCP Server
Query models hosted on Groq for lightning-fast inference directly from Claude and other MCP clients through the Model Context Protocol (MCP).
Use MCP to access vision models for interpreting visual data from images, instantly generate speech from text, process thousands of requests through Groq’s batch processing, and even build apps with full access to Groq’s documentation.
With the Groq MCP server you can try tasks like:
Agentic Tasks, Code Generation & Web Search
- What is Groq’s Compound Beta? Use the compound tool. Summarize with one line then turn into voice
- Please retrieve the current Bitcoin price from CoinGecko API and calculate the value of 0.38474 bitcoins?
- What is the weather in SF right now?
- Generate and run code, which means you can make API calls, get data from webpages, and much more
- This feature uses the new
compound-betaagentic tools system
Vision & Understanding
- “Describe this image [URL to image]”
- “Analyze this image and extract key information as JSON [URL to image]”
Speech & Audio
- “Convert this text to speech using the Arista-PlayAI voice: [text]”
- “Read this text aloud in Arabic: [text]”
- “Transcribe this audio file using whisper-large-v3: [url to mp3]”
- “Translate this foreign language audio to English [url to mp3]”
Batch Processing
- “Process the following batch of prompts: [location of a jsonlines file]” (read more here)
Quickstart with Claude Desktop
- Get a Groq API key for free at console.groq.com
- Install
uv(Python package manager), install withcurl -LsSf https://astral.sh/uv/install.sh | shor see theuvrepo for additional install methods. - Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:
{ "mcpServers": { "groq": { "command": "uvx", "args": ["groq-mcp"], "env": { "GROQ_API_KEY": "your_groq_api_key", "BASE_OUTPUT_PATH": "/path/to/output/directory" # Optional: Where to save generated files (default: ~/Desktop) } } } }
If you’re using Windows, you will have to enable “Developer Mode” in Claude Desktop to use the MCP server. Click “Help” in the hamburger menu in the top left and select “Enable Developer Mode”.
If you want to install the MCP from code, scroll down to “Contributing”.
Other MCP Clients
For other clients like Cursor and Windsurf:
-
Install the package:
# Using UV (recommended) uvx install groq-mcp # Or using pip pip install groq-mcp -
Generate configuration:
# Print config to screen groq-mcp-config --api-key=your_groq_api_key --print # Or save directly to config file (auto-detects location) groq-mcp-config --api-key=your_groq_api_key # Optional: Specify custom output path groq-mcp-config --api-key=your_groq_api_key --output-path=/path/to/outputs
That’s it! Your MCP client can now use these Groq capabilities:
- 🗣️ Text-to-Speech (TTS): Fast, natural-sounding speech synthesis
- 👂 Speech-to-Text (STT): Accurate transcription and translation
- 🖼️ Vision: Advanced image analysis and understanding
- 💬 Chat: Ultra-fast LLM inference with Llama 4 and more
- 📦 Batch: Process large workloads efficiently
Contributing
If you want to contribute or run from source:
Installation Options
Option 1: Quick Setup (Recommended)
-
Clone the repository:
git clone https://github.com/groq/groq-mcp-server cd groq-mcp -
Run the setup script:
./scripts/setup.shThis will:
- Create a Python virtual environment using
uv - Install all dependencies
- Set up pre-commit hooks
- Activate the virtual environment
- Create a Python virtual environment using
-
Run the Claude install script:
./scripts/install.shOn Macs, this will install the Groq MCP server in Claude Desktop, at
~/Library/Application Support/Claude/claude_desktop_config.json. Make sure to refresh or restart Claude Desktop. -
Copy
.env.exampleto.envand add your Groq API key:cp .env.example .env # Edit .env and add your API key
Option 2: Manual Setup
-
Clone the repository:
git clone https://github.com/groq/groq-mcp-server cd groq-mcp -
Create a virtual environment and install dependencies using uv:
uv venv source .venv/bin/activate uv pip install -e ".[dev]" -
Copy
.env.exampleto.envand add your Groq API key:cp .env.example .env # Edit .env and add your API key
Available Scripts
The scripts directory contains several utility scripts for different Groq API functionalities:
Vision & Image Analysis
./scripts/groq_vision.sh <image_file> [prompt] [temperature] [max_tokens] [output_directory]
# Example:
./scripts/groq_vision.sh "./input/image.jpg" "What is in this image?"
Text-to-Speech (TTS)
./scripts/groq_tts.sh "Your text" [voice_name] [model] [output_directory]
# Example:
./scripts/groq_tts.sh "Hello, world!" "Arista-PlayAI"
Speech-to-Text (STT)
./scripts/groq_stt.sh <audio_file> [model] [output_directory]
Utility Scripts
list_groq_voices.sh: Display available TTS voiceslist_groq_stt_models.sh: Show available STT modelsgroq_batch.sh: Process batch operationsgroq_translate.sh: Translate text or audio
Development Scripts
# Run tests
./scripts/test.sh
# Run with options
./scripts/test.sh --verbose --fail-fast
# Run integration tests
./scripts/test.sh --integration
# Debug and test locally
mcp install server.py
mcp dev server.py
Troubleshooting
Logs when running with Claude Desktop can be found at:
- Windows:
%APPDATA%\Claude\logs\groq-mcp.log - macOS:
~/Library/Logs/Claude/groq-mcp.log
Acknowledgments
This project is inspired by the ElevenLabs MCP Server. Thanks!
Dev Tools Supporting MCP
The following are the main code editors that support the Model Context Protocol. Click the link to visit the official website for more information.










