Qmcp
What is Qmcp
Qmcp is a powerful, cross-platform AI chat client built with Flutter, implementing the Model Context Protocol (MCP) for intelligent, context-aware interactions with multiple Large Language Models (LLMs).
Use cases
Use cases for Qmcp include automated customer service chatbots, virtual teaching assistants, content generation tools, and interactive storytelling applications.
How to use
To use Qmcp, install the necessary packages on your device, set up the environment according to the platform (Android, iOS, Desktop), and connect to your desired LLM. Follow the setup instructions provided in the README for detailed steps.
Key features
Key features of Qmcp include universal compatibility across major platforms, model flexibility allowing connection to any supported LLM, context awareness leveraging MCP for maintaining conversation context, and being enterprise-ready with a focus on security and scalability.
Where to use
Qmcp can be used in various fields such as customer support, education, content creation, and any domain requiring intelligent conversational agents that can interact with users in a context-aware manner.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Overview
What is Qmcp
Qmcp is a powerful, cross-platform AI chat client built with Flutter, implementing the Model Context Protocol (MCP) for intelligent, context-aware interactions with multiple Large Language Models (LLMs).
Use cases
Use cases for Qmcp include automated customer service chatbots, virtual teaching assistants, content generation tools, and interactive storytelling applications.
How to use
To use Qmcp, install the necessary packages on your device, set up the environment according to the platform (Android, iOS, Desktop), and connect to your desired LLM. Follow the setup instructions provided in the README for detailed steps.
Key features
Key features of Qmcp include universal compatibility across major platforms, model flexibility allowing connection to any supported LLM, context awareness leveraging MCP for maintaining conversation context, and being enterprise-ready with a focus on security and scalability.
Where to use
Qmcp can be used in various fields such as customer support, education, content creation, and any domain requiring intelligent conversational agents that can interact with users in a context-aware manner.
Clients Supporting MCP
The following are the main client software that supports the Model Context Protocol. Click the link to visit the official website for more information.
Content
Qubase MCP
A powerful, cross-platform AI Chat Client implementing the Model Context Protocol (MCP) for seamless AI interactions.
Overview
Qubase MCP is a sophisticated chat interface that revolutionizes AI interactions on desktop and mobile devices. Built with Flutter and implementing the Model Context Protocol, it provides a unified interface for multiple Large Language Models (LLMs) while ensuring secure, efficient, and context-aware conversations.
Features
Core Capabilities
| Feature | Description |
|---|---|
| Universal Compatibility | Works seamlessly across all major platforms |
| Model Flexibility | Connect to any supported LLM without changing workflow |
| Context Awareness | Leverages MCP for maintaining conversation context |
| Enterprise Ready | Built with security and scalability in mind |
Platform Support
| Platform | Status |
|---|---|
| Desktop (macOS, Windows, Linux) | Available |
| Mobile (iOS, Android) | Available |
| Web | Coming Soon |
Supported AI Models
- OpenAI (GPT-3.5, GPT-4)
- Anthropic Claude
- OLLama (Local Models)
- DeepSeek
- Custom Model Support
Local LLM Setup
Android Setup
-
Install Termux
- Download Termux ARM64 V8 from Termux GitHub
- Install and open Termux
- Run
termux-setup-storageto grant storage permissions - Run
termux-change-repoto select package mirror - Update with
pkg upgrade
-
Required Packages
# Install Tur repository pkg install tur-repo # Install Ollama and Zellij pkg install ollama pkg install zellij -
Android Configuration
- Enable Developer Options: Settings > About device > Tap “Build number” 7 times
- In Developer options, enable “Disable child process restrictions”
-
Model Operations
# Start Ollama server ollama serve # In a new terminal, run models: ollama run deepseek-r1.5b # For DeepSeek ollama run llama3.2 # For Llama3 -
Control Commands
Action Command Stop output CTRL + CExit model CTRL + DClear screen CTRL + LStop server ps aux | grep ollamathenkill [PID]
Desktop Setup
Coming soon…
System Requirements
Hardware Requirements
| Component | Minimum | Recommended |
|---|---|---|
| RAM | 4GB | 8GB |
| Storage | 2GB free | 4GB free |
| Processor | Intel/AMD x64 or ARM64 | Modern multi-core |
Software Requirements
| Component | Version |
|---|---|
| Windows | 10 or later |
| macOS | 10.15 or later |
| Ubuntu | 20.04 or later |
| Flutter SDK | 3.0 or later |
| Git | Latest stable |
Installation Guide
Prerequisites
-
Flutter Setup
git clone https://github.com/flutter/flutter.git export PATH="$PATH:`pwd`/flutter/bin" flutter doctor -
System Dependencies
# Linux sudo apt-get update sudo apt-get install libsqlite3-0 libsqlite3-dev # macOS brew install sqlite3 # Windows # SQLite included in Flutter Windows setup -
Development Tools
# Using uv (Recommended) brew install uv # Alternative: npm brew install node
Application Setup
-
Installation
git clone https://github.com/qubasehq/Qmcp.git cd Qmcp flutter pub get -
Launch
# Desktop platforms flutter run -d <platform> # macos, windows, linux # Mobile development flutter run -d <device-id>
Configuration
Initial Setup
-
API Configuration
- Launch Qubase MCP
- Navigate to Settings > API Configuration
- Enter LLM API credentials
- Configure custom endpoints (if needed)
-
MCP Server Setup
- Access Settings > MCP Server
- Choose installation method
- Configure server settings
- Select default AI model
File Locations
| Purpose | Path |
|---|---|
| Configuration | ~/Library/Application Support/qubase_mcp/mcp_server.json |
| Logs | ~/Library/Application Support/run.daodao.qubase_mcp/logs |
| Application Data | ~/Library/Application Support/qubase_mcp |
Reset Application
# Clear all data (use with caution)
rm -rf ~/Library/Application\ Support/run.daodao.qubase_mcp
rm -rf ~/Library/Application\ Support/qubase_mcp
Troubleshooting
| Issue | Solution |
|---|---|
| Connection Issues | Verify API keys and network connectivity |
| Performance Problems | Check system resources and clear cache |
| Model Errors | Validate model configurations and quotas |
For additional support, visit our Issues page.
Development Roadmap
Upcoming Features
| Feature | Description |
|---|---|
| MCP Server Marketplace | Easy discovery and deployment of community servers |
| Enhanced Integration | Automated server setup and cloud sync support |
| RAG Implementation | Document processing and knowledge base integration |
| UI/UX Improvements | Custom themes, keyboard shortcuts, mobile optimization |
Contributing
- Fork the repository
- Create a feature branch
- Commit your changes
- Push to your branch
- Open a Pull Request
Acknowledgments
- Model Context Protocol (MCP) Team
- MCP CLI Contributors
- Flutter Community
- Open Source AI Community
License
Licensed under Apache License 2.0 - see the LICENSE file for details.
Dev Tools Supporting MCP
The following are the main code editors that support the Model Context Protocol. Click the link to visit the official website for more information.










