Fal
What is Fal
fal is a Model Context Protocol (MCP) server designed for interacting with fal.ai models and services, providing a streamlined interface for model management and execution.
Use cases
Use cases for fal include generating text or data using AI models, managing multiple model executions, integrating with development environments for testing, and facilitating content delivery through CDN.
How to use
To use fal, clone the repository, install the required packages, obtain your fal.ai API key, and run the server locally using the provided scripts. You can connect to it from your LLM IDE by configuring the URL.
Key features
Key features of fal include listing available models, searching by keywords, retrieving model schemas, generating content, managing execution queues, file uploads to CDN, and full streaming support via HTTP.
Where to use
fal can be used in various fields such as artificial intelligence, machine learning, content generation, and any application requiring model interaction and management.
Overview
What is Fal
fal is a Model Context Protocol (MCP) server designed for interacting with fal.ai models and services, providing a streamlined interface for model management and execution.
Use cases
Use cases for fal include generating text or data using AI models, managing multiple model executions, integrating with development environments for testing, and facilitating content delivery through CDN.
How to use
To use fal, clone the repository, install the required packages, obtain your fal.ai API key, and run the server locally using the provided scripts. You can connect to it from your LLM IDE by configuring the URL.
Key features
Key features of fal include listing available models, searching by keywords, retrieving model schemas, generating content, managing execution queues, file uploads to CDN, and full streaming support via HTTP.
Where to use
fal can be used in various fields such as artificial intelligence, machine learning, content generation, and any application requiring model interaction and management.
Content
fal MCP Server
A Model Context Protocol (MCP) server for interacting with fal.ai models and services. This project was inspired by am0y’s MCP server, but updated to use the latest streaming MCP support.
Features
- List all available fal.ai models
- Search for specific models by keywords
- Get model schemas
- Generate content using any fal.ai model
- Support for both direct and queued model execution
- Queue management (status checking, getting results, cancelling requests)
- File upload to fal.ai CDN
- Full streaming support via HTTP transport
Requirements
- Python 3.12+
- fastmcp
- httpx
- aiofiles
- A fal.ai API key
Installation
- Clone this repository:
git clone https://github.com/derekalia/fal.git
cd fal
- Install the required packages:
# Using uv (recommended)
uv sync
# Or using pip
pip install fastmcp httpx aiofiles
Usage
Running the Server Locally
./run_http.sh YOUR_FAL_API_KEY
The server will start and display connection information in your terminal.
- Connect to it from your LLM IDE (Claude Code or Cursor) by adding to your configuration:
{
"Fal": {
"url": "http://127.0.0.1:6274/mcp/"
}
}
Development Mode (with MCP Inspector)
For testing and debugging, you can run the server in development mode:
fastmcp dev main.py
This will:
- Start the server on a random port
- Launch the MCP Inspector web interface in your browser
- Allow you to test all tools interactively with a web UI
The Inspector URL will be displayed in the terminal (typically http://localhost:PORT
).
Environment Variables
The run_http.sh
script automatically handles all environment variables for you. If you need to customize:
PORT
: Server port for HTTP transport (default: 6274)
Setting API Key Permanently
If you prefer to set your API key permanently instead of passing it each time:
- Create a
.env
file in the project root:
echo 'FAL_KEY="YOUR_FAL_API_KEY_HERE"' > .env
- Then run the server without the API key argument:
./run_http.sh
For manual setup:
FAL_KEY
: Your fal.ai API key (required)MCP_TRANSPORT
: Transport mode -stdio
(default) orhttp
Available Tools
models(page=None, total=None)
- List available models with optional paginationsearch(keywords)
- Search for models by keywordsschema(model_id)
- Get OpenAPI schema for a specific modelgenerate(model, parameters, queue=False)
- Generate content using a modelresult(url)
- Get result from a queued requeststatus(url)
- Check status of a queued requestcancel(url)
- Cancel a queued requestupload(path)
- Upload a file to fal.ai CDN