fal-api
fal.ai API Skill
Generate images, videos, and transcripts using fal.ai's API with support for FLUX, Stable Diffusion, Whisper, and more.
Features
- Queue-based async generation (submit → poll → result)
- Support for 600+ AI models
- Image generation (FLUX, SDXL, Recraft)
- Video generation (MiniMax, WAN)
- Speech-to-text (Whisper)
- Stdlib-only dependencies (no
fal_clientrequired)
Setup
- Get your API key from https://fal.ai/dashboard/keys
- Configure with:
export FAL_KEY="your-api-key"
Or via thinkfleetbot config:
thinkfleetbot config set skill.fal_api.key YOUR_API_KEY
Usage
Interactive Mode
You: Generate a cyberpunk cityscape with FLUX
Klawf: Creates the image and returns the URL
Python Script
from fal_api import FalAPI
api = FalAPI()
# Generate and wait
urls = api.generate_and_wait(
prompt="A serene Japanese garden",
model="flux-dev"
)
print(urls)
Available Models
| Model | Endpoint | Type |
|---|---|---|
| flux-schnell | fal-ai/flux/schnell |
Image (fast) |
| flux-dev | fal-ai/flux/dev |
Image |
| flux-pro | fal-ai/flux-pro/v1.1-ultra |
Image (2K) |
| fast-sdxl | fal-ai/fast-sdxl |
Image |
| recraft-v3 | fal-ai/recraft-v3 |
Image |
| sd35-large | fal-ai/stable-diffusion-v35-large |
Image |
| minimax-video | fal-ai/minimax-video/image-to-video |
Video |
| wan-video | fal-ai/wan/v2.1/1.3b/text-to-video |
Video |
| whisper | fal-ai/whisper |
Audio |
For the full list, run:
python3 fal_api.py --list-models
Parameters
| Parameter | Type | Default | Description |
|---|---|---|---|
| prompt | str | required | Image/video description |
| model | str | "flux-dev" | Model name from table above |
| image_size | str | "landscape_16_9" | Preset: square, portrait_4_3, landscape_16_9, etc. |
| num_images | int | 1 | Number of images to generate |
| seed | int | None | Random seed for reproducibility |
Credits
Built following the krea-api skill pattern. Uses fal.ai's queue-based API for reliable async generation.
More from thinkfleetai/thinkfleet-engine
local-whisper
Local speech-to-text using OpenAI Whisper. Runs fully offline after model download. High quality transcription with multiple model sizes.
149flyio-cli-public
Use the Fly.io flyctl CLI for deploying and operating apps on Fly.io: deploys (local or remote builder), viewing status/logs, SSH/console, secrets/config, scaling, machines, volumes, and Fly Postgres (create/attach/manage databases). Use when asked to deploy to Fly.io, debug fly deploy/build/runtime failures, set up GitHub Actions deploys/previews, or safely manage Fly apps and Postgres.
24kagi-search
Web search using Kagi Search API. Use when you need to search the web for current information, facts, or references. Requires KAGI_API_KEY in the environment.
22feishu-bridge
Connect a Feishu (Lark) bot to ThinkFleet via WebSocket long-connection. No public server, domain, or ngrok required. Use when setting up Feishu/Lark as a messaging channel, troubleshooting the Feishu bridge, or managing the bridge service (start/stop/logs). Covers bot creation on Feishu Open Platform, credential setup, bridge startup, macOS launchd auto-restart, and group chat behavior tuning.
13voice-transcribe
Transcribe audio files using OpenAI's gpt-4o-mini-transcribe model with vocabulary hints and text replacements. Requires uv (https://docs.astral.sh/uv/).
10video-subtitles
Generate SRT subtitles from video/audio with translation support. Transcribes Hebrew (ivrit.ai) and English (whisper), translates between languages, burns subtitles into video. Use for creating captions, transcripts, or hardcoded subtitles for WhatsApp/social media.
10