openai-whisper-api
Originally fromsteipete/clawdis
SKILL.md
OpenAI Whisper API (curl)
Transcribe an audio file via OpenAI’s /v1/audio/transcriptions endpoint.
Quick start
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a
Defaults:
- Model:
whisper-1 - Output:
<input>.txt
Useful flags
{baseDir}/scripts/transcribe.sh /path/to/audio.ogg --model whisper-1 --out /tmp/transcript.txt
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --language en
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --prompt "Speaker names: Peter, Daniel"
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --json --out /tmp/transcript.json
API key
Set OPENAI_API_KEY, or configure it in ~/.clawdbot/moltbot.json:
{
skills: {
"openai-whisper-api": {
apiKey: "OPENAI_KEY_HERE"
}
}
}
Weekly Installs
3
Repository
composiohq/open…composioGitHub Stars
11
First Seen
7 days ago
Security Audits
Installed on
openclaw3
gemini-cli3
github-copilot3
codex3
kimi-cli3
cursor3