> openai-whisper-api

Transcribe audio via OpenAI Audio Transcriptions API (Whisper). Use when the user wants to transcribe, convert speech to text, extract words from audio or voice recordings, generate a transcript from an audio file, or perform speech recognition on m4a, ogg, or wav files using the Whisper model.

fetch
$curl "https://skillshub.wtf/elizaOS/eliza/openai-whisper-api?format=md"
SKILL.mdopenai-whisper-api

OpenAI Whisper API (curl)

Transcribe an audio file via OpenAI’s /v1/audio/transcriptions endpoint.

Quick start

{baseDir}/scripts/transcribe.sh /path/to/audio.m4a

Defaults:

  • Model: whisper-1
  • Output: <input>.txt

Useful flags

{baseDir}/scripts/transcribe.sh /path/to/audio.ogg --model whisper-1 --out /tmp/transcript.txt
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --language en
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --prompt "Speaker names: Peter, Daniel"
{baseDir}/scripts/transcribe.sh /path/to/audio.m4a --json --out /tmp/transcript.json

API key

Set OPENAI_API_KEY, or configure it in ~/.otto/otto.json:

{
  skills: {
    "openai-whisper-api": {
      apiKey: "OPENAI_KEY_HERE",
    },
  },
}

┌ stats

installs/wk0
░░░░░░░░░░
github stars17.8K
██████████
first seenMar 17, 2026
└────────────

┌ repo

elizaOS/eliza
by elizaOS
└────────────

┌ tags

└────────────