On this page
article
đ Audio to text
Audio to text models are models that can generate text from an audio file.
The transcription endpoint allows to convert audio files to text. The endpoint is based on whisper.cpp, a C++ library for audio transcription. The endpoint input supports all the audio formats supported by ffmpeg
.
Usage
Once LocalAI is started and whisper models are installed, you can use the /v1/audio/transcriptions
API endpoint.
For instance, with cURL:
Example
Download one of the models from here in the models
folder, and create a YAML file for your model:
The transcriptions endpoint then can be tested like so:
Last updated a year ago.