hnesk/whisper-wordtimestamps

openai/whisper with exposed settings for word_timestamps

Input
Configure the inputs for the AI model.

Audio file

Choose a Whisper model.

language spoken in the audio, specify None to perform language detection

0
100

optional patience value to use in beam decoding, as in https://arxiv.org/abs/2204.05424, the default (1.0) is equivalent to conventional beam search

0
100

temperature to use for sampling

optional text to provide as a prompt for the first window.

comma-separated list of token ids to suppress during sampling; '-1' will suppress most special characters except common punctuations

Extract word-level timestamps using the cross-attention pattern and dynamic time warping, and include the timestamps for each word in each segment.

0
100

if the average log probability is lower than this value, treat the decoding as failed

If word_timestamps is True, merge these punctuation symbols with the previous word

0
100

if the probability of the <|nospeech|> token is higher than this value AND the decoding has failed due to `logprob_threshold`, consider the segment as silence

If word_timestamps is True, merge these punctuation symbols with the next word

if True, provide the previous output of the model as a prompt for the next window; disabling may make the text inconsistent across windows, but the model becomes less prone to getting stuck in a failure loop

0
100

if the gzip compression ratio is higher than this value, treat the decoding as failed

0
100

temperature to increase when falling back when the decoding fails to meet either of the thresholds below

Output
The generated output will appear here.

No output yet

Click "Generate" to create an output.

whisper-wordtimestamps - ikalos.ai