You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
on_chunk_start / on_chunk_end are not called when using return_timestamps: "word".
These callbacks only provide timestamps with return_timestamps: true
I also tried to decode tokens, as I’ve seen it in the demo, but that uses callbacks that no longer exist (e.g. chunk_callback(chunk) and callback_function(item))
The algorithm we use to compute word-level timestamps (dynamic time warping) requires the entire chunk to be processed, so streamed word-level timestamps isn't currently possible.
Since the model is by OpenAI, I wonder if the API could be aligned towards their approach of providing "segment", "word" or both to an array of timestamp_granularities.
The timestamp granularities to populate for this transcription. response_format must be set verbose_json to use timestamp granularities. Either or both of these options are supported: word, or segment. Note: There is no additional latency for segment timestamps, but generating word timestamps incurs additional latency.
Question
Goal
Issue
on_chunk_start
/on_chunk_end
are not called when usingreturn_timestamps: "word"
.These callbacks only provide timestamps with
return_timestamps: true
I also tried to decode tokens, as I’ve seen it in the demo, but that uses callbacks that no longer exist (e.g.
chunk_callback(chunk)
andcallback_function(item)
)Setup
Decoding works, but timestamps are null.
The text was updated successfully, but these errors were encountered: