Releases: BBC-Esq/WhisperS2T-transcriber
Releases · BBC-Esq/WhisperS2T-transcriber
v1.3.3 - WhisperS2T-transcriber
v1.3.2 - user-friendliness
Changes
- The comboboxes are now only populated with the device and whisper model precisions that your system supports, and are updated dynamically if you change the compute device.
v1.3.1 - better-faster-stronger
Improvements
- Add checks for ensuring the correct Whisper model is selected depending on whether CPU or GPU is selected.
- Adjust how many CPU cores are utilized for the best performance.
v1.3.0 - better, faster, stronger
Improvements
- Added distil whisper models as well as
Large-v3
model variants. - Significantly improved the installation speed by using the
uv
library. - Users now choose the installation method of GPU support or Cpu-only.
- You no longer have to separately install
CUDA
on your computer. The GPU installation willpip install
into the virtual environment.
Removals
- Removed
Python 10
support.
v1.2.1 - user friendler fixed
Minor release to address a situation where a user only intends to use cpu, in which case the installation script will not install pynvml nor will the scripts attempt to use it (e.g. the metrics bar).
- Also added a feature to calculate the number of files to be processed and ask a user's permission before proceeding.
v1.2.0 - user friendlier
- Revise GUI
- Add option to only process certain file types.
- Implement stop processing midstream.
- Offload settings to separate
settings.py
script.
v1.1.0 - better/faster/stronger
- CPU support with automatic thread management.
- Process each file individually instead of in a list sent to ctranslate2, which enables error handling for files that can't be processed for any reason. Previously, if one file failed the all other files failed.
- Simplified installation script
setup_windows.py
. - Increase speed max to 200 batch size (USE WITH CAUTION - large-v2 model only supports approximately 21 speed on a GPU with 24 GB of VRAM. You'll have to experiment with the speed setting for smaller Whisper models).
- Revise GUI and add more user-friendly messages.
TO-DO:
- Figure out why the model still isn't being released from memory like it was mid-release versions...You currently have to close the program (not a big deal) to release VRAM memory...).
v1.0.0 - fastest ctranslate2 transcriber
This is the fastest ctranslate2 transcriber while still maintaining superior quality as well. Hats off to the WhisperS2T library for someone finally implementing batch processing with the ctranslate2 library.