https://bugs.kde.org/show_bug.cgi?id=467573

--- Comment #2 from calibre...@gmail.com ---
(In reply to erjiang from comment #1)
> Searching the Web, there's a similar bug reported here against Whisper:
> https://github.com/openai/whisper/discussions/88
> 
> Seems like what you said about not using fp16 (half-precision) can work
> around the issue: maybe you can try modifying the whisper code to not use
> fp16 and see if that fixes it?

Changing data/scripts/whispertosrt.py:
line 44: result = model.transcribe(source, task=sys.argv[5],
language=sys.argv[6], verbose=False, fp16 = False
line 46: result = model.transcribe(source, task=sys.argv[5], verbose=False,
fp16 = False)

Changing data/scripts/whispertotext.py:
line 47: result = model.transcribe(source, task=sys.argv[4],
language=sys.argv[5], verbose=False, fp16 = False)
line 49: result = model.transcribe(source, task=sys.argv[4], verbose=False,
fp16 = False)

This seems to fix it, GPU usage looks good, it's very fast. It took a fair bit
of figuring out with the complete lack of documentation. However, this will be
slower for non-16XX GPUs, a possible improvement would be to detect if it's a
16XX GPU being considered for use in order to use the different version. If you
could make this into a commit, that would be great!

-- 
You are receiving this mail because:
You are watching all bug changes.

Reply via email to