https://bugs.kde.org/show_bug.cgi?id=467573
--- Comment #2 from calibre...@gmail.com --- (In reply to erjiang from comment #1) > Searching the Web, there's a similar bug reported here against Whisper: > https://github.com/openai/whisper/discussions/88 > > Seems like what you said about not using fp16 (half-precision) can work > around the issue: maybe you can try modifying the whisper code to not use > fp16 and see if that fixes it? Changing data/scripts/whispertosrt.py: line 44: result = model.transcribe(source, task=sys.argv[5], language=sys.argv[6], verbose=False, fp16 = False line 46: result = model.transcribe(source, task=sys.argv[5], verbose=False, fp16 = False) Changing data/scripts/whispertotext.py: line 47: result = model.transcribe(source, task=sys.argv[4], language=sys.argv[5], verbose=False, fp16 = False) line 49: result = model.transcribe(source, task=sys.argv[4], verbose=False, fp16 = False) This seems to fix it, GPU usage looks good, it's very fast. It took a fair bit of figuring out with the complete lack of documentation. However, this will be slower for non-16XX GPUs, a possible improvement would be to detect if it's a 16XX GPU being considered for use in order to use the different version. If you could make this into a commit, that would be great! -- You are receiving this mail because: You are watching all bug changes.