New CUDA version #50
Replies: 12 comments 39 replies
-
I did not have any trouble in installing or starting the process of generating a transcript (CPU usage remained below 20%). The file I tried was not in English and I faced the issue you listed in known issues (loop of repeating text) after the 10th minute in a 30 minute audio file. |
Beta Was this translation helpful? Give feedback.
-
Any special specific install instructions on how to install and try this version? |
Beta Was this translation helpful? Give feedback.
-
@VBodrov reported in another thread:
Interesting. There is a good chance that 4GB of VRAM are not sufficient for the large whisper model. You can try the "fast" setting in noScribe. This will use a much smaller model. The quality is not great. But this would confirm that in fact the size of the model is the problem. |
Beta Was this translation helpful? Give feedback.
-
It's really simple
Done
The gif from earliers is from my desktop this is the result from the laptop on which I tried transcribing earlier with the latest drivers installed
I would still need to install the CUDA drivers like I mentioned above because I just tested it and noscribe on the laptop does not use the GPU when transcribing. I just tested it again. I am going to install the CUDA drivers now and report back :) |
Beta Was this translation helpful? Give feedback.
-
Thanks! Works like charm for me! Die Schatzinsel (342 min) works for me on (But i think, I installed Cuda by myself before on all three Systems) Also diarizing used cuda. Just typical small bugs (setting rights on win11, language on auto -> always translating to english automatically) Speech-Translate (https://github.com/Dadangdut33/Speech-Translate (also faster whisper)) did not work on laptop w/ RTX_A500 |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
it works! I will make some suggestions that will be useful 1.Upload video in batch.
As soon as I have some time to do a comparative test CPU Xeon 14 cores 128 GB RAM vs 3060 12gb THANKS! |
Beta Was this translation helpful? Give feedback.
-
Thank you very much for this release ! On my side, it did work better (went from 19mn to transcribe 5mn audio to 4-6mn), and it seems the GPU is working to some extent. Also checked the yml file, everything is correct. I also added the PATH.
Thank you for your help ! |
Beta Was this translation helpful? Give feedback.
-
The 0.4.5 Cuda Version runs like a charm and the quality of transcription results compete with amberscript. Thanks a lot for this release, the CUDA acceleration gains a big performance boost. The CUDA transcription speeds scales massive with RAM speed/timings/bandwith. System 1: Intel i7-13700k, 16GB DDR5-4400 CL36 RAM (One Module - single channel), NVidia RTX4070 System 2.1: AMD Threadripper1950x, 32GB DDR4-2133 CL15 (Four Modules - quad channel), NVidia RTX4070 System 2.2: AMD Threadripper1950x, 32GB DDR4-3600 CL17 (Four Modules - quad channel), NVidia RTX4070 Maybe this should be mentioned in the recomended hardware section, I've spend some hours to get it why the much recent Intel system is much slower than our old Threadripper workstation. |
Beta Was this translation helpful? Give feedback.
-
Each instance consumes an average of 3.5 GB of VRAM. On the 3060 12gb there are times when the machine freezes with three. Any plans to get batch processing one by one in order? |
Beta Was this translation helpful? Give feedback.
-
Hello! This new version is running super smooth on my computer compared to the previous version of noScribe. Before, noScribe would freeze while it was transcribing and it was really terrifying but it always got the transcription done. I only had that weird repeating word issue happen once. This new version can run in the background while I do other work so big fan! I'm running an old Alienware from 2016 with a GTX 1060. Thanks again for making my PhD life easier!! |
Beta Was this translation helpful? Give feedback.
-
I observed a very strange behaviour in the GUI that led to an increase in the speed of the embedding and segmentation process by approximately 50%. For some reason, when minimizing and maximizing the NoScibe window during the embedding and segmentation part, it appears to utilize more cores and result in a time reduction of roughly half. I tested this with a 15-minute audio file, comparing waiting versus minimizing and maximizing the NoScibe window. The results were striking: it took 6 minutes to wait for the process to complete vs. just 3 minutes when toggling the NoScibe window. Initially, I thought that Python's buffering might be slow and only load the UI when in focus, but the time difference tells a different story. |
Beta Was this translation helpful? Give feedback.
-
I've compiled a new version that should support acceleration via NVIDIA graphics cards using cuda. It's not officially released yet. Could you test and report back if it works? Thanks!
Installer here: https://drive.switch.ch/index.php/s/EIVup04qkSHb54j?path=%2FnoScribe%20vers.%200.4.5%2FWindows%2Fcuda%20version
EDIT (09.04.2024): It appears to function quite well. Here are some tips gathered from the discussion below:
Beta Was this translation helpful? Give feedback.
All reactions