Should running on a huge NVIDIA GPU increase transcription speed? #1640
-
I have about a PB of voice data to transcribe. Naturally, based on Whisper's relative model speed documentation, I opted to pay for a huge NVIDIA A100 40GB VRAM instance in GCP to massively speed up the amount it will take me to transcribe all the data. I have had zero issues:
The thing is: when I go to transcribe it's slow as all hecc! It does transcribe, but the GPU appears to make no difference. Is this how it is supposed to work? It runs at the same speed as when I'm transcribing on my Mac! I'm not doing anything complicated:
The file I'm passing is about So yeah I'm lost. Here's some pics of the VM specs below: |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 9 replies
-
Try |
Beta Was this translation helpful? Give feedback.
-
@jsteinberg-rbi Have you tried to run several process(eg. various commands of whisper trancribing ) in paralell or at same time ?.. Im looking for a server with this capability but I dont know if it is feasabilty before buying the server. I tested with RTX 2060 but it crashes when run more than 1 process. Thks in advance. |
Beta Was this translation helpful? Give feedback.
Try
whisper foo.wav --model large-v2 --language English --device cuda