Hi,
does anyone here have experience with https://github.com/openai/whisper, I tried to transcribe mp3 length 13:35 min on different hardware with following results :
i used LARGE MODEL :
1) AC922 GPU 1xV100 utilization (U)=30 % , transcription time (TS) = 5:38 min
2) AC922 32xP9 CPU U=90 + %, TS= 80 min
3) 1050 p311 pytorch-cpu 2.1.1 8CPU TS=28 min U = more less 4.5 cores only few threads work
4) 1050 p311 pytorch-cpu 2.1.1 12CPU TS=25 min
in 4) i used
export OPENBLAS_NUM_THREADS=12
export GOTO_NUM_THREADS=12
export OMP_NUM_THREADS=12
I found out that by incrasing the number of threads (export OPENBLAS....) I can increase CPU utilization but this does not shorten the transition time , but rather increases it
Is it possible to get closer to V100 with some tuning or this is the best which can i expect from p10 with this quite large model ?
Is it normal that with pytorch-cpu 2.1.2 py311_1 rocketce i got this message ?:
/data/miniconda3/lib/python3.11/site-packages/whisper/transcribe.py:126: UserWarning: FP16 is not supported on CPU; using FP32 instead
warnings.warn("FP16 is not supported on CPU; using FP32 instead")
Error in cpuinfo: processor architecture is not supported in cpuinfo
Error in cpuinfo: processor architecture is not supported in cpuinfo
thanks
Tomas
------------------------------
Tomas Kovacik
Rocket Forum Shared Account
------------------------------