Time series classification using horovod on multi-GPUs #3657
Unanswered
Nafees-060
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am implementing some time series classification using horovod on multi-GPUs. I initially used a simple 4-layers convolutional neural network and tested training time on single and double GPUs with 4 CPUs. I have seen a overhead on two GPUs compared to a single. Then I tried increasing parameters using LSTM and got the same result. Then I tried to increase the data, still got the same issue. I am confused about whether multi-GPUs are not efficient for training or if I am doing something wrong in implementation or utilization.
Environment:
Running Command
srun --qos=gpu --gres=gpu:1 -n 2 -c 4 -C rtx2080,ubuntu18 -w gpu41--pty bash
/usr/local/bin/horovodrun -np 2 -H localhost:2 /usr/bin/python3 horov.py
Beta Was this translation helpful? Give feedback.
All reactions