Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to train the network in three GPUs #3

Open
XGBoost opened this issue Apr 12, 2019 · 1 comment
Open

How to train the network in three GPUs #3

XGBoost opened this issue Apr 12, 2019 · 1 comment

Comments

@XGBoost
Copy link

XGBoost commented Apr 12, 2019

I have noticed you have mentioned that it takes four hours to finish the training on three NVIDIA GTX 1080 Ti GPUs. However, you do not describe how to train the network on three GPUs in README.md.
When I run
python train.py --id resnet50_rnn --use_rnn
, it only takes a single GPU, and the batch size of it is eight which is different from that mentioned in your paper.
Could you please describe the process of training in detail.

@sunset1995
Copy link
Owner

Hi @XGBoost
To train on multiple GPU, you have to modify few lines of code.
You can check https://pytorch.org/tutorials/beginner/former_torchies/parallelism_tutorial.html
This repo only work on single GPU.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants