这是indexloc提供的服务,不要输入任何密码
Skip to content

Conversation

@mattmacy
Copy link

This re-factoring may well be too intrusive for your taste. I wanted to quickly test DataParallel support on a network that I knew worked well before applying it to my own work. I hope it's useful to you.

  • re-factor the DenseNet constructor so it can be passed to nn.parallel.data_parallel
  • scale batch size with the number of gpus
  • make plot.py robust to larger batch sizes
  • move weight initialization to train.py and add optional kaiming fan-in initialization

Matt Macy added 2 commits March 15, 2017 18:05
- scale batch size with ngpu
- call plot.py with the batchsize so that it can scale it's rolling N
  value accordingly (and not break when batchsize is greater than 64)
- add optional kaiming initialization
- do weight initialization in train.py as opposed to the network
  constructor
@mattmacy mattmacy closed this Mar 16, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant