You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@yikangshen@shawntan Is there an easy way to train the model to replicate the experiments using main.py using multiple GPUs?
When using model = nn.DataParallel(model) before train(), the initialization goes into the LSTM stack and then the ONLSTM cell to return the weights but it throws an error.
We also tried doing the model = nn.DataParallel(model) after the hidden = model.init_hidden(args.batch_size) and it seems like the LinearDropConnect layer can't access the .weight tensors.
The text was updated successfully, but these errors were encountered:
@yikangshen@shawntan Is there an easy way to train the model to replicate the experiments using main.py using multiple GPUs?
When using model = nn.DataParallel(model) before train(), the initialization goes into the LSTM stack and then the ONLSTM cell to return the weights but it throws an error.
We also tried doing the model = nn.DataParallel(model) after the hidden = model.init_hidden(args.batch_size) and it seems like the LinearDropConnect layer can't access the .weight tensors.
Hi,
Just want to know have you figured this out?
Best
@yikangshen @shawntan Is there an easy way to train the model to replicate the experiments using main.py using multiple GPUs?
When using
model = nn.DataParallel(model)
beforetrain()
, the initialization goes into the LSTM stack and then the ONLSTM cell to return the weights but it throws an error.We also tried doing the
model = nn.DataParallel(model)
after thehidden = model.init_hidden(args.batch_size)
and it seems like theLinearDropConnect
layer can't access the.weight
tensors.The text was updated successfully, but these errors were encountered: