Skip to content

Commit

Permalink
Added gradient clipping instructions
Browse files Browse the repository at this point in the history
  • Loading branch information
armandmcqueen committed Jun 5, 2019
1 parent 1c9d4c5 commit 600cc22
Showing 1 changed file with 5 additions and 1 deletion.
6 changes: 5 additions & 1 deletion RESULTS.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,4 +16,8 @@ You can use a improved bounding box regression weight (`cfg.FRCNN.BBOX_REG_WEIGH

### SyncBN

You can use SyncBN to train with very large batch sizes without getting NaN losses. However, currently the accuracy is generally lower than when using FreezeBN and the throughput is significantly worse.
You can use SyncBN to train with very large batch sizes without getting NaN losses. However, currently the accuracy is generally lower than when using FreezeBN and the throughput is significantly worse.

### Large batch size

When training in the 32x4 configuration, you will get NaN ~5% of the time if you do not use gradient clipping. To enable gradient clipping, you need to add `TRAIN.GRADIENT_CLIP=1.5` to the config. This has a minor throughput impact, but eliminates NaN runs.

0 comments on commit 600cc22

Please sign in to comment.