Skip to content

train loss about training_perception #36

@Ahnhojin1223

Description

@Ahnhojin1223

Hi, thank you for sharing GoalFlow!

I'm training the perception model using run_goalflow_training_perception.sh,
but my training loss keeps oscillating and doesn't converge even after 40 epochs.
(Attached TensorBoard screenshot shows losses for agent box/class and BEV semantic.)
Image

Environment:

  • 4× V100 (32GB, NVIDIA DGX)
  • Batch size: 15 (to fully use VRAM)
  • Epochs: 40
  • Other configs: default

Questions:

  1. Is this oscillating loss behavior expected, or am I missing any training setting (LR, warmup, grad clip, etc.)?
  2. Why does the provided script use batch_size=2?
    Does GoalFlow assume a specific global batch size or gradient accumulation setting?

Thanks for any advice!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions