Skip to content

Non-record: AWQ 2xH100 proxy no-compile quantized eval#1976

Open
Devchandrasen wants to merge 1 commit intoopenai:mainfrom
Devchandrasen:awq2gpu-nonrecord-20260430
Open

Non-record: AWQ 2xH100 proxy no-compile quantized eval#1976
Devchandrasen wants to merge 1 commit intoopenai:mainfrom
Devchandrasen:awq2gpu-nonrecord-20260430

Conversation

@Devchandrasen
Copy link
Copy Markdown

Non-record submission for the 16MB track.

This adds a 2xH100 proxy run based on the PR #1956 / PR #1908 AWQ + GPTQ stack. It is not intended as a leaderboard/SOTA claim because it was run on 2xH100 rather than the official 8xH100 setup.

Result:

  • val_bpb: 1.15828615
  • val_loss: 2.53478079
  • total counted bytes: 15,998,289
  • train time: 599917 ms
  • seed: 42

The original run produced an under-16MB artifact. On my PyTorch 2.8.0 + CUDA 12.8 HPC setup, compiled quantized evaluation crashed after decompression, so the included train_gpt.py adds an environment-gated no-compile quantized eval path. The saved artifact was then reloaded and evaluated successfully with exit status 0.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant