This repository contains the code used in the paper "Distillation of Discrete Diffusion through Dimensional Correlations":
- Paper is available on arXiv.
- It was also presented at the NeurIPS 2024 Compression Workshop.
This repository is organized as follows (Section numbers follow the arXiv version):
tauldr/
contains the code for Section 5.1, which is based on tauLDR.maskgit-pytorch/
contains the code for Section 5.2, which is based on MaskGIT-pytorch.sdtt/
contains the code for Section 5.3, which is based on SDTT.
In each repository, we provide an implementation of mixture modeling on top of the teacher model and the Di4C training/inference scripts.
The Di4C-distilled model checkpoints are available on Zenodo as follows:
tldr-di4c.pt
is thestudent
model in Section 5.1 (Table 1).maskgit-di4c-d.pth
is thedi4c-d
model in Section 5.2 (Figure 3).sdtt6-di4c2.ckpt
is thesdtt-6 + di4c^2
model in Section 5.3 (Figure 4).sdtt7-di4c2.ckpt
is thesdtt-7 + di4c^2
model in Section 5.3 (Figure 4).
@article{hayakawa2024distillation,
title={Distillation of Discrete Diffusion through Dimensional Correlations},
author={Hayakawa, Satoshi and Takida, Yuhta and Imaizumi, Masaaki and Wakaki, Hiromi and Mitsufuji, Yuki},
journal={arXiv preprint arXiv:2410.08709},
year={2024}
}