You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
It seems like by default, the sentence order prediction weights of ALBERT models are randomly initialized (huggingface/transformers#12196). Did you use some custom export of the ALBERT model for your project? The logs below confirm the random initialization of the ALBERT SOP head when I run the eval/coherence.py file.
(tables-venv) kalpesh@node105:CoRPG$ python eval/coherence.py --coh --pretrain_model albert-xxlarge-v2 --text_file abc.txt
Some weights of AlbertForPreTraining were not initialized from the model checkpoint at albert-xxlarge-v2 and are newly initialized: ['sop_classifier.classifier.weight', 'sop_classifier.classifier.bias']
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 5/5 [00:00<00:00, 54.82it/s]
COH-p: 0.327211 COH: 0.000000
(tables-venv) kalpesh@node105:CoRPG$
The text was updated successfully, but these errors were encountered:
The ALBERT model provided by huggingface does not include the parameters of the final classification layer, and transformers will random initial its parameters. So you should finetune the ALBERT with your training dataset. And we also release our finetuned ALBERT on the release page.
NOTE: You should only leverage the training dataset rather than the whole dataset include validation and test dataset to finetune the ALBERT model, this can make sure the accuracy and fairness of the COH and COH-p.
Hi,
It seems like by default, the sentence order prediction weights of ALBERT models are randomly initialized (huggingface/transformers#12196). Did you use some custom export of the ALBERT model for your project? The logs below confirm the random initialization of the ALBERT SOP head when I run the
eval/coherence.py
file.The text was updated successfully, but these errors were encountered: