Skip to content

We introduce UltraLLaDA , a scaled variant of LLaDA-8B-Base that extends the context length up to 128K tokens with light-weight post-training, enabling long-context comprehension and generation.

License

Notifications You must be signed in to change notification settings

Relaxed-System-Lab/UltraLLaDA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

UltraLLaDA

Hugging Face

We introduce UltraLLaDA , a scaled variant of LLaDA-8B-Base that extends the context length up to 128K tokens with light-weight post-training, enabling long-context comprehension and generation.

Eval

export PYTHONPATH=/path/to/eval/opencompass-0.4.2:$PYTHONPATH && \
export TIKTOKEN_CACHE_DIR=/path/to/eval/tiktoken/tiktoken_cache && \
export COMPASS_DATA_CACHE='/path/to/eval/data/'
  • lonbench:

     python3 run.py /path/to/eval/opencompass-0.4.2/opencompass/eval/longbench/16K/eval_llada_long.py --dump-eval-details -r
  • NIAH:

    python3 run.py /path/to/eval/opencompass-0.4.2/opencompass/eval/niah/128K/eval_llada_niah.py --dump-eval-details -r
  • ruler:

    python3 run.py /path/to/eval/opencompass-0.4.2/opencompass/eval/ruler/32K/eval_llada_ruler.py --dump-eval-details -r

About

We introduce UltraLLaDA , a scaled variant of LLaDA-8B-Base that extends the context length up to 128K tokens with light-weight post-training, enabling long-context comprehension and generation.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages