In the ChatSampler, it would be good if users could resize the cache, like: ```python sampler = gm.text.ChatSampler(cache_length=1024) sampler.resize_cache(2048) ``` On Colab when reaching the full cache length, it would allow to still run more prompts without having to restart from scratch