in order to have research agents that can run for days, we need context compaction
i used RL to have LLMs naturally learn their own 10x compression! Qwen learned to pack more info per token (ie use Mandarin tokens, prune text)
read the technical blog: https://t.co/3pzRvt4zvr
rajan agarwalx.com