Good catch actually.
Okay maybe not exactly caveman dialect, but text compression using LLM is definitely possible to save on tokens in deep research.