This is really cool. Having offline Wikipedia + local LLMs in a single bundle is a great combo for emergency preparedness. Do you have any benchmarks on how it performs on lower-end hardware? Curious about minimum specs.