▲ | psadri 5 days ago | |||||||||||||||||||||||||||||||||||||
I do miss the earlier "heavy" models that had encyclopedic knowledge vs the new "lighter" models that rely on web search. Relying on web search surfaces a shallow layer of knowledge (thanks to SEO and all the other challenges of ranking web results) vs having ingested / memorized basically the entirety of human written knowledge beyond what's typically reachable within the first 10 results of a web search (eg: digitized offline libraries). | ||||||||||||||||||||||||||||||||||||||
▲ | hamdingers 4 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||
I feel the opposite. Before I can use information from a model's "internal" knowledge I have to engage in independent research to verify that it's not a hallucination. Having an LLM generate search strings and then summarize the results does that research up front and automatically, I need only click the sources to verify. Kagi Assistant does this really well. | ||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||
▲ | mastercheif 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
I kept search off for a long time due to it tanking the quality of the responses from ChatGPT. I recently added the following to my custom instructions to get the best of both worlds: # Modes When the user enters the following strings you should follow the following mode instructions: 1. "xz": Use the web tool as needed when developing your answer. 2. "xx": Exclusively use your own knowledge instead of searching the internet. By default use mode "xz". The user can switch between modes during a chat session. Stay with the current mode until the user explicitly switches modes. | ||||||||||||||||||||||||||||||||||||||
▲ | ants_everywhere 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
Most real knowledge is stored outside the head, so intelligent agents can't rely solely on what they've remembered. That's why libraries are so fundamental to universities. | ||||||||||||||||||||||||||||||||||||||
▲ | stephen_cagle 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
I think this is partially something I have felt myself as well. It would be interesting if these lighter web search models would highlight the distinction between information that has been seen elsehwere vs information that is novel for each page? Like, a view that lets me look at the things that have been asserted and see how many of the different pages show those facts asserted (vs unmentioned vs contradicted). | ||||||||||||||||||||||||||||||||||||||
▲ | simianwords 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
There is a tradeoff here: the non search models are internally heavy but the search models are light but also depend on real data. I keep switching between both but I think I'm starting to prefer the lighter one that is based on the sources instead. | ||||||||||||||||||||||||||||||||||||||
▲ | killerstorm 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
These models are still available: GPT-4.5, Gemini 2.5 Pro (at least the initial version - not sure if they optimized it away). From what I can tell, they are pretty damn big. Grok 4 is quite large too. | ||||||||||||||||||||||||||||||||||||||
▲ | gerdesj 4 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||
"encyclopedic knowledge" Have you just hallucinated that? |