| ▲ | pdimitar an hour ago | |||||||
Eh, tearing down a straw man is not an impressive argument from you either. As a counter-point, LLMs still do embarrassing amounts of hallucinations, some of which are quite hilarious. When that is gone and it starts doing web searches -- or it has any mechanisms that mimic actual research when it does not know something -- then the agents will be much closer to whatever most people imagine AGI to be. Have LLMs learned to say "I don't know" yet? | ||||||||
| ▲ | in-silico 11 minutes ago | parent [-] | |||||||
> When that is gone and it starts doing web searches -- or it has any mechanisms that mimic actual research when it does not know something ChatGPT and Gemini (and maybe others) can already perform and cite web searches, and it vastly improves their performance. ChatGPT is particularly impressive at multi-step web research. I have also witnessed them saying "I can't find the information you want" instead of hallucinating. It's not perfect yet, but it's definitely climbing human percentiles in terms of reliability. I think a lot of LLM detractors are still thinking of 2023-era ChatGPT. If everyone tried the most recent pro-level models with all the bells and whistles then I think there would be a lot less disagreement. | ||||||||
| ||||||||