| ▲ | robotresearcher 5 days ago | |||||||
I tried that out in my field of expertise, to calibrate my expectations. ChatGPT invented multiple references to non-existent but plausibly-titled papers written by me. I think of that when asking questions about areas I don’t know. That was about 18mo ago, so maybe this kind of hallucination is under control these days. | ||||||||
| ▲ | wat10000 5 days ago | parent | next [-] | |||||||
LLMs are good for tasks where you can verify the result. And you must verify the result unless you're just using it for entertainment. | ||||||||
| ||||||||
| ▲ | wahnfrieden 5 days ago | parent | prev | next [-] | |||||||
I would use an agent (Codex) for this task: use the Pro model in ChatGPT for deep research and to assemble the information and citations, then have Codex systematically go through the citations with a task list to web search and verify or correct each. Codex can be used like a test suite. | ||||||||
| ▲ | SoftTalker 5 days ago | parent | prev | next [-] | |||||||
Turns out Gell-Mann amnesia applies to LLMs too. | ||||||||
| ||||||||
| ▲ | 5 days ago | parent | prev [-] | |||||||
| [deleted] | ||||||||