▲ | NothingAboutAny 4 days ago | |||||||||||||||||||||||||||||||||||||
I tried to use perplexity to find ideal settings for my monitor, it responded with concise list of distinct settings and why. When I investigated the source it was just people guessing and arguing with each other in the Samsung forums, no official or even backed up information. I'd love if it had a confidence rating based on the sources it found or something, but I imagine that would be really difficult to get right. | ||||||||||||||||||||||||||||||||||||||
▲ | Moosdijk 4 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||
I asked gemini to do a deep research on the role of healthcare insurance companies in the decline of general practicioners in the Netherlands. It based its premise mostly on blogs and whitepapers on company websites, who's job it is to sell automation-software. AI really needs better source-validation. Not just to combat the hallucination of sources (which gemini seems to do 80% of the time), but also to combat low quality sources that happen to correlate well to the question in the prompt. It's similar to Google having to fight SEO spam blogs, they now need to do the same in the output of their models. | ||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||
▲ | ugh123 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
Seems like the right outcome was had, by reviewing sources. I wish it went one step further and loaded those source pages and scroll/highlight the snippets where it pulled information from. That way we can easily double check at least some aspects of it's response, and content+ads can be attributed to the publisher. | ||||||||||||||||||||||||||||||||||||||
▲ | stocksinsmocks 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
In the absence of easily found authoritative information from the manufacturer, this would have been my source of information. Internet banter might actually be the best available information. | ||||||||||||||||||||||||||||||||||||||
▲ | wodenokoto 4 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
But the really tricky thing is, that sometimes it _is_ these kinds of forums where you find the best stuff. When LLMs really started to show themselves, there was a big debate about what is truth, with even HN joining in on heated debates on the number of sexes or genders a dog may have and if it was okay or not for ChatGPT to respond with a binary answer. On one hand, I did found those discussions insufferable, but the deeper question - what is truth and how do we automated the extraction of truth from corpora - is super important and somehow completely disappeared from the LLM discourse. | ||||||||||||||||||||||||||||||||||||||
▲ | simonw 4 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||
It would be interesting to see if that same question against GPT-5 Thinking produces notably better results. |