Remix.run Logo
AndyNemmity 3 days ago

Before the internet we asked people around us in our sphere. If we wanted to know the answer to a question, we asked, they made up an answer, and we believed it and moved on.

Then the internet came, and we asked the internet. The internet wasn't correct, but it was a far higher % correct than asking a random person who was near you.

Now AI comes. It isn't correct, but it's far higher % correct than asking a random person near you, and often asking the internet which is a random blog page which is another random person who may or may not have done any research to come up with an answer.

The idea that any of this needs to be 100% correct is weird to me. I lived a long period in my life where everyone accepted what a random person near them said, and we all believed it.

buellerbueller 3 days ago | parent | next [-]

If you are asking random people, then your approach is incorrect. You should be asking the domain experts. Not gonna ask my wife about video games. Not gonna ask my dad about computer programming.

There, I've shaved a ton of the spread off of your argument. Possibly enough to moot the value of the AI, depending on the domain.

skybrian 3 days ago | parent | next [-]

This all assumes you have experts that you can talk to. But they might be difficult to find or expensive to hire. You wouldn't want to waste your lawyer's time on trivia.

skydhash 3 days ago | parent | next [-]

That is why experts often publish books and articles, which is then corrected by other experts (or random people if it’s a typo). I’ve read a lot of books and I haven’t met any of their authors. But I’ve still learned stuff.

skybrian 3 days ago | parent [-]

Yep. At that point you're doing research, and become familiar enough with the literature to know what's right is work.

Much like with Wikipedia, using AI to start on this journey (rather than blindly using quick answers) makes a lot of sense.

buellerbueller 2 days ago | parent | prev [-]

Most of us don't have the domain expert in range, particularly pre-internet. I was merely suggesting you ask the most expert of the domain you have access to and work your way up the tree of knowledge from there.

However, the sibling commenter about books, journals, etc., is also an excellent suggestion.

AndyNemmity 3 days ago | parent | prev [-]

Before the internet, I didn't have the phone number of domain experts to just call and ask these questions. perhaps you did. For a lot of us, it was an entirely foreign experience to have domain experts at your finger tips.

skydhash 3 days ago | parent [-]

Didn’t you have books? And teachers?

Gormo 3 days ago | parent | prev [-]

How is an LLM making stochastic inferences based on aggregations of random blog pages more likely to be correct than looking things up on decidedly non-random blog pages written by people with relevant domain knowledge?

xpe 3 days ago | parent [-]

Is the above comment a genuine question? I’m concerned it’s a rhetorical question that isn’t really getting to the heart of the matter; namely, what is the empirical performance? One’s ability to explain said performance doesn’t always keep up.

How about we pick an LLM evaluation and get specific? They have strengths and weaknesses. Some do outperform humans in certain areas.

Often I see people latching on to some reason that “proves” to them “LLMs cannot do X”. Stop and think about how powerful such a claim has to be. Such claims are masquerading as impossibility proofs.

Cognitive dissonance is a powerful force. Hold your claims lightly.

There are often misunderstandings here on HN about the kinds of things transformer based models can learn. Many people use the phrase “stochastic parrots” derisively; most of the time I think these folks are getting it badly wrong. A careful reading of the original paper is essential, not to mention follow up work.

Gormo 2 days ago | parent [-]

I'm not making a blanket statement against LLMs for all use cases. I'm certain that LLMs are, for example, much more performant at indexing already-curated documents and locating information within them than humans operating manually are.

What I'm skeptical about isn't LLMs as a utilitarian tool to enhance productivity in specific use cases, but rather treating LLMs as sources of information in their own right, especially given their defining characteristic of generating novel text through stochastic inference.

I'm 100% behind RAG powering the search engines of the future. Using LLMs to find reliable sources within the vast ocean of dubious information on the modern internet? Perfect -- ChatGPT, find me those detailed blog posts by people competent in the problem domain. Asking LLMs to come up with their own answers to questions? No thanks. That's just an even worse version of "ask a random person to make up an answer on the spot".