Remix.run Logo
lblume 18 hours ago

Imagine telling a person from five years ago that the programs that would basically solve NLP, perform better than experts at many tasks and are hard not to anthropomorphize accidentally are actually "trivial". Good luck with that.

hyperbovine 4 hours ago | parent | next [-]

It still doesn't pass the Turing test, and is not close. Five years ago me would be impressed but still adamant that this is not AI, nor is it on the path to AI.

jrflowers 17 hours ago | parent | prev | next [-]

>programs that would basically solve NLP

There is a load-bearing “basically” in this statement about the chat bots that just told me that the number of dogs granted forklift certification in 2023 is 8,472.

lblume 17 hours ago | parent [-]

Sure, maybe solving NLP is too great a claim to make. It is still not at all ordinary that beforehand we could not solve referential questions algorithmically, that we could not extract information from plain text into custom schemas of structured data, and context-aware mechanical translation was really unheard of. Nowadays LLMs can do most of these tasks better than most humans in most scenarios. Many NLP questions at least I find interesting reduce to questions of the explanability of LLMs.

clarinificator 17 hours ago | parent | prev | next [-]

Yeah it solved NLP about 50% of the time, and also mangles data badly and in often hard-to-detect ways.

Applejinx 13 hours ago | parent | prev [-]

"hard not to anthropomorphize accidentally' is a you problem.

I'm unhappy every time I look in my inbox, as it's a constant reminder there are people (increasingly, scripts and LLMs!) prepared to straight-up lie to me if it means they can take my money or get me to click on a link that's a trap.

Are you anthropomorphizing that, too? You're not gonna last a day.

lblume 12 hours ago | parent [-]

I didn't mean typical chatbot output, these are luckily still fairly recognizable due to stylistic preferences learned during fine-tuning. I mean actual base model output. Take a SOTA base model and give it the first two paragraphs of some longer text you wrote, and I would bet on many people being unable to distinguish your continuation from the model's autoregressive guesses.