Remix.run Logo
trollbridge 8 hours ago

The amount of effort to click an LLM’s sources is, what, 20 seconds? Was a human in the loop for sourcing that article at all?

phire 7 hours ago | parent | next [-]

Humans aren't very diligent in the long term. If an LLM does something correctly enough times in a row (or close enough), humans are likely to stop checking its work throughly enough.

This isn't exactly a new problem we do it with any bit of new software/hardware, not just LLMs. We check its work when it's new, and then tend to trust it over time as it proves itself.

But it seems to be hitting us worse with LLMs, as they are less consistent than previous software. And LLM hallucinations are partially dangerous, because they are often plausible enough to pass the sniff test. We just aren't used to handling something this unpredictable.

Waterluvian 7 hours ago | parent | next [-]

It’s a core part of the job and there’s simply no excuse for complacency.

jatora 7 hours ago | parent | next [-]

There's not a human alive that isnt complacent in many ways.

emmelaich 7 hours ago | parent [-]

You're being way too easy on a journalist.

nradov 6 hours ago | parent | next [-]

And too easy on the editor who was supposed to personally verify that the article was properly sourced prior to publication. This is like basic stuff that you learn working on a high school newspaper.

jatora 3 hours ago | parent | prev [-]

lol true

pixl97 7 hours ago | parent | prev | next [-]

The words on the page are just a medium to sell ads. If shit gets ad views then producing shit is part of the job... unless you're the one stepping up to cut the checks.

Marsymars 2 hours ago | parent [-]

Ars also sells ad-free subscriptions.

intended 6 hours ago | parent | prev [-]

This is a first degree expectation of most businesses.

What the OP pointed out is a fact of life.

We do many things to ensure that humans don’t get “routine fatigue”- like pointing at each item before a train leaves the station to ensure you don’t eyes glaze over during your safety check list.

This isn’t an excuse for the behavior. Its more about what the problem is and what a corresponding fix should address.

potatoman22 7 hours ago | parent | prev | next [-]

https://en.wikipedia.org/wiki/Automation_bias

zahlman 7 hours ago | parent | prev | next [-]

There's a weird inconsistency among the more pro-AI people that they expect this output to pass as human, but then don't give it the review that an outsourced human would get.

kaibee 6 hours ago | parent [-]

> but then don't give it the review that an outsourced human would get.

Its like seeing a dog play basketball badly. You're too stunned to be like "no don't sign him to <home team>".

vidarh 7 hours ago | parent | prev [-]

The irony is that while from perfect, an LLM-based fact-checking agent is likely to be far more dilligent (but still needs human review as well) by nature of being trivial to ensure it has no memory of having done a long list of them (if you pass e.g. Claude a long list directly in the same context, it is prone to deciding the task is "tedious" and starting to take shortcuts).

But at the same time, doing that makes it even more likely the human in the loop will get sloppy, because there'll be even fewer cases where their input is actually needed.

I'm wondering if you need to start inserting intentional canaries to validate if humans are actually doing sufficiently torough reviews.

prussia 7 hours ago | parent | prev | next [-]

The kind of people to use LLM to write news article for them tend not to be the people who care about mundane things like reading sources or ensuring what they write has any resemblance to the truth.

adamddev1 an hour ago | parent | prev | next [-]

The problem is that the LLM's sources can be LLM generated. I was looking up some health question and tried clicking to see the source for one of the LLMs claim. The source was a blog post that contained an obvious hallucination or false elaboration.

kortilla 7 hours ago | parent | prev [-]

The source would just be the article, which the Ars author used an LLM to avoid reading in the first place.