Remix.run Logo
JBorrow 5 hours ago

From my perspective as a journal editor and a reviewer these kinds of tools cause many more problems than they actually solve. They make the 'barrier to entry' for submitting vibed semi-plausible journal articles much lower, which I understand some may see as a benefit. The drawback is that scientific editors and reviewers provide those services for free, as a community benefit. One example was a submission their undergraduate affiliation (in accounting) to submit a paper on cosmology, entirely vibe-coded and vibe-written. This just wastes our (already stretched) time. A significant fraction of submissions are now vibe-written and come from folks who are looking to 'boost' their CV (even having a 'submitted' publication is seen as a benefit), which is really not the point of these journals at all.

I'm not sure I'm convinced of the benefit of lowering the barrier to entry to scientific publishing. The hard part always has been, and always will be, understanding the research context (what's been published before) and producing novel and interesting work (the underlying research). Connecting this together in a paper is indeed a challenge, and a skill that must be developed, but is really a minimal part of the process.

SchemaLoad 3 hours ago | parent | next [-]

GenAI largely seems like a DDoS on free resources. The effort to review this stuff is now massively more than the effort to "create" it, so really what is the point of even submitting it, the reviewer could have generated it themself. Seeing it in software development where coworkers are submitting massive PRs they generated but hardly read or tested. Shifting the real work to the PR review.

I'm not sure what the final state would be here but it seems we are going to find it increasingly difficult to find any real factual information on the internet going forward. Particularly as AI starts ingesting it's own generated fake content.

cryzinger 2 hours ago | parent | next [-]

More relevant than ever:

> The amount of energy needed to refute bullshit is an order of magnitude bigger than that needed to produce it.

https://en.wikipedia.org/wiki/Brandolini%27s_law

trees101 19 minutes ago | parent | next [-]

The P≠NP conjecture in CS says checking a solution is easier than finding one. Verifying a Sudoku is fast; solving it from scratch is hard. But Brandolini's Law says the opposite: refuting bullshit costs way more than producing it.

Not actually contradictory. Verification is cheap when there's a spec to check against. 'Valid Sudoku?' is mechanical. But 'good paper?' has no spec. That's judgment, not verification.

monkaiju 2 hours ago | parent | prev [-]

Wow the 3 comments from OC to here are all bangers, they combine into a really nice argument against these toys

overfeed 2 hours ago | parent | prev | next [-]

> The effort to review this stuff is now massively more than the effort to "create" it

I don't doubt the AI companies will soon announce products that will claim to solve this very problem, generating turnkey submission reviews. Double-dipping is very profitable.

It appears LLM-parasitism isn't close to being done, and keeps finding new commons to spoil.

Spivak 2 hours ago | parent | prev | next [-]

In some ways it might be a good thing that shorthand signals of quality are being destroyed because it forces all of us to meaningfully engage with the work. No more LGTM +1 when every PR looks good.

toomuchtodo 2 hours ago | parent | prev [-]

AI slop security reports submitted to curl - https://gist.github.com/bagder/07f7581f6e3d78ef37dfbfc81fd1d...

HN Search: curl AI slop - https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...

Cornbilly an hour ago | parent [-]

This one is hilarious. https://hackerone.com/reports/3516186

If I submitted this, I'd have to punch myself in the face repeatedly.

toomuchtodo 36 minutes ago | parent [-]

The great disappointment is that the humans submitting these just don’t care it’s slop and they’re wasting another human’s time. To them, it’s a slot machine you just keep cranking the arm of until coins come out. “Prompt until payout.”

InsideOutSanta 4 hours ago | parent | prev | next [-]

I'm scared that this type of thing is going to do to science journals what AI-generated bug reports is doing to bug bounties. We're truly living in a post-scarcity society now, except that the thing we have an abundance of is garbage, and it's drowning out everything of value.

willturman 3 hours ago | parent | next [-]

In a corollary to Sturgeon's Law, I'd propose Altman's Law: "In the Age of AI, 99.999...% of everything is crap"

SimianSci 2 hours ago | parent [-]

Altman's Law: 99% of all content is slop

I can get behind this. This assumes a tool will need to be made to help determine the 1% that isn't slop. At which point I assume we will have reinvented web search once more.

Has anyone looked at reviving PageRank?

Imustaskforhelp 2 hours ago | parent [-]

I mean Kagi is probably the PageRank revival we are talking about.

I have heard from people here that Kagi can help remove slop from searches so I guess yeah.

Although I guess I am DDG user and I love using DDG as well because its free as well but I can see how for some price can be a non issue and they might like kagi more.

So Kagi / DDG (Duckduckgo) yeah.

jll29 2 hours ago | parent [-]

Does anyone have kept an eye of who uses what back-end?

DDG used to be meta-search on top of Yahoo, which doesn't exist anymore. What do Gabriel and co-workers use now?

selectodude 2 hours ago | parent [-]

I think they all use Bing now.

techblueberry 3 hours ago | parent | prev | next [-]

There's this thing where all the thought leaders in software engineering ask "What will change about building about building a business when code is free" and while, there are some cool things, I've also thought, like it could have some pretty serious negative externalities? I think this question is going to become big everywhere - business, science, etc. which is like - Ok, you have all this stuff, but do is it valuable? Which of it actually takes away value?

SequoiaHope 2 hours ago | parent [-]

To be fair, the question “what will change” does not presume the changes will be positive. I think it’s the right question to ask, because change is coming whether we like it or not. While we do have agency, there are large forces at play which impact how certain things will play out.

jplusequalt 3 hours ago | parent | prev | next [-]

Digital pollution.

jll29 2 hours ago | parent | prev | next [-]

Soon, poor people will talk to a LLM, rich people will get human medical care.

Spivak 2 hours ago | parent [-]

I mean I'm currently getting "expensive" medical care and the doctors are still all using AI scribes. I wouldn't assume there would be a gap in anything other than perception. I imagine doctors that cater to the fuck you rich will just put more effort into hiding it.

No one, at all levels, wants to do notes.

jcranmer 3 hours ago | parent | prev [-]

The first casualty of LLMs was the slush pile--the unsolicited submission pile for publishers. We've since seen bug bounty programs and open source repositories buckle under the load of AI-generated contributions. And all of these have the same underlying issue: the LLM makes it easy to do things that don't immediately look like garbage, which makes the volume of submission skyrocket while the time-to-reject also goes up slightly because it passes the first (but only the first) absolute garbage filter.

storystarling 2 hours ago | parent [-]

I run a small print-on-demand platform and this is exactly what we're seeing. The submissions used to be easy to filter with basic heuristics or cheap classifiers, but now the grammar and structure are technically perfect. The problem is that running a stronger model to detect the semantic drift or hallucinations costs more than the potential margin on the book. We're pretty much back to manual review which destroys the unit economics.

jll29 2 hours ago | parent | prev | next [-]

I totally agree. I spend my whole day from getting up to going to bed (not before reading HN!) on reviews for a conference I'm co-organizing later this year.

So I was not amused about this announcement at all, however easy it may make my own life as an author (I'm pretty happy to do my own literature search, thank you very much).

Also remember, we have no guarantee that these tools will still exist tomorrow, all these AI companies are constantly pivoting and throwing a lot of things at the wall to see what sticks.

OpenAI chose not to build a serious product, as there is no integration with the ACM DL, the IEEE DL, SpringerNatureLink, the ACL Anthology, Wiley, Cambridge/Oxford/Harvard University Press etc. - only papers that are not peer reviewed (arXiv.org) are available/have been integrated. Expect a flood of BS your way.

When my student submit a piece of writing, I can ask them to orally defend their opus maximum (more and more often, ChatGPT's...); I can't do the same with anonymous authors.

bloppe 5 hours ago | parent | prev | next [-]

I wonder if there's a way to tax the frivolous submissions. There could be a submission fee that would be fully reimbursed iff the submission is actually accepted for publication. If you're confident in your paper, you can think of it as a deposit. If you're spamming journals, you're just going to pay for the wasted time.

Maybe you get reimbursed for half as long as there are no obvious hallucinations.

JBorrow 4 hours ago | parent | next [-]

The journal that I'm an editor for is 'diamond open access', which means we charge no submission fees and no publication fees, and publish open access. This model is really important in allowing legitimate submissions from a wide range of contributors (e.g. PhD students in countries with low levels of science funding). Publishing in a traditional journal usually costs around $3000.

NewsaHackO 4 hours ago | parent | next [-]

Those journals are really good for getting practice in writing and submitting research papers, but sometimes they are already seen as less impactful because of the quality of accepted papers. At least where I am at, I don't think the advent of AI writing is going to affect how they are seen.

agnishom 5 minutes ago | parent [-]

In the field of Programming Languages and Formal Methods, many of the top journals and conference proceedings are open access

methuselah_in 4 hours ago | parent | prev [-]

Welcome to new world of fake stuff i guess

willturman an hour ago | parent | prev | next [-]

If the penalty for a crime is a fine, then that law exists only for the lower class

In other words, such a structure would not dissuade bad actors with large financial incentives to push something through a process that grants validity to a hypothesis. A fine isn't going to stop tobacco companies from spamming submissions that say smoking doesn't cause lung cancer or social media companies from spamming submissions that their products aren't detrimental to the mental health.

s0rce 4 hours ago | parent | prev | next [-]

That would be tricky, I often submitted to multiple high impact journals going down the list until someone accepted it. You try to ballpark where you can go but it can be worth aiming high. Maybe this isn't a problem and there should be payment for the efforts to screen the paper but then I would expect the reviewers to be paid for their time.

noitpmeder 4 hours ago | parent [-]

I mean your methodology also sounds suspect. You're just going down a list until it sticks. You don't care where it ends up (I'm sure within reason) just as long as it is accepted and published somewhere (again, within reason).

niek_pas 4 hours ago | parent | next [-]

Scientists are incentivized to publish in as high-ranking a journal as possible. You’re always going to have at least a few journals where your paper is a good fit, so aiming for the most ambitious journal first just makes sense.

antasvara 2 hours ago | parent | prev | next [-]

No different from applying to jobs. Much like companies, there are a variety of journals with varying levels of prestige or that fit your paper better/worse. You don't know in advance which journals will respond to your paper, which ones just received submissions similar to yours, etc.

Plus, the t in me from submission to acceptance/rejection can be long. For cutting edge science, you can't really afford to wait to hear back before applying to another journal.

All this to say that spamming 1,000 journals with a submission is bad, but submitting to the journals in your field that are at least decent fits for your paper is good practice.

jll29 2 hours ago | parent | prev | next [-]

It's standard practice, nothing suspect about their approach - and you won't go lower and lower and lower still because at some point you'll be tired of re-formatting, or a doctoral candidate's funding will be used up, or the topic has "expired" (= is overtaken by reality/competition).

mathematicaster 3 hours ago | parent | prev [-]

This is effectively standard across the board.

throwaway85825 4 hours ago | parent | prev | next [-]

Pay to publish journals already exist.

bloppe 4 hours ago | parent | next [-]

This is sorta the opposite of pay to publish. It's pay to be rejected.

olivia-banks 4 hours ago | parent | prev [-]

I would think it would act more like a security deposit, and you'd get back 100%, no profit for the journal (at least in that respect).

pixelready 4 hours ago | parent | prev | next [-]

I’d worry about creating a perverse incentive to farm rejected submissions. Similar to those renter application fee scams.

petcat 4 hours ago | parent | prev | next [-]

> There could be a submission fee that would be fully reimbursed if the submission is actually accepted for publication.

While well-intentioned, I think this is just gate-keeping. There are mountains of research that result in nothing interesting whatsoever (aside from learning about what doesn't work). And all of that is still valuable knowledge!

ezst 3 hours ago | parent [-]

Sure, but now we can't even assume that such research is submitted in good faith anymore. There just seems to be no perfect solution.

Maybe something like a "hierarchy/DAG? of trusted-peers", where groups like universities certify the relevance and correctness of papers by attaching their name and a global reputation score to it. When it's found that the paper is "undesirable" and doesn't pass a subsequent review, their reputation score deteriorates (with the penalty propagating along the whole review chain), in such a way that:

- the overall review model is distributed, hence scalable (everybody may play the certification game and build a reputation score while doing so) - trusted/established institutions have an incentive to keep their global reputation score high and either put a very high level of scrutiny to the review, or delegate to very reputable peers - "bad actors" are immediately punished and universally recognized as such - "bad groups" (such as departments consistently spamming with low quality research) become clearly identified as such within the greater organisation (the university), which can encourage a mindset of quality above quantity - "good actors within a bad group" are not penalised either because they could circumvent their "bad group" on the global review market by having reputable institutions (or intermediaries) certify their good work

There are loopholes to consider, like a black market of reputation trading (I'll pay you generously to sacrifice a bit of your reputation to get this bad science published), but even that cannot pay off long-term in an open system where all transactions are visible.

Incidentally, I think this may be a rare case where a blockchain makes some sense?

jll29 2 hours ago | parent | next [-]

You have some good ideas there, it's all about incentives and about public reputation.

But it should also fair. I once caught a team at a small Indian branch of a very large three letter US corporation violating the "no double submission" rule of two conferences: they submitted the same paper to two conferences, both naturally landed in my reviewer inbox, for a topic I am one of the experts in.

But all the other employees should not be penalized by the violations of 3 researchers.

gus_massa 3 hours ago | parent | prev | next [-]

This idea looks very similar to journals! Each journal has a reputation, if they publish too much crap, the crap is not cited and the impact factors decrease. Also, they have an informal reputation, because impact index also has problems.

Anyway, how will universities check the papers? Somone must read the preprints, like the current reviewers. Someone must check the incoming preprints, find reviewers and make the final decition, like the current editors. ...

amitav1 2 hours ago | parent | prev [-]

How would this work for independent researchers?

(no snark)

mathematicaster 3 hours ago | parent | prev | next [-]

Pay to review is common in Econ and Finance.

skissane 2 hours ago | parent [-]

Variation I thought of on pay-to-review:

Suppose you are an independent researcher writing a paper. Before submitting it for review to journals, you could hire a published author in that field to review it for you (independently of the journal), and tell you whether it is submission-worthy, and help you improve it to the point it was. If they wanted, they could be listed as coauthor, and if they don't want that, at least you'd acknowledge their assistance in the paper.

Because I think there are two types of people who might write AI slop papers: (1) people who just don't care and want to throw everything at the wall and see what sticks; (2) people who genuinely desire to seriously contribute to the field, but don't know what they are doing. Hiring an advisor could help the second group of people.

Of course, I don't know how willing people would be to be hired to do this. Someone who was senior in the field might be too busy, might cost too much, or might worry about damage to their own reputation. But there are so many unemployed and underemployed academics out there...

utilize1808 4 hours ago | parent | prev [-]

Better yet, make a "polymarket" for papers where people can bet on which paper can make it, and rely on "expertise arbitrage" to punish spams.

ezst 3 hours ago | parent [-]

Doesn't stop the flood, i.e. the unfair asymmetry between the effort to produce vs. effort to review.

utilize1808 36 minutes ago | parent [-]

Not if submissions require some small mandatory bet.

pickleRick243 an hour ago | parent | prev | next [-]

I'm curious if you'd be in favor of other forms of academic gate keeping as well. Isn't the lower quality overall of submissions (an ongoing trend with a history far pre-dating LLMs) an issue? Isn't the real question (that you are alluding to) whether there should be limits to the democratization of science? If my tone seems acerbic, it is only because I sense cognitive dissonance between the anti-AI stance common among many academics and the purported support for inclusivity measures.

"which is really not the point of these journals at all"- it seems that it very much is one of the main points? Why do you think people publish in journals instead of just putting their work on the arxiv? Do you think postdocs and APs are suffering through depression and stressing out about their publications because they're agonizing over whether their research has genuinely contributed substantively to the academic literature? Are academic employers poring over the publishing record of their researchers and obsessing over how well they publish in top journals in an altruistic effort to ensure that the research of their employees has made the world a better place?

agnishom 7 minutes ago | parent [-]

> whether there should be limits to the democratization of science?

That is an interesting philosophical question, but not the question we are confronted with. A lot of LLM assisted materials have the _signals_ of novel research without having its _substance_.

Rperry2174 4 hours ago | parent | prev | next [-]

This keeps repeating in different domains: we lower the cost of producing artifacts and the real bottleneck is evaluating them.

For developers, academics, editors, etc... in any review driven system the scarcity is around good human judgement not text volume. Ai doesn't remove that constraint and arguably puts more of a spotlight on the ability to separate the shit from the quality.

Unless review itself becomes cheaper or better, this just shifts work further downstream and disguising the change as "efficiency"

SchemaLoad 3 hours ago | parent | next [-]

This has been discussed previously as "workslop", where you produce something that looks at surface level like high quality work, but just shifts the burden to the receiver of the workslop to review and fix.

vitalnodo 4 hours ago | parent | prev [-]

This fits into the broader evolution of the visualization market. As data grows, visualization becomes as important as processing. This applies not only to applications, but also to relating texts through ideas close to transclusion in Ted Nelson’s Xanadu. [0]

In education, understanding is often best demonstrated not by restating text, but by presenting the same data in another representation and establishing the right analogies and isomorphisms, as in Explorable Explanations. [1]

[0] https://news.ycombinator.com/item?id=40295661

[1] https://news.ycombinator.com/item?id=22368323

jasonfarnon 40 minutes ago | parent | prev | next [-]

I'm certain your journal will be using LLMs in reviewing incoming articles, if they aren't already. I also don't think this is in response to the flood of LLM generated articles. Even if authors were the same as pre-LLM, journals would succumb to the temptation, at least at the big 5 publishers, which already have a contentious relationship with the referees.

jjcm 3 hours ago | parent | prev | next [-]

The comparison to make here is that a journal submission is effectively a pull request to humanities scientific knowlegde base. That PR has to be reviewed. We're already seeing the effects of this with open source code - the number of PR submissions have skyrocketed, overwhelming maintainers.

This is still a good step in a direction of AI assisted research, but as you said, for the moment it creates as many problems as it solves.

mrandish 4 hours ago | parent | prev | next [-]

As a non-scientist (but long-time science fan and user), I feel your pain with what appears to be a layered, intractable problem.

> > who are looking to 'boost' their CV

Ultimately, this seems like a key root cause - misaligned incentives across a multi-party ecosystem. And as always, incentives tend to be deeply embedded and highly resistant to change.

maxkfranz 4 hours ago | parent | prev | next [-]

I generally agree.

On the other hand, the world is now a different place as compared to when several prominent journals were founded (1869-1880 for Nature, Science, Elsevier). The tacit assumptions upon which they were founded might no longer hold in the future. The world is going to continue to change, and the publication process as it stands might need to adapt for it to be sustainable.

ezst 4 hours ago | parent [-]

As I understand it, the problem isn't publication or how it's changing over time, it's about the challenges of producing new science when the existing one is muddied in plausible lies. That warrants a new process by which to assess the inherent quality of a paper, but even if it comes as globally distributed, the cheats have a huge advantage considering the asymmetry between the effort to vibe produce vs. the tedious human review.

maxkfranz 3 hours ago | parent [-]

That’s a good point. On the other hand, we’ve had that problem long before AI. You already need to mentally filter papers based on your assessment of the reputability of the authors.

The whole process should be made more transparent and open from the start, rather than adding more gatekeeping. There ought to be openness and transparency throughout the entire research process, with auditing-ability automatically baked in, rather than just at the time of publication. One man’s opinion, anyway.

boplicity 4 hours ago | parent | prev | next [-]

Is it at all possible to have a policy that bans the submission of any AI written text, or text that was written with the assistance of AI tools? I understand that this would, by necessity, be under an "honor system" but maybe it could help weed out papers not worth the time?

currymj 2 hours ago | parent [-]

this is probably a net negative as there are many very good scientists with not very strong English skills.

the early years of LLMs (when they were good enough to correct grammar but not enough to generate entire slop papers) were an equalizer. we may end up here but it would be unfortunate.

2 hours ago | parent | prev | next [-]
[deleted]
jascha_eng 3 hours ago | parent | prev | next [-]

Why not filter out papers from people without credentials? And also publicly call them out and register them somewhere, so that their submission rights can be revoked by other journals and conferences after "vibe writing".

These acts just must have consequences so people stop doing them. You can use AI if you are doing it well but if you are wasting everyones time you should just be excluded from the discourse altogether.

keithnz 2 hours ago | parent | prev | next [-]

wouldn't AI actually be good for filtering given it's going to be a lot better at knowing what has been published? Also seems possible that it could actually work out papers that have ideas that are novel, or at least come up with some kind of likely score.

usefulposter 5 hours ago | parent | prev | next [-]

Completely agree. Look at the independent research that gets submitted under "Show HN" nowadays:

https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru...

https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru...

lupsasca 5 hours ago | parent | prev | next [-]

I am very sympathetic to your point of view, but let me offer another perspective. First off, you can already vibe-write slop papers with AI, even in LaTeX format--tools like Prism are not needed for that. On the other hand, it can really help researchers improve the quality of their papers. I'm someone who collaborates with many students and postdocs. My time is limited and I spend a lot of it on LaTeX drudgery that can and should be automated away, so I'm excited for Prism to save time on writing, proofreading, making TikZ diagrams, grabbing references, etc.

CJefferson 3 hours ago | parent | next [-]

What the heck is the point of a reference you never read?

lupsasca 3 hours ago | parent [-]

By "grabbing references" I meant queries of the type "add paper [bla] to the bibliography" -- that seems useful to me!

nestes 2 hours ago | parent [-]

Focusing in on "grabbing references", it's as easy as drag-and-drop if you use Zotero. It can copy/paste references in BibTeX format. You can even customize it through the BetterBibTeX extension.

If you're not a Zotero user, I can't recommend it enough.

noitpmeder 4 hours ago | parent | prev [-]

AI generating references seems like a hop away from absolute unverifiable trash.

SecretDreams 4 hours ago | parent | prev [-]

I appreciate and sympathize with this take. I'll just note that, in general, journal publications have gone considerably downhill over the last decade, even before the advent of AI. Frequency has gone up, quality has gone down, and the ability to actually check if everything in the article is actually valid is quite challenging as frequency goes up.

This is a space that probably needs substantial reform, much like grad school models in general (IMO).