| ▲ | atleastoptimal 5 days ago |
| This is all true. The best way to treat LLM's as they are now is one step above the abstraction offered by compiled languages over assembly. You can describe something in plain english, note its explicit requirements, inputs and outputs, and an LLM can effectively write the code as a translation of the logic you are specifying. Using LLM's, you are best served minimizing the entropy they have to deal with. The transformer is essentially a translation engine, so use it as a translator, not as a generator. That being said, every few months a new model comes out that is a little less encumbered by the typical flaws of LLM's, a little more "intuitively" smart and less needing of hand-holding, a little more reliable. I feel that this is simply a natural course of evolution, as more money is put into LLM's they get better because they're essentially a giant association machine, and those associations give rise to larger abstractions, more robust conceptions of how to wield the tools of understanding the world, etc. Over time it seems inevitable that providing an LLM any task it will be able to perform that task better than any human programmer given it, and the same will go for the rest of what humans do. |
|
| ▲ | nickm12 5 days ago | parent | next [-] |
| This is a false analogy. LLMs do not "compile" natural language to high level code in the same way that a compiler or interpreter implements a high-level programming language in terms of machine instructions (or, for that matter, how a CPU implements machine instructions in hardware). Programming and machine languages aim for a precise and unambiguous semantics, such that it's meaningful to talk about things like whether the semantics are actually precise or whether the compiler has a bug in failing to implement the spec. Natural language is not just a higher level of abstraction on our existing stack. If a new model comes out, or you even run an existing model with a new seed, you can get different code out that behaves differently. This is not how compilers work. |
| |
| ▲ | atleastoptimal 5 days ago | parent | next [-] | | If a program calls an API like search_engine.get_search_results(query, length, order) It doesn't "care" about the algorithm that produced that list of results, only that it fits the approximation of how the algorithm works as defined by the schema. There are thousands of ways the engine could have been implemented to produce the schema that returns relevance-based results from a web-crawler-sourced database. In the same way, if I prompt an LLM "design a schema with [list of requirements] that works in [code context and API calls]", there are thousands of ways it could produce that code, but within a margin of error a high quality LLM should be able to produce the code that fits those requirements. Of course the difference is that there is a stochastic element to LLM generated code. However it is useful to think of LLM's this way because it allows being able to leverage their probability of being correct, even if they aren't as precise as calling APIs but being explicit in how those abstractions are used. | |
| ▲ | devnullbrain 5 days ago | parent | prev [-] | | This is a false interpretation, you've put "compile" in quotes when it doesn't appear in the parent comment and the actual phrasing used is more correct. | | |
| ▲ | nickm12 12 hours ago | parent [-] | | The parent comment was talking about the "abstraction offered by compiled languages over assembly". I quote an alternate form of "compiled" from that sentence. |
|
|
|
| ▲ | voxl 5 days ago | parent | prev [-] |
| No, LLMs are not an "abstraction" like a compiler is. This is bullshit. LLMs are stochastic token generators. I have NEVER met someone in real life that has produced something I wouldn't throw in the trash using LLMs, and I have had the displeasure of eating cookies baked from an LLM recipe. No, LLMs will not get better. The singularity bullshit has been active since 2010s. LLMs have consumed the entire fucking Internet and are still useless. Where the fuck is the rest of the data going to come from? All these emails from people wanting high quality data from PhDs only for them to be scammy. People only want to train these things on easily stolen garbage, not quality input, because quality is expensive. Go figure! This optimistic horeshit hype is embarrassing. |
| |
| ▲ | atleastoptimal 5 days ago | parent | next [-] | | >No, LLMs will not get better. What makes you so sure of this? They've been getting better like clockwork every few months for the past 5 years. | | |
| ▲ | bigstrat2003 5 days ago | parent | next [-] | | I don't claim that they won't get better, but they certainly haven't gotten better. From the original release of ChatGPT to now, they still suck in the same exact ways. | | |
| ▲ | johnisgood 5 days ago | parent [-] | | I don't think they have gotten better either (at least in the past 1 year), because I remember how much better ChatGPT or even Claude used to be before. Perhaps they are nerfed now for commercial use, who knows. |
| |
| ▲ | otabdeveloper4 5 days ago | parent | prev [-] | | No they haven't. The hallucinate exactly as much as they did five years ago. | | |
| ▲ | atleastoptimal 5 days ago | parent | next [-] | | Absolutely untrue. Claiming GPT-3 hallucinates as much as o3 over the same token horizon on the same prompts is a silly notion and easily disproven by the dozens of benchmarks. You can code a complete web-app with models now, something far beyond the means of models so long ago. | | |
| ▲ | otabdeveloper4 5 days ago | parent [-] | | > caveats and weasel words > "benchmarks" Stop drinking the coolaid and making excuses for LLM limitations, and learn to use the tools properly given their limits instead. |
| |
| ▲ | antihero 5 days ago | parent | prev [-] | | They really don’t though. | | |
| ▲ | otabdeveloper4 5 days ago | parent [-] | | Larger context lengths are awesome, but they don't fundamentally change the failure modes of LLMs. |
|
|
| |
| ▲ | anshumankmr 5 days ago | parent | prev [-] | | > LLMs have consumed the entire fucking Internet and are still useless. They aren't useless. Otherwise, ChatGPT would have died a long time back > Where the fuck is the rest of the data going to come from? Good question. Personally, I think companies will start paying more for high quality data or what is at least perceived as high quality data.
I think Reddit and some other social media companies like it are poised to reap the rewards of this. Whether this will be effective in the long run remains to be seen. | | |
| ▲ | misnome 5 days ago | parent [-] | | > They aren't useless. Otherwise, ChatGPT would have died a long time back Isn’t the entire industry being fuelled by orders of magnitude more VC funding than revenue? | | |
| ▲ | anshumankmr 5 days ago | parent | next [-] | | >Isn’t the entire industry being fuelled by orders of magnitude more VC funding than revenue? Because people want to use it, right? And it is a matter of time before they start limiting the ChatGPT "free" or "logged out" accounts, I feel. In the consumer AI chat apps, it is still the dominant brand, at least in my anecdotal experience, and they will basically make the Plus version the one version of the app to definitely use. Plus they are planning on selling it to enterprises, and at least a couple of them are signing up for sure. | | |
| ▲ | johnisgood 5 days ago | parent [-] | | I think they are already limiting / nerfing "free" vs "logged out" vs "paid" vs "non-commercial". |
| |
| ▲ | arthens 5 days ago | parent | prev [-] | | Isn't that an argument against the sustainability of the LLM business model rather than their usefulness? People use them because they are useful, not because they are VC funded. | | |
| ▲ | skydhash 5 days ago | parent [-] | | When the product is free, that put the barrier at ground level. I have more confidence in Kagi userbase, than OpenAI’s. |
|
|
|
|