| |
| ▲ | tikhonj 7 days ago | parent | next [-] | | Can't speak for anyone else, but for me, AI/LLMs have been firmly in the "nice but forgettable" camp. Like, sometimes it's marginally more convenient to use an LLM than to do a proper web search or to figure out how to write some code—but that's a small time saving at best, it's less of a net impact than Stack Overflow was. I'm already a pretty fast writer and programmer without LLMs. If I hadn't already learned how to write and program quickly, perhaps I would get more use out of LLMs. But the LLMs would be saving me the effort of learning which, ultimately, is an O(1) cost for O(n) benefit. Not super compelling. And what would I even do with a larger volume of text output? I already write more than most folks are willing to read... So, sure, it's not strictly zero utility, but it's far less utility than a long series of other things. On the other hand, trains are fucking amazing. I don't drive, and having real passenger rail is a big chunk of why I want to move to Europe one day. Being able to get places without needing to learn and then operate a big, dangerous machine—one that is statistically much more dangerous for folks with ADHD like me—makes a massive difference in my day-to-day life. Having a language model... doesn't. And that's living in the Bay Area where the trains aren't great. Bart, Caltrain and Amtrak disappearing would have an orders of magnitude larger effect on my life than if LLMs stopped working. And I'm totally ignoring the indirect but substantial value I get out of freight rail. Sure, ships and trucks could probably get us there, but the net increase in costs and pollution should not be underestimated. | | |
| ▲ | knowitnone2 7 days ago | parent [-] | | No matter how good or fast you are, you will never beat the LLM. What you're saying is akin to "your math is faster than a calculator" and I'm willing to bet it's not. LLMs are not perfect and will require intervention and fixing but if it can get you 90% there, that's pretty good. In the coming years, you'll soon find your peers are performing much faster than you (assuming you program for a living) and you will have no choice but you do you. | | |
| ▲ | tikhonj 7 days ago | parent | next [-] | | Fun story: when I interned at Jane Street, they gave out worksheets full of put-call parity calculations to do in your head because, when you're trading, being able to do that sort of calculation at a glance is far faster and more fluid than using a calculator or computer. So for some professionals, mental math really is faster. Make of that what you will. | |
| ▲ | globular-toast 7 days ago | parent | prev | next [-] | | Beat an LLM at what? Lines of code per minute? Certainly not. But that's not my job. If anything I try to minimise the amount of code I output. On a good day my line count will be negative. Mathematicians are not calculators. Programmers are not typists. | |
| ▲ | WD-42 7 days ago | parent | prev | next [-] | | LLMs do not work the same way calculators do, not even close. | |
| ▲ | haganomy 7 days ago | parent | prev [-] | | So now programmers add value when they write more code faster? Curious how this was anathema but now is a clear evidence of LLM-driven coding superiority. The math that isn't mathing is even more basic tho. This is a Concorde situation all over again. Yes, supersonic passenger jets would be amazing. And they did reach production. But the economics were not there. Yeah, using GPU farms delivers some conveniences that are real. But after 1.6 trillion dollars it's not clear at all that they are a net gain. |
|
| |
| ▲ | bluefirebrand 7 days ago | parent | prev | next [-] | | > Can you explain how you don't find clear utility, at the personal level, from LLMs? Sure. They don't meaningfully improve anything in my life personally. They don't improve my search experience, they don't improve my work experience, they don't improve the quality of my online interactions, and I don't think they improve the quality of the society I live in either | | |
| ▲ | knowitnone2 7 days ago | parent | next [-] | | so you never read the summary at the top of Google search results to get the answer because it provides the answer to most of my searches. "they don't improve my work experience" that's fair but perhaps you haven't really given it a try? "they don't improve the quality of my online interactions" but how do you know? LLMs are being used to create websites, generate logos, images, memes, art videos, stories - you've already been entertained by them and not even know it. "I don't think they improve the quality of the society I live in either" That's a feeling, not a fact. | | |
| ▲ | AngryData 7 days ago | parent | next [-] | | I never do because I still don't trust its answers enough without also seeing a secondary source to confirm it and the first result or two is already correct 99% of the time and often also has source citations to tell me how that conclusion or information was made or gathered if im dealing with a potential edge case. | |
| ▲ | GJim 6 days ago | parent | prev | next [-] | | > so you never read the summary at the top of Google search results to get the answer No. Because I cannot trust it. (Especially when it gives no attributions). | |
| ▲ | whatarethembits 6 days ago | parent | prev | next [-] | | I generally recognise utility of AI, but on this particular point, it has been a net negative if I were to add up the time I wasted by believing a summarised answer, got some length further on given task only to find that the answer was wrong and having to backtrack and redo all that work. | |
| ▲ | bluefirebrand 7 days ago | parent | prev | next [-] | | > so you never read the summary at the top of Google search results to get the answer because it provides the answer to most of my searches Unfortunately yes I do, because it is placed in a way to immediately hijack my attention Most of the time it is just regurgitating the text of the first link anyways, so I don't think it saves a substantial amount of time or effort. I would genuinely turn it off if they let me > That's a feeling, not a fact So? I'm allowed to navigate my life by how I feel | | |
| ▲ | ryao 7 days ago | parent [-] | | If you find it annoying, why not configure a custom blocking rule in an adblocker to remove it? | | |
| |
| ▲ | incone123 7 days ago | parent | prev [-] | | I've read some complete nonsense in those summaries. I use LLMs for other things but I don't find this application useful because I would need to trust it, and I don't. |
| |
| ▲ | rockemsockem 7 days ago | parent | prev [-] | | Have you even tried using them though? Like in earnest? Or do you see yourself as a conscientious objector of sorts? | | |
| ▲ | ruszki 7 days ago | parent | next [-] | | This whole topic makes me remember the argument for vi, and quick typing. I was always baffled because for the 25 years since I can code, typing was never that huge block of my time that it would matter. I have the same feeling with AI. It clearly cannot produce the quality of code, architecture, features which I require from myself. And I also want to understand what’s written, and not saying “it works, it’s fine <inserting dog with coffee image here>”, and not copy-pasting a terrible StackOverflow answer which doesn’t need half of the code in reality, and clearly nobody who answered sat down and tried to understand it. Of course, not everybody wants these, and I’ve seen several people who were fine with not understanding what they were doing. Even before AI. Now they are happy AI users. But it clears to me that it’s not beneficial salary, promotion, and political power wise. So what’s left is that it types faster… but that was never an issue. It can be better however. There was the first case just about a month ago when one of them could answer better to a problem than anything else which I knew or could find via Kagi/Google. But generally speaking it’s not there at all. Yet. | |
| ▲ | bluefirebrand 7 days ago | parent | prev [-] | | I have tried using them frequently. I've tried many things for years now, and while I am impressed I'm not impressed enough to replace any substantial part of my workflow with them At this point I am somewhat of a conscientious objector though Mostly from a stance of "these are not actually as good as people say and we will regret automating away jobs held by competent people in favor of these low quality automations" |
|
| |
| ▲ | agent_turtle 7 days ago | parent | prev | next [-] | | There was a study recently that showed how not only did devs overestimate the time saved using AI, but that they were net negative compared to the control group. Anyway, that about sums up my experience with AI. It may save some time here and there, but on net, you’re better off without it. | | |
| ▲ | keeda 7 days ago | parent | next [-] | | That study gets mentioned all the time, somehow this one and many of the others it cites don't get much airtime: https://www.stlouisfed.org/on-the-economy/2024/sep/rapid-ado... >This implies that each hour spent using genAI increases the worker’s productivity for that hour by 33%. This is similar in magnitude to the average productivity gain of 27% from several randomized experiments of genAI usage (Cui et al., 2024; Dell’Acqua et al., 2023; Noy and Zhang, 2023; Peng et al., 2023) Our estimated aggregate productivity gain from genAI (1.1%) exceeds the 0.7% estimate by Acemoglu (2024) based on a similar framework. To be clear, they are surmising that GenAI is already having a productivity gain. | | |
| ▲ | agent_turtle 7 days ago | parent [-] | | The article you gave is derived from a poll, not a study. As for the quote, I can’t find it in the article. Can you point me to it? I did click on one of the studies and it indicated productivity gains specifically on writing tasks. Which reminded me of this recent BBC article about a copywriter making bank fixing expensive mistakes caused by AI: https://www.bbc.com/news/articles/cyvm1dyp9v2o | | |
| ▲ | keeda 7 days ago | parent [-] | | The quote is from the paper linked in the article: https://s3.amazonaws.com/real.stlouisfed.org/wp/2024/2024-02... It's actually based on the results of three surveys conducted by two different parties. While surveys are subject to all kinds of biases and the gains are self-reported, their findings of 25% - 33% producitivity do match the gains shown by at least 3 other randomized studies, one of which was specifically about programming. Those studies are worth looking at as well. | | |
| ▲ | foolswisdom 7 days ago | parent [-] | | It's worth noting that the METR paper that found decreased productivity also found that many of the developers thought the work was being sped up. | | |
| ▲ | keeda 6 days ago | parent [-] | | Yes, self-reporting has biases and estimating tasks is still a fool's errand, which is why I noted that the estimates from these surveys matched the findings from other RTC studies. However, what doesn't get discussed enough about the METR study is that there was a spike in overall idle time as they waited for the AI to finish. I haven't run the numbers so I don't know how much of the increased completion time it accounts for, but if your cognitive load drops almost to 0, it will of course feel like your work is sped up, even though calendar time has increased. I wonder if that is the more important finding of that paper. |
|
|
|
| |
| ▲ | rockemsockem 7 days ago | parent | prev | next [-] | | I'm not talking about time saving. AI seems to speed up my searching a bit since I can get results quicker without having to find the right query then find a site that actually answers my question, but that's minor, as nice as it is. I use AI in my personal life to learn about things I never would have without it because it makes the cost of finding any basic knowledge basically 0. Diet improvement ideas based on several quick questions about gut functioning, etc, recently learning how to gauge tsunami severity, and tons of other things. Once you have several fundamental terms and phrases for new topics it's easy to then validate the information with some quick googling too. How much have you actually tried using LLMs and did you just use normal chat or some big grand complex tool? I mostly just use chat and prefer to enter my code in artisanally. | | |
| ▲ | lisbbb 7 days ago | parent | next [-] | | How much of that is junk knowledge, though? I mean, sure, I love looking up obscure information, particularly about cosmology and astronomy, but in reality, it's not making me better or smarter, it's just kind of "science junk food." It feels good, though. I feel smarter. I don't think I am, though, because the things I really need to work on about myself are getting pushed aside. | | |
| ▲ | rockemsockem 6 days ago | parent [-] | | For me it's pretty much all knowledge that I'm immediately operationalizing. I occasionally use it to look up actors and stuff too, but most of the time it's information that provides direct value to me |
| |
| ▲ | flkiwi 7 days ago | parent | prev | next [-] | | This is kind of how I use it: 1. To work through a question I'm not sure how to ask yet
2. To give me a starting point/framework when I have zero experience with an issue
3. To automate incredibly stupid monkey-level tasks that I have to do but are not particularly valuable It's a remarkable accomplishment that has the potential to change a lot of things very quickly but, right now, it's (by which I mean publicly available models) only revolutionary for people who (a) have a vested interest in its success, (b) are easily swayed by salespeople, (c) have quite simple needs (which, incidentally, can relate to incredible work!), or (d) never really bothered to check their work anyway. | |
| ▲ | fzeroracer 7 days ago | parent | prev | next [-] | | Why not just look up the information directly instead of asking a machine that you can never truly validate? If I need information, I can just keyword search wikipedia, then follow the chain there and then validate the sources along with outside information. An LLM would actually cost me time because I would still need to do all of the above anyways, making it a meaningless step. If you don't do the above then it's 'cheaper' but you're implicitly trusting the lying machine to not lie to you. | | |
| ▲ | rockemsockem 7 days ago | parent | next [-] | | See my previous statement > Once you have several fundamental terms and phrases for new topics it's easy to then validate the information with some quick googling too. You're practically saying that looking at an index in the back of a book is a meaningless step. It is significantly faster, so much so that I am able to ask it things that would have taken an indeterminate amount of time to research before, for just simple information, not deep understanding. Edit: Also I can truly validate literally any piece of information it gives me. Like I said previously, it makes it very easy to validate via Wikipedia or other places with the right terms, which I may not have known ahead of time. | | |
| ▲ | fzeroracer 7 days ago | parent [-] | | Again, why would you just not use Wikipedia as your index? I'm saying why would you use the index that lies and hallucinates to you instead of another perfectly good index elsewhere. You're using the machine that ingests and regurgitates stuff like Wikipedia to you. Why not skip the middleman entirely? | | |
| ▲ | rockemsockem 7 days ago | parent [-] | | Because the middleman is faster and practically never lies/hallucinates for simple queries, the middleman can handle vague queries that Google and Wikipedia cannot. The same reasons you use Wikipedia instead of reading all the citations on Wikipedia. | | |
| ▲ | fzeroracer 7 days ago | parent [-] | | > Because the middleman is faster and practically never lies/hallucinates for simple queries How do you KNOW it doesn't lie/hallucinate? In order to know that, you have to verify what it says. And in order to verify what it says, you need to check other outside sources, like Wikipedia. So what I'm saying is: Why bother wasting time with the middle man? 'Vague queries' can be distilled into simple keyword searches: If I want to know what a 'Tsunami' is I can simply just plug that keyword into a Wikipedia search and skim through the page or ctrl-f for the information I want instantly. If you assume that it doesn't lie/hallucinate because it was right on previous requests then you fall into the exact trap that blows your foot off eventually, because sometimes it can and will hallucinate over even benign things. | | |
| ▲ | rockemsockem 6 days ago | parent [-] | | I feel like you're coming from a very strange place of both using advanced technology that saves you time and expands your personal knowledge base and at the same time saying that the more advanced technology that saves you even more time and expands your knowledge base further is useless and a time sink. For most questions it is so much faster to validate a correct answer than to figure out the answer to begin with. Vague queries CANNOT be distilled to simple keyword searches when you don't know where to start without significant time investment. Ctrl-f relies on you and the article having the exact same preferred vocabulary for the exact same concepts. I do not assume that LLMs don't lie or hallucinate, I start with the assumption that they will be wrong. Which for the record is the same assumption I take with both websites and human beings. |
|
|
|
| |
| ▲ | lisbbb 7 days ago | parent | prev [-] | | A lot of formerly useful search tools, particularly Google, are just trash now, absolute trash. |
| |
| ▲ | agent_turtle 7 days ago | parent | prev [-] | | OpenAI is currently being evaluated in terms of hundreds of billions. That’s an insane number for creating a product that “speeds up searching a bit”. | | |
| ▲ | rockemsockem 6 days ago | parent [-] | | Did you miss the part where I said it helps me find new knowledge that I wouldn't have otherwise? That is pretty significant in my book. | | |
| ▲ | agent_turtle 4 days ago | parent [-] | | I don't know how you use search but I often find incredible information that I didn't explicitly search for. How do you quantify such things? How can you say with a straight face that this magic box gives you more relevant information (which may be wrong!) and that will revolutionize the workforce? | | |
| ▲ | rockemsockem 4 days ago | parent [-] | | I am searching for the incredible information and I can't find it without the LLM because I don't know the proper terminology ahead of time and search isn't that good unless you know exactly what you want. |
|
|
|
| |
| ▲ | ares623 7 days ago | parent | prev [-] | | Is the "here and there" tasks that were previously so little value that they are always stuck in the backlog? i.e. the parts where it helps have very little value in the first place. |
| |
| ▲ | decimalenough 7 days ago | parent | prev | next [-] | | I actually do get clear utility, with major caveats, namely that I only ask things where the answer is both well known and verifiable. I still do 10-20x regular Kagi searches for every LLM search, which seems about right in terms of the utility I'm personally getting out of this. | |
| ▲ | roncesvalles 7 days ago | parent | prev | next [-] | | As a dev, I find that the personal utility of LLMs is still very limited. Analyze it this way: Are LLMs enabling something that was impossible before? My answer would be No. Whatever I'm asking of the LLM, I'd have figured it out from googling and RTFMing anyway, and probably have done a better job at it. And guess what, after letting the LLM do it, I probably still need to google and RTFM anyway. You might say "it's enabling the impossible because you can now do things in less time", to which I would say, I don't really think you can do it in less time. It's more like cruise control where it takes the same time to get to your destination but you just need to expend less mental effort. Other elephants in the room: - where is the missing explosion of (non-AI) software startups that should've been enabled by LLM dev efficiency improvements? - why is adoption among big tech SWEs near zero despite intense push from management? You'd think, of all people, you wouldn't have to ask them twice. The emperor has no clothes. | | |
| ▲ | ryao 7 days ago | parent | next [-] | | > Are LLMs enabling something that was impossible before? I would say yes when the LLM is combined with function calling to allow it to do web searches and read web pages. It was previously impossible for me to research a subject within 5 minutes when it required doing several searches and reviewing dozens of search results (not just reading the list entries, but reading the actual HTML pages). I simply cannot read that fast. A LLM with function calling can do this. The other day, I asked it to check the Linux kernel sources to tell me which TCP connection states for a closing connection would not return an error to send() with MSG_NOSIGNAL. It not only gave me the answer, but made citations that I could use to verify the answer. This happened in less than 2 minutes. Very few developers could find the answer that fast, unless they happen to already know it. I doubt very many know it offhand. Beyond that, I am better informed than I have ever been since I have been offloading previously manual research to LLMs to do for me, allowing me to ask questions that I previously would not ask due to the amount of time it took to do the background research. What previously would be a rabbit hole that took hours can be done in minutes with minimal mental effort on my part. Note that I am careful to ask for citations so I can verify what the LLM says. Most of the time, the citations vouch for what the LLM said, but there are some instances where the LLM will provide citations that do not. | |
| ▲ | knowitnone2 7 days ago | parent | prev [-] | | do cars enable something that was impossible before? bikes? shoes? clothing? Your answer would be No. | | |
| ▲ | danlitt 7 days ago | parent | next [-] | | Yes, obviously. Commuting between cities would be an example. | | | |
| ▲ | roncesvalles 7 days ago | parent | prev [-] | | If your implication is that LLM-assisted coding to non-LLM-assisted coding is like motorcar to horse buggy, that is just not the case. | | |
| ▲ | ryao 7 days ago | parent [-] | | I think he was referring to the ability to go from A to B within a certain amount of time. There is a threshold at which it is possible for a car, yet impossible for a horse and buggy. That said, I recently saw a colleague use a LLM to make a non-trivial UI for electron in HTML/CSS/JS, despite knowing nothing about any of those technologies, in less time than it would have taken me to do it. We had been in the process of devising a set of requirements, he fed his version of them into the LLM, did some back and forth with the LLM, showed me the result, got feedback, fed my feedback back into the LLM and got a good solution. I had suggested that he make a mockup (a drawing in kolourpaint for example) for further discussion, but he had surprised me by using a LLM to make a functional prototype in place of the mockup. It was a huge time saver. | | |
| ▲ | roncesvalles 7 days ago | parent | next [-] | | The issue is that the 'B' is not very consequential. Consider something like Shopify - someone with zero knowledge of programming can wow you with an incredible ecommerce site built through Shopify. It's probably like a 1000x efficiency improvement versus building one from scratch (or even using the popular lowcode tools of the era like Magento and Drupal). But it won't help you build Amazon.com, or even Nike.com. It won't even get you part of the way there. And LLMs, while more general/expressive than Shopify, are inferior to Shopify at doing what Shopify does i.e. you're still better off using Shopify instead of trying to vibe-code an e-commerce website. I would say the same line of thinking extends to general software engineering. | | |
| ▲ | ryao 6 days ago | parent | next [-] | | What was described was offloading portions of software development to a LLM to reach B faster. This works very well and is an improvement over the traditional method of implementing everything yourself. Shopify is tangential to this. I will add that having had experience with similar platforms in the past (for building websites, not e-commerce), I can say that you must be either naive or a masochist to use them. They tend to be mediocre compared to what you can get from self hosted solutions and the vendor lock-in always will be used to bite those foolish enough to use them in the end. | |
| ▲ | satyrun 7 days ago | parent | prev [-] | | Or you are just not creative at all and not making anything that interesting yourself. |
| |
| ▲ | 7 days ago | parent | prev [-] | | [deleted] |
|
|
|
| |
| ▲ | ehnto 7 days ago | parent | prev | next [-] | | Much work is not "text in, text out", and much work is not digital at all. But even in the digital world, inserting an LLM into a workflow is just not always that useful. In fact much automation, code or otherwise, benefits from or even requires explicit, concise rules. It is far quicker for me to already know, and write, an SQL statement, than it is to explain what I need to an LLM. It is also quite difficult to get LLMs into a lot of processes, and I think big enterprises are going to really struggle with this. I would absolutely love AI to manage some Windows servers that are in my care, but they are three VMs deep in a remote desktop stack that gets me into a DMZ/intranet. There's no interface, and how would an LLM help anyway. What I need is concise, discreet automations. Not a chat bot interface to try and instruct every day. To be clear I do try to use AI most days, I have Claude and I am a software developer so ideally it could be very helpful, but I have far less use for it than say people in the strategy or marketing departments for example. I do a lot of things, but not really all that much writing. | |
| ▲ | kazinator 7 days ago | parent | prev | next [-] | | Gemini wasted my time today assuring me that if I want a git bundle that only has the top N commits, yet is cleanly clone-able, I can just make a --depth N clone of the original repo, and and do a git bundle create ... --all. Nope; cloning a bundle created from a depth-limited clone results in error messages about missing commit objects. So I tell the parrot that, and it comes back with: of course, it is well-known that it doesn't work, blah blah. (Then why wasn't it well known one prompt ago, when it was suggested as the definitive answer?) Obviously, I wasn't in the "the right mindset" today. This mindset is one of two things: - the mindset of a complete n00b asking a n00b question that it will nail every time, predicting it out of its training data richly replete with n00b material. - the mindset of a patient data miner, willing to expend all they keystrokes. needed to build up enough context to in effect create a query which zeroes in on the right nugget of information that made an appearance in the training data. It was interesting to go down this #2 rabbit hole when this stuff was new, which it isn't any more. Basically do most of the work, while it looks as if it solved the problem. I had the right mindset for AI, but most of it has worn off. If I don't get something useful in one query with at most one follow up, I quit. The only shills who continue to hype AI are either completely dishonest assholes, or genuine bros bearing weapons-grade confirmation bias. Let's try something else: Q: "What modes of C major are their own reflection?" A: "The Lydian and Phrygian modes are reflections of each other, as are the Ionian and Aeolian modes, and the Dorian and Mixolydian modes. The Locrian mode is its own reflection." Very nice sounding and grammatical, but gapingly wrong in every point. The only mode that is its own reflection is Dorian. Furthermore, Lydian and Phrygian are not mutual reflections. Phrygian reflected around is root is Ionian. The reflection of Lydian is Locrian; and of Aeolian, Mixolydian. I once loaded a NotebookLM with materials about George Russel's concept of the Lydian Chromatic, and Tonal Gravity. It made an incomprehensible mess of explaining the stuff, worse than the original sources. AI performs well on whatever is the focus of its purveyors. When they want to shake down entry-level coding, they beef it up on entry-level coding and let it loose, leaving it unable to tell Mixolydian from mixing console. | | |
| ▲ | lisbbb 7 days ago | parent | next [-] | | Thank you! This is what I've been trying to tell people about LLMs. They don't hold up. They're like those Western movie set towns that look normal from the front, but when you walk around behind them, you see it is all just scaffolding with false fronts. | | | |
| ▲ | ishyaboibro 7 days ago | parent | prev [-] | | what model did you ask? here's the exact reply I received from Claude Sonnet, which appears to be exactly the answer you were expecting: "Among the seven modes of C major, only Dorian is its own reflection. Understanding Mode Reflections
When we reflect a mode, we reverse its interval pattern. The modes of C major and their interval patterns are: Ionian: W-W-H-W-W-W-H Dorian: W-H-W-W-W-H-W Phrygian: H-W-W-W-H-W-W Lydian: W-W-W-H-W-W-H Mixolydian: W-W-H-W-W-H-W Aeolian: W-H-W-W-H-W-W Locrian: H-W-W-H-W-W-W The Palindromic Nature of Dorian
Dorian mode is palindromic, meaning it produces the same scale whether you read its interval pattern forwards or backwards. When you reverse the Dorian interval pattern W-H-W-W-W-H-W, you get exactly the same sequence: W-H-W-W-W-H-W. Mirror Pairs Among the Other Modes
The remaining modes form mirror pairs with each other: Ionian-Phrygian: Mirror pair Lydian-Locrian: Mirror pair Mixolydian-Aeolian: Mirror pair For example, when you reflect the C major scale (Ionian), which has the interval pattern W-W-H-W-W-W-H, you get H-W-W-W-H-W-W, which corresponds to the Phrygian mode. This symmetrical relationship exists because the whole diatonic scale system can be symmetrically inverted, creating these natural mirror relationships between the modes" | | |
| ▲ | kazinator 7 days ago | parent [-] | | > what model did you ask? Are you hoping to disprove my point by cherry picking the AI that gets the answer? I used Gemini 2.5 Flash. Where can I get an exact list of stuff that Gemini 2.5 Flash does not know that Claude Sonnet does, and vice versa? Then before deciding to consult with AI, I can consult the list? | | |
| ▲ | simianwords 7 days ago | parent | next [-] | | 2.5 flash is particularly cheap and fast, I think 2.5 pro would have got all the answers correct - at least it gets this one correct. | | |
| ▲ | Yokolos 7 days ago | parent | next [-] | | I get a lot of garbage out of 2.5 Pro and Claude Sonnet and ChatGPT. There's always this "this is how you solve it", I take a close look and it's clearly broken, I point it out and it's all "you're right, this is a common issue". Okay, so why do we have to do this song and dance a million times to arrive at the actually correct answer? | |
| ▲ | kazinator 7 days ago | parent | prev [-] | | Why doesn't Flash get it correct, yet comes up with plausible sounding nonsense? That means it is trained on some texts in the area. What would make 2.5 Pro (or anything else) categorically better would be if it could say "I don't know". There will be things that Claude 3.7 or Gemini Pro will not know, and the interpolations they come up with will not make sense. | | |
| ▲ | simianwords 7 days ago | parent [-] | | Model accuracy goes up as you use heavier models. Accuracy is always preferable and the jump from Flash to Pro is considerable. You must rely on your own internal model in your head to verify the answers it gives. On hallucination: it is a problem but again, it reduces as you use heavier models. | | |
| ▲ | Macha 6 days ago | parent [-] | | > You must rely on your own internal model in your head to verify the answers it gives This is what significantly reduces the utility, if it can only be trusted to answer things I know the answer to, why would I ask it anything? | | |
|
|
| |
| ▲ | ryao 7 days ago | parent | prev | next [-] | | Gemini 2.5 Flash is meant for things that have a higher tolerance for mistakes as long as the costs are low and responses are quick. Claude Sonnet is similar, although the trade off it makes between mistake tolerance and cost/speed is more in favor of fewer mistakes. Lately, I have been using Grok 4 and I have had very good results from it. | |
| ▲ | iusewindows 7 days ago | parent | prev | next [-] | | Today I read a stupid Hackernews comment about how AI is useless. Therefore Hackernews is stupid. Oh, I need a filtered list of which comments to read? Do you build computers by ordering random parts off Alibaba and complaining when they are deficient? You are complaining that you need to RTFM for a piece of high tech? | | |
| ▲ | kazinator 7 days ago | parent [-] | | > Oh, I need a filtered list of which comments to read? If they are about something you're not sure about, and you're making decisions based on them ... maybe it would actually help, so yes? > Do you build computers by ordering random parts off Alibaba and complaining when they are deficient? We build computers using parts which are carefully documented by data sheets, which tell you exactly for what ranges of parameters their operation is defined and in what ways. (temperatures, voltages, currents, frequencies, loads, timings, typical circuits, circuit board layouts, programming details ...) |
| |
| ▲ | 7 days ago | parent | prev [-] | | [deleted] |
|
|
| |
| ▲ | danlitt 7 days ago | parent | prev | next [-] | | Like other commenters here, when I try to use them to help with my work, they don't. It's that simple. I have tried AI coding assistants, and they just guess incorrectly. If I know the answer, they generally give me the same answer. If I don't know the answer, they give me gibberish that ends up wasting time. I would love to look over the shoulder of an AI booster who had a really good interaction, because it's hard for me to believe they didn't just already know what they were looking for. | |
| ▲ | harimau777 7 days ago | parent | prev | next [-] | | I think that there's a strong argument to be made that the negatives of having to wade through AI slop outweights the benefits that AI may provide. I also suspect that AI could contribute to enshittification of society; e.g. AI therapy being substituted for real therapy, AI products displacing industrial design, etc. | | |
| ▲ | fc417fc802 7 days ago | parent | next [-] | | > e.g. AI therapy being substituted for real therapy, AI products displacing industrial design, etc. That depends on the quality of the end product and the willingness to invest the resources necessary to achieve a given quality of result. If average quality goes up in practice then I'd chalk that up as a net win. Low quality replacing high quality is categorically different than low quality filling a previously empty void. Therapy in particular is interesting not just because of average quality in practice (therapists are expensive experts) but also because of user behavior. There will be users who exhibit both increased and decreased willingness to share with an LLM versus a human. There's also a very strong privacy angle. Querying a local LLM affords me an expectation of privacy that I don't have when it comes to Google or even Wikipedia. (In the latter case I could maintain a local mirror but that's similar to maintaining a local LLM from a technical perspective making it a moot point.) | |
| ▲ | rockemsockem 7 days ago | parent | prev [-] | | What is this AI slop that you're wading through and where is it? Spam emails are not any worse for being verbose, I don't recognize the sender, I send it straight to spam. The volume seems to be the same. You don't want an AI therapist? Go get a normal therapist. I have not heard of any AI product displacing industrial design, but if anything it'll make it easier to make/design stuff if/when it gets there. Like are these real things you are personally experiencing? |
| |
| ▲ | mopsi 7 days ago | parent | prev | next [-] | | Last night, I asked a LLM to produce an /etc/fstab entry for connecting to a network share with specific options. I was too lazy to look up the options from the manual. It gave me the options separated by semicolons, which is invalid because the config file requires commas as separators. I honestly don't see technology that stumbles over trivial problems like these as something that will replace my job, or any job that is not already automatable within ten thousand lines of Python, anytime soon. The gap between hype and actual capabilities is insane. The more I've tried to apply LLMs to real problems, the more disillusioned I've become. There is nothing, absolutely nothing, no matter how small the task, that I can trust LLMs to do correctly. | | |
| ▲ | ryao 7 days ago | parent [-] | | Which one? There are huge variations between LLMs. Was this a frontier thinking model with tool use? Did you ask it review online references before presenting an answer? |
| |
| ▲ | satyrun 7 days ago | parent | prev | next [-] | | IMO I think it is a combination of being a really great programmer already and then either not all that intellectually curious or so well read and so intellectually curious that LLMs are a step down from being a voracious reader of books and papers. For me, LLMs are also the most useful thing ever but I was a C student in all my classes. My programming is a joke. I have always been intellectually curious but I am quite lazy. I have always had tons of ideas to explore though and LLMS let me explore these ideas that I either wouldn't be able to otherwise or would be too lazy to bother. | | |
| ▲ | rockemsockem 6 days ago | parent [-] | | I don't totally follow what you're saying. Are you saying that LLMs are most useful if you're not intellectually curious, and therefore most interested in immediate answers, but also that they're very useful if you're a really great programmer for an unstated reason? |
| |
| ▲ | shusaku 7 days ago | parent | prev | next [-] | | The concerning thing is that AI contrarianism is being left wing coded. Imagine you’re fighting a war and one side decides “guns are overhyped, let’s stick with swords”. While there is a lot of hype about AI, even the pessimistic take has to admit it’s a game changing tech. If it isn’t doing anything useful for you, that’s because you need to get off your butt and start building tools on top of it. Especially people on the left need to realize how important their vision is to the future if AI. Right now you can see the current US admin having zero concern for AI safety or carbon use. If you keep your head in the dirt saying “bubble!” that’s no problem. But if this is here to stay then you need to get involved. | | |
| ▲ | rockemsockem 6 days ago | parent [-] | | That's a good point and even worse we'll eventually end up with yet another issue where both left and right offer terrible options and there's no nuanced middle ground :/ |
| |
| ▲ | m0llusk 5 days ago | parent | prev | next [-] | | The worst thing in the information space is bullshit. It's worse than lies because it's slicker. LLMs generate extremely dangerous bullshit that can be difficult to correct. Advice given may apply to old versions, outdated legal precedent, or radical opinions. Fully understanding nuance or mistakes can take a lot of effort and still be off. | |
| ▲ | 6 days ago | parent | prev [-] | | [deleted] |
|
| |
| ▲ | decimalenough 7 days ago | parent | next [-] | | Railroads move people and cargo quickly and cheaply from point A to point B. Mechanized textile production made clothing, a huge sink of time and resources before the industrial age, affordable to everybody. What does AI get the consumer? Worse spam, more realistic scams, hallucinated search results, easy cheating on homework? AI-assisted coding doesn't benefit them, and the jury is still out on that too (see recent study showing it's a net negative for efficiency). | | |
| ▲ | azeirah 7 days ago | parent | next [-] | | For learning with self-study it has been amazing. | | |
| ▲ | gamblor956 7 days ago | parent [-] | | Until you dive deeper and discover that most of what the AI agents provided you was completely wrong... There's a reason that AI is already starting to fade out of the limelight with customers (companies and consumers both). After several years, the best they can offer is slightly better chatbots than we had a decade ago with a fraction of the hardware. | | |
| ▲ | simonw 7 days ago | parent | next [-] | | "Until you dive deeper and discover that most of what the AI agents provided you was completely wrong..." Oddly enough, I don't think that actually matters too much to the dedicated autodidact. Learning well is about consulting multiple sources and using them to build up your own robust mental model of the truth of how something works. If you can really find the single perfect source of 100% correct information then great, I guess... but that's never been my experience. Every source of information has its flaws. You need to build your own mental model with a skeptical eye from as many sources as possible. As such, even if AI makes mistakes it can still accelerate your learning, provided you know how to learn and know how to use tips from AI as part of your overall process. Having an unreliable teacher in the mix may even be beneficial, because it enforces the need for applying critical thinking to what you are learning. | | |
| ▲ | JimDabell 7 days ago | parent [-] | | > > "Until you dive deeper and discover that most of what the AI agents provided you was completely wrong..." > Oddly enough, I don't think that actually matters too much to the dedicated autodidact. I think it does matter, but the problem is vastly overstated. One person points out that AIs aren’t 100% reliable. Then the next person exaggerates that a little and says that AIs often get things wrong. Then the next person exaggerates that a little and says that AIs very often get things wrong. And so on. Before you know it, you’ve got a group of anti-AI people utterly convinced that AI is totally unreliable and you can’t trust it at all. Not because they have a clear view of the problem, but because they are caught in this purity spiral where any criticism gets amplified every time it’s repeated. Go and talk to a chatbot about beginner-level, mainstream stuff. They are very good at explaining things reliably. Can you catch them out with trick questions? Sure. Can you get incorrect information when you hit the edges of their knowledge? Sure. But for explaining the basics of a huge range of subjects, they are great. “Most of what they told you was completely wrong” is not something a typical beginner learning a typical subject would encounter. It’s a wild caricature of AI that people focused on the negatives have blown out of all proportion. |
| |
| ▲ | Gud 7 days ago | parent | prev | next [-] | | That has not been the case for me. I use LLMs to study German, so far it’s been an excellent teacher. I also use them to help me write code, which it does pretty well. | |
| ▲ | rockemsockem 7 days ago | parent | prev [-] | | I almost always validate what I get back from LLMs and it's usually right. Even when it isn't it still usually gets me closer to my goal (e.g maybe some UX has changed where a setting I'm looking for in an app has changed, etc). IDK where you're getting the idea that it's fading out. So many people are using the "slightly better chatbots" every single day. Btw if you only think chat GPT is slightly better than what we had a decade ago then I do not believe that you have used any chat bots at all, either 10 years ago or recently because that's actually a completely insane take. | | |
|
| |
| ▲ | fc417fc802 7 days ago | parent | prev | next [-] | | At a minimum, presumably once it arrives it will provide the consumer custom software solutions which are clearly a huge sink of time and resources (prior to the AI age). You're looking at the prototype while complaining about an end product that isn't here yet. | |
| ▲ | osigurdson 7 days ago | parent | prev [-] | | I don't have that negative of a take but agree to some extent. The internet, mobile, AI have all been useful but not in the same way as earlier advancements like electricity, cars, aircraft and even basic appliances. Outside of things that you can do on screens, most people live exactly the same way as they did in the 70s and 80s. For instance, it still takes 30-45 minutes to clean up after dinner - using the same kind of appliances that people used 50 years ago. The same goes for washing clothes, sorting socks and other boring things that even fairly rich people still do. Basically, the things people dreamed about in the 50s - more wealth, more leisure time, robots and flying cars really were the right dream. |
| |
| ▲ | AngryData 7 days ago | parent | prev | next [-] | | The luddites were often the ones that built the mechanized looms. They had nothing against mechanized looms, they had everything against the business owners using their workers talents and knowledge to build an entire operation only to later undercut their wages and/or replace them with lesser paid unskilled workers and reduce the quality of life of their entire community. Getting people to associate the luddites as anti-technology zealots rather than pro-labor organization is one of the most successful pieces of propaganda in history. | | |
| ▲ | Macha 6 days ago | parent | next [-] | | People will also use "look society was fine afterwards" as proof the luddites were wrong, but if you look at the fact the growth of industrial revolution cities was driven by importing more people from the countryside than died of disease, it's not clear at all that they were wrong about it's impact on their society, even if it worked out alright for us in the aftermath. | |
| ▲ | gruez 6 days ago | parent | prev | next [-] | | >The luddites were often the ones that built the mechanized looms. Source? Skimming the wikipedia article it definitely sounds like most were made up of former skilled textile workers that were upset they were replaced with unskilled workers operating the new machines. > They had nothing against mechanized looms, they had everything against the business owners using their workers talents and knowledge to build an entire operation only to later undercut their wages and/or replace them with lesser paid unskilled workers and reduce the quality of life of their entire community. Sounds a lot like the anti-AI sentiment today, eg. "I'm not against AI, I'm just against it being used by evil corporations so they don't have to hire human workers". The "AI slop" argument also resembles luddites objecting to the new machines on the quality of "quality" (also from wikipedia), although to be fair that was only a passing mention. | |
| ▲ | GJim 6 days ago | parent | prev [-] | | > Getting people to associate the luddites as anti-technology zealots Interestingly.... ..... the fact that luddites also called for unemployment compensation and retraining for workers displaced by the new machinery, probably makes them amongst the most forward thinking and progressive people of the 1800's. |
| |
| ▲ | no_wizard 7 days ago | parent | prev | next [-] | | Luddites weren’t anti technology at all[0] in fact they were quite adept at using technology. It was a labor movement that fought for worker rights in the face of new technologies. [0]: https://www.newyorker.com/books/page-turner/rethinking-the-l... | |
| ▲ | trod1234 7 days ago | parent | prev | next [-] | | Apples to oranges. Luddites weren't at a point where every industry sees individual capital formation/demand for labor trend towards zero over time. Prices are ratios in the currency between factors and producers. What do you suppose happens when the factors can't buy anything because there is nothing they can trade. Slavery has quite a lot of historic parallels with the trend towards this. Producers stop producing when they can make no profit. You have a deflationary (chaotic) spiral towards socio-economic collapse, under the burden of debt/money-printing (as production risk). There are limits to systems, and when such limits are exceeded; great destruction occurs. Malthus/Catton pose a very real existential threat when such disorder occurs, and its almost inevitable that it does without action to prevent it. One cannot assume action will happen until it actually does. | |
| ▲ | shadowgovt 7 days ago | parent | prev | next [-] | | With this generation of AI, it's too early to tell whether it's the next railroad, the next textile machine, or the next way to lock your exclusive ownership of an ugly JPG of a multicolored ape into a globally-referenceable, immutable datastore backed by a blockchain. | |
| ▲ | bgwalter 7 days ago | parent | prev | next [-] | | The mechanical loom produced a tangible good. That kind of automation was supposed to free people from menial work. Now they are trying to replace interesting work with human supervised slop, which is a stolen derivative work in the first place. The loom wasn't centralized in four companies. Customers of textiles did not need an expensive subscription. Obviously average people would benefit more if all that investment went into housing or in fact high speed railways. "AI" does not improve their lives one bit. | |
| ▲ | wat10000 7 days ago | parent | prev | next [-] | | They laughed at Einstein, but they also laughed at Bozo the Clown. This sort of “other people were wrong once, so you might be too” comment is really pointless. | |
| ▲ | 7 days ago | parent | prev | next [-] | | [deleted] | |
| ▲ | harimau777 7 days ago | parent | prev [-] | | I mean, for them it probably was. |
|