| ▲ | stavros 6 hours ago |
| I see this fallacy being committed a lot these days. "Because LLMs, you will no longer need a skill you don't need any more, but which you used to need, and handwaves that's bad". Academia doesn't want to produce astrophysics (or any field) scientists just so the people who became scientists can feel warm and fuzzy inside when looking at the stars, it wants to produce scientists who can produce useful results. Bob produced a useful result with the help of an agent, and learned how to do that, so Bob had, for all intents and purposes, the exact same output as Alice. Well, unless you're saying that astrophysics as a field literally does not matter at all, no matter what results it produces, in which case, why are we bothering with it at all? |
|
| ▲ | djaro 6 hours ago | parent | next [-] |
| The problem is that LLMs stop working after a certain point of complexity or specificity, which is very obvious once you try to use it in a field you have deep understanding of. At this point, your own skills should be able to carry you forward, but if you've been using an LLM to do things for you since the start, you won't have the necessary skills. Once they have to solve a novel problem that was not already solved for all intentes and purposes, Alice will be able to apply her skillset to that, whereas Bob will just run into a wall when the LLM starts producing garbage. It seems to me that "high-skill human" > "LLM" > "low-skill human", the trap is that people with low levels of skills will see a fast improvement of their output, at the hidden cost of that slow build-up of skills that has a way higher ceiling. |
| |
| ▲ | stavros 6 hours ago | parent | next [-] | | Then test Bob on what you actually want him to produce, ie novel problems, instead of trivial things that won't tell you how good he is. Why is it a problem of the LLM if your test is unrelated to the performance you want? | | |
| ▲ | skydhash 5 hours ago | parent | next [-] | | What people forget about programming is it is a notation for formal logic, one that can be executed by a machine. That formal logic is for solving a problem in the real world. While we have a lot of abstractions that solve some subproblems, there still need to connect those solutions to solve the main problem. And there’s a point where this combination becomes its own technical challenge. And the skill that is needed is the same one as solving simpler problems with common algorithms. | |
| ▲ | troupo 6 hours ago | parent | prev [-] | | How can Bob produce novel things when he lacks the skills to do even trivial things? I didn't get to be a senior engineer by immediately being able to solve novel problems. I can now solve novel problems because I spent untold hours solving trivial ones. | | |
| ▲ | stavros 6 hours ago | parent [-] | | Because trivial things aren't a prerequisite for novel things, as any theoretical mathematician who can't do long division will tell you. | | |
| ▲ | sgarland 5 hours ago | parent | next [-] | | I would love to see someone attempt to do multiplication who never learned addition, or exponentiation without having learned multiplication. There is a vast difference between “never learned the skill,” and “forgot the skill from lack of use.” I learned how to do long division in school, decades ago. I sat down and tried it last year, and found myself struggling, because I hadn’t needed to do it in such a long time. | | |
| ▲ | thepasch 2 hours ago | parent | next [-] | | > There is a vast difference between “never learned the skill,” and “forgot the skill from lack of use.” This sentence contains the entire point, and the easiest way to get there, as with many, many things, is to ask “why?” | |
| ▲ | ipaddr 4 hours ago | parent | prev [-] | | Most people learn multiplication by memorizing a series of cards 2x2,2x3.. 9x9. Later this gets broken down to addition in higher grades. | | |
| ▲ | Jensson 4 hours ago | parent | next [-] | | Most people learn multiplication by counting, it has been in basic mathbooks since forever. "1 box has 4 cookies. Jenny ha 4 boxes of cookies. How many cookies do Jenny have?" and so on, the kids solve that by counting 4 cookies in every of the 4 boxes and reaching 16. Only later do you learn those tables. | |
| ▲ | sgarland 2 hours ago | parent | prev [-] | | That’s definitely not how I learned it, nor how my kids have learned it. I vividly remember writing out “2 x 3 = 2 + 2 + 2 = 3 + 3.” I later memorized the multiplication table up to 12, yes, but that was not a replacement of understanding what multiplication was |
|
| |
| ▲ | Folcon 5 hours ago | parent | prev | next [-] | | There's a difference between needing no trivial skills to do novel things and not needing specific prerequisite trivial skills to do a novel thing | |
| ▲ | troupo 6 hours ago | parent | prev [-] | | Ah yes. The famous theoretical mathematicians who immediately started on novel problems in theoretical mathematics without first learning and understanding a huge number of trivial things like how division works to begin with, what fractions are, what equations are and how they are solved etc. Edit: let's look at a paper like Some Linear Transformations on Symmetric Functions Arising From a Formula of
Thiel and Williams https://ecajournal.haifa.ac.il/Volume2023/ECA2023_S2A24.pdf and try and guess how many of trivial things were completely unneeded to write a paper like this. | | |
| ▲ | stavros 5 hours ago | parent [-] | | Seems that teaching Bob trivial things would be a simple solution to this predicament. | | |
| ▲ | sumeno 4 hours ago | parent [-] | | That's what the program he just took was supposed to be for, learning not output. You've just reinvented the article from first principles, congrats | | |
| ▲ | HauntingPin 3 hours ago | parent | next [-] | | Sometimes I wonder how deeply some people actually read these articles. What's the point of the comments if all we're doing is re-explaining what's already explained in such a precise and succint manner? It's a fantastic article. It's so well-written and clear. And yet we're stuck going in a circle repeating what's in the article to people who either didn't read it, or didn't read it with the care it deserves. | |
| ▲ | thepasch 2 hours ago | parent | prev [-] | | > That’s what the program he just took was supposed to be for, learning not output. If you send a kid to an elementary school, and they come back not having learned anything, do you blame the concept of elementary schools, or do you blame that particular school - perhaps a particular teacher _within_ that school? |
|
|
|
|
|
| |
| ▲ | brookst 5 hours ago | parent | prev [-] | | This whole argument can be made for why every programmer needs to deeply understand assembly language and computer hardware. At a certain point, higher level languages stop working. Performance, low level control of clocks and interrupts, etc. I’m old enough dropping into assembly to be clever with the 8259 interrupt controller really was required. Programmers today? The vast majority don’t really understand how any of that works. And honestly I still believe that hardware-up understanding is valuable. But is it necessary? Is it the most important thing for most programmers today? When I step back this just reads like the same old “kids these days have it so easy, I had to walk to school uphill through the snow” thing. | | |
| ▲ | imtringued 3 hours ago | parent [-] | | Teaching how computer hardware works is pretty smart. There is no need to do it in depth though. Writing assembly is probably completely irrelevant. You should still know how programming language concepts map to basic operations though. Simple things like strict field offsets, calling conventions, function calls, dynamic linking, etc. | | |
| ▲ | sgarland 2 hours ago | parent [-] | | > Writing assembly is probably completely irrelevant. ffmpeg disagrees. More broadly, though, it’s a logical step if you want to go from “here’s how PN junctions work” to “let’s run code on a microprocessor.” There was a game up here yesterday about building a GPU, in the same vein of nand2tetris, Turing Complete, etc. I find those quite fun, and if you wanted to do something like Ben Eater’s 8-bit computer, it would probably make sense to continue with assembly before going into C, and then a higher-level language. |
|
|
|
|
| ▲ | nandomrumber 6 hours ago | parent | prev | next [-] |
| > why are we bothering with it at all? Because we largely want people who have committed to tens of thousands of dollars of debt to feel sufficiently warm and fuzzy enough to promote the experience so that the business model doesn’t collapse. It’s difficult to think anyone would end up truly regretting doing a course in astrophysics, or any of the liberal arts and sciences if they have a modicum of passion, but it’s very believable that a majority of them won’t go on to have a career in it, whatever it is, directly. They’re probably more likely to gain employment on their data science skills, or whether core competencies they honed, or just the fact that they’ve proven they can learn highly abstract concepts, or whatever their field generalises to. Most of the jobs are in not-highly-specific academic-outcome. |
| |
| ▲ | imtringued 3 hours ago | parent [-] | | Even if you land a job in your field, you will encounter that academia is backwards vs industry in some aspects and decades ahead of what is adopted in the industry in other aspects to the point where both of these mean that you won't make much use of the skills you learned in university. |
|
|
| ▲ | pards 6 hours ago | parent | prev | next [-] |
| > Take away the agent, and Bob is still a first-year student who hasn't started yet. The year happened around him but not inside him. He shipped a product, but he didn't learn a trade. We're minting an entire generation of people completely dependent on VC funding. What happens if/when the AI companies fail to find a path to profitability and the VC funding dries up? |
| |
| ▲ | Paradigma11 5 hours ago | parent | next [-] | | What will happen is pretty obvious. Those companies will either be classified as too important to fail and get government support or go bankrupt and will be bought for pennies on the dollar. For the customers nothing much will change since tokens are getting cheaper every year and the business is already pretty profitable. Progress will slow down massively till local open weight models catch up to pre-crash SotA and go on from there. | | | |
| ▲ | stavros 6 hours ago | parent | prev [-] | | Do you think that'll take a generation to happen? | | |
| ▲ | rafterydj 6 hours ago | parent [-] | | ChatGPT 3.5 came out coming on 4 years ago now. I don't think a human generation (~20-30 years) needs to be the benchmark here, but new juniors in the industry for a handful of years can be said to be a whole "generation". That how I was reading OP. |
|
|
|
| ▲ | hirako2000 6 hours ago | parent | prev | next [-] |
| I was reading in the article that what matters is the process that leads to the (typically useless) result, what the people get out of it. Once I realized that this white on black contrast was hurting my eyes, I decided to stop as I didn't want to see stripes for too long when looking away. Some activity has outcomes that aren't strictly in the results. |
| |
| ▲ | stavros 6 hours ago | parent [-] | | Yeah, it was saying that what matters is the process of training people to be good scientists, so they can produce other, more useful, results. That's literally what training is, everywhere. This argument boils down to "don't use tools because you'll forget how to do things the hard way", which nobody would buy for any other tool, but with LLMs we seem to have forgotten that line of reasoning entirely. | | |
| ▲ | rglullis 6 hours ago | parent | next [-] | | > so they can produce other, more useful, results But to even *know* what is more useful, it is crucial to have walked the walk. Otherwise we will all end up with a bunch of people trying to reinvent the wheel, over and over again, like JavaScript "developers" who keep reinventing frameworks every six months. > which nobody would buy for any other tool I don't know about you, but I wasn't allowed to use calculators in my calculus classes precisely to learn the concepts properly. "Calculators are for those who know how to do it by hand" was something I heard a lot from my professors. | | |
| ▲ | thepasch an hour ago | parent [-] | | > But to even know what is more useful, it is crucial to have walked the walk. I feel like people tend to forget that among the many things LLMs can do these days, “using a search engine” is among them. In fact, they use them better than the majority of people do! The conversation people think they’re having here and the conversation that actually needs to be had are two entirely different conversations. > I don’t know about you, but I wasn’t allowed to use calculators in my calculus classes precisely to learn the concepts properly. “Calculators are for those who know how to do it by hand” was something I heard a lot from my professors. Suppose I never learned how to derive a function. I don’t even know what a function is. I have no idea how to do make one, write one, or what it even does. So I start gathering knowledge: - A function is some math that allows you to draw a picture of how a number develops if you do that math on it. - A derivative is a function that you feed a function and a number into, and then it tells you something about what that function is doing to that number at that number. - “What it’s doing” specifically means not the result of the math for that particular number, but the results for the immediate other numbers behind and in front of it. - This can tell us about how the function works. Now I go tell ClaudeGPTimini “hey, can you derive f(x) at 5 so that we can figure out where it came from and where it goes from there?”, and it gives me a result. I’ve now ostensibly understood what a derivative does and what it’s used for, yet I have zero idea how to mathematically do it. Does that make any results I gain from this intuitive understanding any less valuable? What I’ll give you is this: if I knew exactly how the math worked, then it would be far easier for me to instantly spot any errors ClaudeGPTimini produced. And the understanding of functions and derivatives outlined above may be simplistic in some places (intentionally so), in ways that may break it in certain edge cases. But that only matters if I take its output at face value. If I get a general understanding of something and run a test with it, I’ll generally have some sort of hypothesis of what kind of result I’m expecting, given that my understanding is correct. If I know that a lot of unknown unknowns exist around a thing I’m working with, then I also know that unexpected results, as well as expected ones, require more thorough verification. Science is what happens when you expect something, test something, and get a result - expected OR unexpected - and then systematically rule out that anything other than the thing you’re testing has had an effect on that result. This is not a problem with LLMs. It’s a thing we should’ve started teaching in schools decades ago: how to understand that there are things you don’t understand. In my view, the vast majority of problems plaguing us as a species lies in this fundamental thing that far too many people are just never taught the concept of. |
| |
| ▲ | defrost 6 hours ago | parent | prev | next [-] | | > This argument boils down to "don't use tools because you'll forget how to do things the hard way", which nobody would buy for any other tool, This is false. There absolutely are people that fall back on older tools when fancy tools fail. You will find such people in the military, in emergency services, in agriculture, generally in areas where getting the job done matters. Perhaps you're unfamiliar. They other week I finished putting holes in fence posts with a bit and brace as there was no fuel for the generator to run corded electric drills and the rechargable batteries were dead. Ukrainians, and others, need to fall back on no GPS available strategies and have done so for a few years now. etc. | | |
| ▲ | thijson 5 hours ago | parent | next [-] | | In the 80's the Americans thought that the Russians were backwards to be still using vacuum tubes in their military vehicles. Later they found out that they were being used because they are more tolerant to EMP from a nuclear blast. | |
| ▲ | Kon5ole 5 hours ago | parent | prev [-] | | >This is false. There absolutely are people that fall back on older tools when fancy tools fail.
>They other week I finished putting holes in fence posts with a bit and brace as there was no fuel for the generator to run corded electric drills and the rechargable batteries were dead. It depends on the task though. If you are in a similar scenario as with your fence posts and want to edit computer programs, you can't. (Not even with xkcd's magnetic needle and a steady hand). ;-) As technology marches on it seems inevitable that we will get increasingly large and frequent knowledge gaps. Otherwise progress would stop - we need the giant shoulders to stand on. How many people in the world can recreate a ASML lithography machine vs how many people are surviving by doing something that requires that machine to exist? |
| |
| ▲ | hirako2000 6 hours ago | parent | prev | next [-] | | There is an argument to make that tools that speed up a process whilst keeping acuity intact are legitimate. LLMs, the way they typically get used, are solely to save time by handing over nearly the entire process. In that sense acuity can't remain intact, even less so improving over time. | | |
| ▲ | stavros 6 hours ago | parent [-] | | So? | | |
| ▲ | hirako2000 5 hours ago | parent [-] | | You previous comment reads as if LLMs get some unjustified different treatment. Do you agree the different treatment is justified ? (Many do not). Or are you asking , so what if acuity is diminished so long as an LLM does the job equally well? |
|
| |
| ▲ | nathan_compton 6 hours ago | parent | prev [-] | | People say this in a very large number of other contexts. Mathematica has been able to do many integrals for decades and yet we still make students learn all the tricks to integrate by hand. This pattern is very common. | | |
| ▲ | FrojoS 4 hours ago | parent [-] | | Yes. But to be fair to your specific point, symbolic solving of integrals used to be a huge skill in the engineering education. Nowadays, it is not a focus anymore, because numerical solutions are either sufficiently accurate or, more importantly, the only feasible approach anyway. | | |
| ▲ | nathan_compton 3 hours ago | parent [-] | | There is much more to life than engineering. | | |
| ▲ | FrojoS 3 hours ago | parent [-] | | Sorry, I should have quoted properly in my reply.
My first sentence ("Yes.") was in general agreement with you, the second sentence was specifically about > Mathematica has been able to do many integrals for decades and yet we still make students learn all the tricks to integrate by hand But maybe, integrating by hand is still as big as ever in other parts of academia. Or were you thinking about high school? I'm fairly sure, that symbolic solving of integrals is treated as less important in education these days, than it was before digital computers, but I could be wrong. Mathematica's symbolic solve sure is very useful, but numeric solutions are what really makes the art of finding integrals much less relevant. | | |
| ▲ | nathan_compton 2 hours ago | parent [-] | | I studied physics and mathematics and finding analytic solutions to problems is still useful and enlightening. |
|
|
|
|
|
|
|
| ▲ | asHg19237 5 hours ago | parent | prev | next [-] |
| The arguments of the LLM psychosis afflicted get more and more desperate. Astrophysics is about understanding and thinking, this comment paints it as result oriented (whatever that means). The industrialization of academia hasn't even produced more results, it has produced more meaningless papers. Just like LLMs produce the 10.000th note taking app, which for the LLM psychosis afflicted is apparently enough. |
| |
|
| ▲ | dwa3592 3 hours ago | parent | prev | next [-] |
| Hard sciences play this crucial and often unseen role in our society : they help train humans to develop critical thinking. Not everyone with PhD in Astrophysics ends up doing Astrophysics in life; it's a discipline, or a training regime for our minds. After that PhD; the result is a human being who can tackle hard problems. We have many other such disciplines (basically any PhD in hard sciences) which produces this outcome. |
|
| ▲ | mzhaase 6 hours ago | parent | prev | next [-] |
| Why should we only do things that produce some sort of value? Do we really want to reduce all of human existence to increasing profits? |
| |
| ▲ | stavros 6 hours ago | parent | next [-] | | You said "value" and "profit". I said "useful". | |
| ▲ | nemo44x 6 hours ago | parent | prev [-] | | What’s a better method for determining how to utilize and distribute resources? To determine where energy should be used and where it should be moved from? | | |
| ▲ | sgarland 5 hours ago | parent [-] | | Some things are just enjoyable. I get no real utility from photography - it’s not my career, it’s not a side gig, and I’m not giving prints out as gifts. Most of the shots never get printed at all. I do it because I enjoy the act itself, of knowing how to make an image frozen in time look a particular way by tweaking parameters on the camera, and then seeing the result. I furthermore enjoy the fact that I could achieve the same result on a dumb film camera, because I spent time learning the fundamentals. |
|
|
|
| ▲ | cmiles74 3 hours ago | parent | prev | next [-] |
| Until the LLM is wrong and Bob passes the erroneous result off as accurate, reliable and vetted by a knowledgeable person. At that point Bob is not producing a useful result. Then it becomes a trap other people might get caught in, wasting valuable time and energy. |
|
| ▲ | dsqrt 5 hours ago | parent | prev | next [-] |
| The goal of academic research is to create understanding, not papers. If we outsource all research to LLMs, then we are only producing the latter. |
|
| ▲ | sega_sai 5 hours ago | parent | prev | next [-] |
| You missed the argument. When we are talking about faculty, yes their result is the only thing that matters, so if it was produced quicker with a LLM, that's great. But when we are talking about the student, there is a drastic difference in the student in the with LLM vs without LLM cases. In the latter they have much better understanding. And that matters in the system when we are educating future physicists. |
|
| ▲ | nathan_compton 6 hours ago | parent | prev | next [-] |
| Is that what "academia" wants? Last I checked "academia" is not a dude I can call and ask for an opinion or definition of what it was interested in. I will make an explicit, plausible, counterpoint: academia wants to produce understanding. This is, more or less, by definition, not possible with an AI directly (obviously AIs can be useful in the process). Take GR as an example. The vast majority of the dynamical character of the theory is inaccessible to human beings. We study it because we wanted to understand it, and only secondarily because we had a concrete "result" we were trying to "achieve." A person who cares only about results and not about understanding is barely a person, in my opinion. |
|
| ▲ | selimthegrim 5 hours ago | parent | prev | next [-] |
| Completely missed the point of the blog post which was that the point was producing the scientist not the result |
|
| ▲ | gedy 5 hours ago | parent | prev [-] |
| We aren't talking pocket calculators here (I see the irony of phone app in pocket), LLMs are hugely expensive things made and controlled behind costly commercial subscriptions. And likely in the middle of a huge investment bubble and stability is uncertain. So we all need to be careful about "gee we don't need that skill or person anymore", etc. |
| |
| ▲ | danielbln 4 hours ago | parent [-] | | Open weight models that run under your desk are not frontier model level, but they are getting closer. Improvements in agentic post training and things like TurboQuant mean that even if all frontier labs pull the plug tomorrow, we will still have agents to work with. | | |
| ▲ | zozbot234 3 hours ago | parent | next [-] | | TurboQuant is not a step change, it's more of a smaller incremental improvement to KV quantization, and possibly (unsure) to quantization more generally. I'm actually more positive about SSD weights offload, which opens up very large local models for slow inference (good enough for slow chat) to virtually any hardware or amount of RAM. | |
| ▲ | gedy 4 hours ago | parent | prev [-] | | I'm definitely looking forward to that, as I really want people to control their own tools. |
|
|