| ▲ | There is no spoon – A software engineers primer for demystified ML(github.com) |
| 93 points by jmatthews a day ago | 32 comments |
| |
|
| ▲ | janalsncm 20 hours ago | parent | next [-] |
| I personally think it is much more important to have strong statistical intuitions rather than intuitions about what neural networks are doing. The latter isn’t wrong or useless. It’s simply not something a typical software engineer will need. On the other hand, wiring up LLMs into an application is very popular and may be an engineer’s first experience with systems that are fundamentally chaotic. Knowing the difference between precision and recall and when you care about them will get you a lot more bang for your buck. I would suggest the gateway drug into ML for most engineers is something like: we have a task and it can currently be done for X dollars. But maybe we can do it for a tenth of the price with a different API call. Or maybe there’s something on Huggingface that does the same thing for a fixed hourly cost, hundreds of times cheaper in practice. |
| |
| ▲ | jmatthews 20 hours ago | parent [-] | | I'm just trying to develop the lens where I can see a problem and know what properties of it are meaningful from an ML standpoint. Coming from a specific domain where I have a sharpened instinct for how things are haven't really given me the ability to decompose the problem using ML primitives. That's what I'm working on. |
|
|
| ▲ | bonoboTP 19 hours ago | parent | prev | next [-] |
| Just read a good textbook instead of this LLM-written stuff. For example those by Murphy or Prince or Bishop. Or one of many YouTube lecture series from MIT or Stanford. There are many primer 101 tutorials and Medium posts. But if you actually want to learn instead of procrastinating, pick up a real textbook or work through a course. |
| |
| ▲ | mememememememo 17 hours ago | parent | next [-] | | Or just train some NNs. If more time write the code and understand the tensor operations. | |
| ▲ | jmatthews 19 hours ago | parent | prev [-] | | I've bounced off of many good textbooks. Even Karpathy's YouTube series was too dense for me. I'm trying to come in at a more palatable level. This was a two day exploration where I provided the syllabus and ran through it with Claude Code, asking questions, trying to anchor it to stuff I understand well. I feel like the artifact has value. | | |
| ▲ | bonoboTP 19 hours ago | parent | next [-] | | I think chatting with an llm alongside a textbook can be helpful but producing learning material when you yourself are a novice is not really that valuable. | | | |
| ▲ | profsummergig 13 hours ago | parent | prev | next [-] | | FWIW, I found it quite useful. I liked that a huge amount of AI/LLM concepts are mentioned and compared. So it's a handy reference. | |
| ▲ | antonvs 19 hours ago | parent | prev [-] | | It's AI slop. You're letting a machine gaslight you. |
|
|
|
| ▲ | hilliardfarmer 15 hours ago | parent | prev | next [-] |
| Please stop trying to trick us into reading AI generated text. "This isn't a textbook or a tutorial. It's a mental model — the abstractions you need to reason about ML systems the way you already reason about software systems." |
| |
| ▲ | jmatthews 13 hours ago | parent | next [-] | | It's not a trick bud. The github page shows my user name and Claude. The content is intended to be read by an AI agent and explored through a text interface. That is explicit in the readme and the primer itself. If you think you can generate this artifact with a prompt then show me. This was 2 days of exploration and research. | | |
| ▲ | bonoboTP 12 hours ago | parent | next [-] | | If you think that "2 days" makes it sound a lot... You'd be surprised how long it takes to actually make learning materials. I don't want to be too harsh, in case you're a high school student etc. I see it's good faith, but do note the reaction here. | | |
| ▲ | jmatthews 6 hours ago | parent [-] | | I'm trying to untangle the "this content isn't valuable" signal from the luddite "Anything with AI is low effort slop" I appreciate the former and am trying to filter the latter. | | |
| ▲ | bonoboTP 5 hours ago | parent [-] | | I read a couple of good analogies to predict how you and others will feel about your AI content: 1) telling people at the breakfast table about the dream you just had, 2) showing all your loose acquaintances the photos of your newborn baby. That is, it's very precious and interesting to you, but it really isn't to anyone else. This is true about generated text, images and songs. I've generated a lot of what I think of as bangers with Suno but learned quickly that they have zero value to anyone else. Part of the value to me is the thrill and dopamine hits of having generated it. This simply doesn't translate to anyone else. It will take a while until society internalizes this. This is not to say that AI can't have any role in the creative process. But the effort will be still high and original human thinking and intent and input is still very important. | | |
| ▲ | jmatthews 5 hours ago | parent [-] | | it's a worthwhile lesson. thank you. There was a great deal of effort on my part, but not in the prose. You've taught me something and I appreciate it. |
|
|
| |
| ▲ | bonoboTP 12 hours ago | parent | prev [-] | | An AI agen won't need this, it has been trained on a lot of ML knowledge already. It's basic stuff. | | |
| ▲ | jmatthews 6 hours ago | parent [-] | | it's not that you're teaching the AI, it's that you're framing the conversation on a reference material and having a conversation around it. Exploring a problem with referential framing, like a white paper or a dense blog post is a useful cognitive hack. You just have to be careful to pin extraordinary claims to extraordinary evidence. |
|
| |
| ▲ | thirtygeo 14 hours ago | parent | prev [-] | | I got to that and just stopped reading! |
|
|
| ▲ | whoamii 19 hours ago | parent | prev | next [-] |
| Feature request for HN, Adblocker, etc: please block pages with the text “it isn’t X, it is Y”. |
| |
| ▲ | tzs 6 hours ago | parent | next [-] | | Block pages that contain a construct that is commonly used by humans in human writing and has been for centuries because AI that trained on human writing also uses it? Should we also block pages that contain no spelling errors or no grammatical errors? | |
| ▲ | ggambetta 15 hours ago | parent | prev | next [-] | | Can we also ban anything where the second line is "let that sink in"? And anything claiming that "X is a masterclass in Y" (especially for (tweet, empathy))? | |
| ▲ | thirtygeo 14 hours ago | parent | prev | next [-] | | Careful, you'll just make the output generators harder to spot! Www
We need to keep their 'Tells' hidden from them.. | |
| ▲ | jmatthews 13 hours ago | parent | prev [-] | | my apologies it wasn't up to your standards. In fairness to me, that line is exactly what my effort was. I wasn't trying to "learn ML" I am trying to build a mental model that let's me decompose real problem into ML primitives. It's unclear to me if you think the resource has no value or if it bother you that I wrote it using a coding agent. I wrote the syllabus and worked through each section. Where my understanding was weak I explored the space, pulled in research, referred the model to other sources, and just generally tried to ground the topic in something I understood. What resulted was something that helped a lot of subjects click together for me. Especially when to reach for a particular activation function and the section on gating. This enter survey was motivated by an ML expirent I ran with assosicative memories that just failed horribly. So rather than post mortem that I set about understanding why it failed. Anyhow, thank you for the feedback. I submitted this in good faith that it may help others. | | |
| ▲ | 2sb 10 hours ago | parent [-] | | “Decompose a real problem into ML primitives”
What does this actually mean? Be careful of AI hallucinations they are dangerous. | | |
| ▲ | jmatthews 6 hours ago | parent [-] | | I'll give a software example because I'm just better at it, but if you describe Trello and tell me to decompose it on a white board I start thinking in terms of queues and write|edit|save tuples. I don't have to invent queues in my head and I have to wonder if it is possible to proactively assign a series of tranforms or schedules on an fresh input. I know how to do it and it's all internalized. Even if I've never needed to do it. that's the toolbox I'm trying to develop in ML. For example. I've studies LSTM and implemented one. What I didn't know was if gating was turing complete and essentially unbound. I didn't know if gates could be arbitrarily complex. Importantly I had no idea how to translate "I need a switch here" to a gate, or if a switch was even possible given the need to be differentiable for backprop. |
|
|
|
|
| ▲ | jmatthews a day ago | parent | prev | next [-] |
| This is my weekend project. I am building up my pattern recognition in machine learning. By that I mean see X problem, instantly think of Y solution. The primer markdown file is the artifact of that exploration. read it from top to bottom or better have your favorite language model read it and then explore the space with a strong guided syllabus. |
| |
| ▲ | janalsncm 20 hours ago | parent | next [-] | | Framing a business problem in terms of ML is indeed important. Where does classification come in, where does regression come in, when to use retrieval, when to use generative solutions. Would be a good section to add imo. | | | |
| ▲ | TheTaytay 16 hours ago | parent | prev [-] | | I quite liked this. It feels approachable and to-the-point. |
|
|
| ▲ | zar1048576 19 hours ago | parent | prev | next [-] |
| Nice weekend project! Even though there are copious resources out there (textbooks, videos, etc.), those may not appeal to everyone. People have different preferred modalities for consuming information and there is always value in (correctly) reframing concepts in a way that can be better understood by people who don’t resonate with traditional textbooks and YouTube videos. I’m
glad you found a formulation that works for you, and judging by the number of upvotes, it resonated with others as well. At the very least, I’m sure that working on this improved your understanding as well! |
|
| ▲ | oleggromov 21 hours ago | parent | prev [-] |
| Thank you for sharing! Saved to bookmarks to read on my free time. Hopefully I'll get some soon :) |
| |