| ▲ | SimpleFold: Folding proteins is simpler than you think(github.com) |
| 297 points by kevlened 9 hours ago | 104 comments |
| https://arxiv.org/abs/2509.18480 |
|
| ▲ | hashta 7 hours ago | parent | next [-] |
| One caveat that’s easy to miss: the "simple" model here didn’t just learn folding from raw experimental structures. Most of its training data comes from AlphaFold-style predictions. Millions of protein structures that were themselves generated by big MSA-based and highly engineered models. It’s not like we can throw away all the inductive biases and MSA machinery, someone upstream still had to build and run those models to create the training corpus. |
| |
| ▲ | aDyslecticCrow 7 hours ago | parent | next [-] | | What i take away is the simplicity and scaling behavior. The ML field often sees an increase in module complexity to reach higher scores, and then a breakthrough where a simple model performs on-par with the most complex. That such a "simple" architecture works this well on its own, means we can potentially add back the complexity again to reach further. Can we add back MSA now? where will that take us? My rough understanding of field is that a "rough" generative model makes a bunch of decent guesses, and more formal "verifiers" ensure they abide by the laws of physics and geometry. The AI reduce the unfathomably large search-space so the expensive simulation doesn't need to do so much wasted work on dead-ends. If the guessing network improves, then the whole process speeds up. - I'm recalling the increasingly complex transfer functions in redcurrant networks, - The deep pre-processing chains before skip forward layers. - The complex normalization objectives before Relu. - The convoluted multi-objective GAN networks before diffusion. - The complex multi-pass models before full-convolution networks. So basically, i'm very excited by this. Not because this itself is an optimal architecture, but precisely because it isn't! | | |
| ▲ | nextos 7 hours ago | parent [-] | | > Can we add back MSA now? Using MSAs might be a local optimum. ESM showed good performance on some protein problems without MSAs. MSAs offer a nice inductive bias and better average performance. However, the cost is doing poorly on proteins where MSAs are not accurate. These include B and T cell receptors, which are clinically very relevant. Isomorphic Labs, Oxford, MRC, and others have started the OpenBind Consortium (https://openbind.uk) to generate large-scale structure and affinity data. I believe that once more data is available, MSAs will be less relevant as model inputs. They are "too linear". |
| |
| ▲ | slashdave 4 hours ago | parent | prev | next [-] | | Correct. For those that might not follow, the MSA is used to generalize from known PDB structures to new sequences. If you train on AlphaFold2 results, those results include that generalization, so that your model no longer needs that capability (you can rely on rote memorization). This simple conclusion seems to have escaped the authors. | |
| ▲ | godelski 7 hours ago | parent | prev | next [-] | | Is this so unusual? Almost everything that is simple was once considered complex. That's the thing about emergence, you have to go through all the complexities first to find the generalized and simpler formulations. It should be obvious that things in nature run off of relatively simple rulesets, but it's like looking at a Game of Life and trying to reverse engineer those rules AND the starting parameters. Anyone telling you such a task is easy is full of themselves. But then again, who seriously believes that P=NP? | | |
| ▲ | hashta 6 hours ago | parent [-] | | To people outside the field, the title/abstract can make it sound like folding is just inherently simple now, but this model wouldn’t exist without the large synthetic dataset produced by the more complex AF. The "simple" architecture is still using the complex model indirectly through distillation. We didn’t really extract new tricks to design a simpler model from scratch, we shifted the complexity from the model space into the data space (think GPT-5 => GPT-5-mini, there’s no GPT-5-mini without GPT-5) | | |
| ▲ | stavros 6 hours ago | parent | next [-] | | But this is just a detail, right? If we went and painstakingly catalogued millions of proteins, we'd be able to use the simple model without needing a complex model to generated data, no? | |
| ▲ | godelski 4 hours ago | parent | prev [-] | | > To people outside the field
So what?It's a research paper. That's not how you communicate to a general audience. Just because the paper is accessible in terms of literal access doesn't mean you're the intended audience. Papers are how scientists communicate to other scientists. More specifically, it is how communication happens between peers. They shouldn't even be writing for just other scientists. They shouldn't be writing for even the full set of machine learning researchers nor the full set of biologists. Their intended audience is people researching computational systems that solve protein folding problems. I'm sorry, but where do you want scientists to be able to talk directly to their peers? Behind closed doors? I just honestly don't understand these types of arguments. Besides, anyone conflating "Simpler than You Think" as "Simple" is far from qualified from being able to read such a paper. They'll misread whatever the authors say. Conflating those two is something we'd expect from an Elementary School level reader who is unable to process comparative statements. I don't think we should be making that the bar... | | |
| ▲ | hashta 3 hours ago | parent [-] | | It’s literally called "SimpleFold". But that’s not really my point, from your earlier comment (".. go through all the complexities first to find the generalized and simpler formulations"), I got the impression you thought the simplicity came purely from architectural insights. My point was just that to compare apples to apples, a model claiming "simpler but just as good" should ideally train on the same kind of data as AF or at least acknowledge very clearly that substantial amount of its training data comes from AF. I’m not trying to knock the work, I think it’s genuinely cool and a great engineering result. I just wanted to flag that nuance for readers who might not have the time or background to spot it, and I get that part of the "simple/simpler" messaging is also about attracting attention which clearly worked! | | |
| ▲ | godelski 2 hours ago | parent [-] | | > I got the impression you thought the simplicity came purely from architectural insights.
I'm unsure where I indicated that, but apologize for the confusion. I was initially pushing back against your original criticism of something like Alphafold having needed to be built first.Like you suggest, simple can mean many things. I think it's clear that in this context they mean "simple" (not from an absolute sense) in terms of the architectural design. I think the abstract is more than sufficient to convey this. > My point was just that to compare apples to apples
As a ML researcher who does a lot of work on architecture and efficiency, I think they are. Consider this from the end of the abstract | SimpleFold shows efficiency in deployment and inference on consumer-level hardware.
To me they are clearly stating that their goal isn't to get the top score on a benchmark. Their appendix shows that the 100M param is apples to apples to alphafold2 by size but not by compute. Even their 3B model uses less compute then alphafold2.So being someone in a neighboring niche, I don't understand your claim. There's no easy way to make your comparisons "apples to apples" because we shouldn't be evaluating on a single metric. Sure, alphafold2 gives better results on the benchmarks but does that mean people wouldn't sacrifice performance for a 450x reduction in compute? (20x for their largest model. But note that compute, not memory). > messaging is also about attracting attention
Yeah this is an unfortunate thing and I'm incredibly frustrated with this in academia and especially in ML. But it's also why I'm pushing against you. The problem stems from needing to get people to read your paper. There's a perverse incentive because you could have a paper that is groundbreaking but ends up having little to no impact because it didn't get read. A common occurrence is that less innovative papers will get magnitudes more citations by using similar methods but scale and beat benchmarks. So unfortunately as long as we use citation metrics as a significant measure of our research impact then marketing will be necessary. A catchy title is a good way to get more eyeballs. But I think you're being too nitpicky here and there's far more egregious/problematic examples. I'm not going to pick my fight with a title when the abstract is sufficiently clear. Could it be more clear? Certainly. But if the title is all that's wrong then it's a pretty petty problem. Especially if it's only confusing people who are significantly outside the target audience.Seriously, what's the alternative? That researchers write to the general public? To the general technical public? I'm sorry, I don't think that's a good solution. It's already difficult to communicate to people in the same domain (but not niche) in the page limit. It's hard to be them to read everything as it is. I'd rather papers be written strongly for the niche peers and enough generalization that domain experts can get through it with effort. For the general public, that's what science communicators are for |
|
|
|
| |
| ▲ | mapmeld 7 hours ago | parent | prev [-] | | And AlphaFold was validated with experimental observation of folded proteins using X-rays |
|
|
| ▲ | stephenpontes 8 hours ago | parent | prev | next [-] |
| I remember first hearing about protein folding with the Folding @Home project (https://foldingathome.org) back when I had a spare media server and energy was cheap (free) in my college dorm. I'm not knowledgable on this, but have we come a long way in terms of making protein folding simpler on today's hardware, or is this only applicable to certain types of problems? It seems like the Folding @Home project is still around! |
| |
| ▲ | roughly 7 hours ago | parent | next [-] | | As I understand it, folding at home was a physics based simulation solver, whereas alphafold and its progeny (including this) are statistical methods. The statistical methods are much, much cheaper computationally, but rely on existing protein folds and can’t generate strong predictions for proteins that don’t have some similarities to proteins in their training set. In other words, it’s a different approach that trades off versatility for speed, but that trade off is significant enough to make it viable to generate protein folds for really any protein you’re interested in - it moves folding from something that’s almost computationally infeasible for most projects to something that you can just do for any protein as part of a normal workflow. | | |
| ▲ | cowsandmilk 2 hours ago | parent [-] | | 1. I would be hesitant to not categorize folding@home as statistics based; they use Markov state models which is very much based on statistics. And their current force fields are parameterized via machine learning ( https://pubs.acs.org/doi/10.1021/acs.jctc.0c00355 ). 2. The biggest difference between folding@home and alphafold is that folding@home tries to generate the full folding trajectory while alphafold is just protein structure prediction; only looking to match the folded crystal structure. Folding@home can do things like look into how a mutation may make a protein take longer to fold or be more or less stable in its folded state. Alphafold doesn’t try to do that. | | |
| ▲ | roughly an hour ago | parent [-] | | You’re right, that’s true - I’d glossed over the folding@ methodology a bit. I think the core distinction is still that Folding is trying to divine the fold via simulation, while Alphafold is playing closer to a gpt-style predictor relying on training data. I actually really like Alphafold because of that - the core recognition that an amino acid string’s relationship to the structure and function of the protein was akin to the cross-interactions of words in a paragraph to the overall meaning of the excerpt is one of those beautiful revelations that come along only so often and are typically marked by leaps like what Alphafold was for the field. The technique has a lot of limitations, but it’s the kind of field cross-pollination that always generates the most interesting new developments. |
|
| |
| ▲ | _joel 8 hours ago | parent | prev | next [-] | | Yep, that and SETI@Home. I loved the eye candy, even if I didn't know what it fully meant. | | | |
| ▲ | jffry 4 hours ago | parent | prev | next [-] | | Apparently from a F@H blog post [1] they say it's still useful to know the dynamics of how it folded, in addition to the final folded shape. And that having ML-folded proteins is a rich target for simulation to validate and to understand how the protein works [1] https://foldingathome.org/2024/05/02/alphafold-opens-new-opp... | |
| ▲ | EasyMark 4 hours ago | parent | prev | next [-] | | They're still going and have made some great discoveries over the years. https://foldingathome.org/papers-results/?lng=en | |
| ▲ | ge96 6 hours ago | parent | prev | next [-] | | I contributed a lot on there too used my 3080Ti-FE as a small heater in the winter | | |
| ▲ | EasyMark 4 hours ago | parent [-] | | lol I still run it in the winter but I feel bad running it in the summer, so I don't run it when A/C or heating is not necessary. I figure some contribution is infinitely more than 0 contribution. |
| |
| ▲ | nkjoep 8 hours ago | parent | prev [-] | | Team F@H forever! |
|
|
| ▲ | tzumby 37 minutes ago | parent | prev | next [-] |
| Flow-matching, the technique they describe is incredibly interesting. I studied it in the context of generative AI and found it fascinating. It’s so fitting that a technique that borrows from thermodynamics and uses Brownian motion would go full circle to solve for protein folding. |
|
| ▲ | vbarrielle 8 hours ago | parent | prev | next [-] |
| A paper that says: "our approach is simpler than the state of the art". But also does not loudly say "our approach is significantly behind the state of the art on all metrics". Not easy to get published, but I guess putting it as a preprint with a big company's name will help... |
|
| ▲ | IAmBroom 8 hours ago | parent | prev | next [-] |
| Link goes the github repository behind the article you might want to read. https://arxiv.org/abs/2509.18480 |
| |
|
| ▲ | barbarr 9 hours ago | parent | prev | next [-] |
| Why is apple doing protein folding? |
| |
| ▲ | robotresearcher 7 hours ago | parent | next [-] | | Apple has an ML research group. They do a mixture of obviously-Apple things, other applications, generally useful optimizations, and basic research. https://machinelearning.apple.com/ | |
| ▲ | giancarlostoro 8 hours ago | parent | prev | next [-] | | No idea, but can I be signed up for R&D jobs where you don't necessarily build something generating revenue? Maybe these are just projects they use to test and polish their AI chips? Not sure. | |
| ▲ | cowsandmilk 3 hours ago | parent | prev | next [-] | | To sell computers? 20 years ago, Apple had scientific poster sessions at WWDC and worked to bring PyMol to the Mac. The pictures of proteins you see in the paper were generated with PyMol as are probably >50% of the protein images in scientific papers for the last 15 years. | |
| ▲ | nextos 8 hours ago | parent | prev | next [-] | | Local inference. I imagine they have an interest in making this and other cutting edge models small enough to be possible to do quick inference on their desktop machines. The article shows that, with Figure 1E demonstrating inference on an M2 Max 64 GB. Frankly, it's a great idea. If you are a small pharma company, being able to do quick local inference removes lots of barriers and gatekeeping. You can even afford to do some Bayesian optimization or RL with lab feedback on some generated sequences. In comparison, running AlphaFold requires significant resources. And IMHO, their usage of multiple alignments is a bit hacky, makes performance worse on proteins without close homologs, and requires tons of preprocessing. A few years back, ESM from Meta already demonstrated that alignment-free approaches are possible and perform well. AlphaFold has no secret sauce, it's just a seq2seq problem, and many different approaches work well, including attention-free SSMs. | | |
| ▲ | Zacharias030 2 hours ago | parent | next [-] | | I think people often interpret a bit too much.
Perhaps it’s just some researchers who got enough freedom to run and publish interesting work within apple. For a company like apple it makes sense to have a research lab with considerable freedoms even if protein folding is not a core interest, which is why you see it published but not the formula for the new Corning Gorilla glass… | |
| ▲ | mensetmanusman 34 minutes ago | parent | prev [-] | | Will be fascinating to see how the market breaks down in the future, will enough people want a third best model they can run on prem, or will people all be fighting in line for the top models that are a few cents more per token on supercomputers. |
| |
| ▲ | bobmarleybiceps 6 hours ago | parent | prev | next [-] | | This may not be the actual reason in this case, but I think it's good to be aware of: A non-zero chunk of "ai for science" research done at tech companies is basically done for marketing. Even in cases where it's not directly beneficial for the companies products or is unlikely to really lead to anything substantial, it is still good for "prestige" | |
| ▲ | Forbo 8 hours ago | parent | prev | next [-] | | Reputation laundering? | | |
| ▲ | EasyMark 4 hours ago | parent | next [-] | | They have a much better reputation that most companies. I think they're doing okay compared to google, facebook, oracle, etc. Few people are going to think a corp is "doing good" but reputation does still matter somewhat. | | |
| ▲ | leptons 4 hours ago | parent [-] | | If more people read the cases against Apple by the DOJ & the EU, they probably wouldn't have such a high opinion of Apple. |
| |
| ▲ | jama211 8 hours ago | parent | prev [-] | | What’s there to launder? Perhaps they shouldn’t have as good a reputation as they do, but you can’t deny they do have a good reputation. | | |
| ▲ | amelius 8 hours ago | parent [-] | | Reputation of what? They are just an office appliance company. | | |
| ▲ | axoltl 7 hours ago | parent | next [-] | | You're confusing your opinion of the company with the perception by the general public. Apple's definitely not perceived as 'an office appliance company' by your average person. It's considered a high-end luxury brand by many[1]. 1: https://www.researchgate.net/publication/361238549_Consumer_... | |
| ▲ | robotresearcher 7 hours ago | parent | prev [-] | | I think their public sales data shows Apple sells mainly to consumers, and mainly iPhones at that. Like 1980s SONY, they are the top of the line consumer electronics giant of the time. The iPhone is even more successful than the Walkman or Trinitron TVs. They also sell the most popular laptops,to consumers as well as corporate. Like SONY’s VAIO but more popular again. | | |
| ▲ | robotresearcher 5 hours ago | parent [-] | | The move in consumer electronics leadership from Japan to the US, Korea, and now China is probably pretty interesting to understand. Can anyone recommend a good book or article about this? |
|
|
|
| |
| ▲ | shpongled 8 hours ago | parent | prev | next [-] | | Probably because ByteDance and Facebook (spun out into EvolutionaryScale) are doing it | |
| ▲ | IncreasePosts 8 hours ago | parent | prev | next [-] | | They're jealous they haven't won a Nobel prize | |
| ▲ | mabedan 9 hours ago | parent | prev | next [-] | | Prowlly cuz Siri didn’t work out | |
| ▲ | lovasoa 8 hours ago | parent | prev [-] | | How do you call the opposite of green washing? When you want to show that you are burning as much energy on training models as the others. |
|
|
| ▲ | foodevl 8 hours ago | parent | prev | next [-] |
| I was curious what the protein picture was showing:
"Figure 1 Example predictions of SimpleFold on targets ... with ground truth shown in light aqua and prediction in deep teal." and now I'm even more curious why they thought "light aqua" vs "deep teal" would be a good choice |
| |
| ▲ | gilleain 8 hours ago | parent [-] | | Well, figure a) shows a ribbon representation of the fold (as helices and strands) of the protein 7QSW (https://www.ebi.ac.uk/pdbe/entry/pdb/7qsw) which is RubisCO (https://en.wikipedia.org/wiki/RuBisCO), an plant protein that plays a key role in photosynthesis. The different colours are for the predicted and 'real' (ground truth) models. The fact that it is hard to distinguish is partly the - as you point out - weird colour choice, but also because they are so close together. An inaccurate prediction would have parts that stand out more as they would not align well in 3D space. |
|
|
| ▲ | shpongled 8 hours ago | parent | prev | next [-] |
| It's not totally novel, but it's very cool to see the continued simplification of protein folding models - AF2 -> AF3 was a reduction in model architecture complexity, and this is a another step in the direction of the bitter lesson. |
| |
| ▲ | hashta 7 hours ago | parent [-] | | I’m not sure AF3’s performance would hold up if it hadn’t been trained on data from AF2 which itself bakes in a lot of inductive bias like equivariance |
|
|
| ▲ | frenchie4111 8 hours ago | parent | prev | next [-] |
| I am curious to hear an expert weigh in on this approach's implications for protein folding research. This sounds cool but it's really unclear to me what the implications are |
| |
| ▲ | geremiiah 8 hours ago | parent | next [-] | | Their representation is simpler, just a transformer. That means you can just plug in all the theory and tools that have been developed specifically for transformers, most importantly you can scale the model easier. But more than that, I think, it shows that there was no magic to AlphaFold. The details of the architecture and training method didn't matter much. All that was needed was training a big enough model on a large enough dataset. Indeed lots of people who have experimented with AlphaFold have found it to behave similiar to LLMs, i.e. it performs well on inputs close to the training dataset and but it doesn't generalize well at all. | | |
| ▲ | johncolanduoni 8 minutes ago | parent | next [-] | | Except their dataset is mostly the output of AlphaFold, which had to use the much smaller dataset of proteins analyzed by crystallography as input. This is really an exercise in model distillation - a worthy endeavor but it's not like they could have just taken their architecture and the dataset AlphaFold had and expect to get the same results. If that was the case, that's what they would have done because it would've been much more impressive. | |
| ▲ | aDyslecticCrow 5 hours ago | parent | prev | next [-] | | I think the sentiment that simplicity is good, is a false conclusion. Simplicity is simply good scientific methodology. Doing too many things at once makes methods hard to adopt and makes conclusions harder to draw. So we try to find simple methods that show measurable gain, so we can adapt it to future approaches. Its a cycle between complexity and simplicity. When a new simple and scalable approach beats the previous state of art, that just means we discovered a new local maxima hill to climp up. | |
| ▲ | visarga 6 hours ago | parent | prev [-] | | > But more than that, I think, it shows that there was no magic to AlphaFold. The details of the architecture and training method didn't matter much. All that was needed was training a big enough model on a large enough dataset. People often like to say that we just need one more algorithmic breakthrough or two for AGI. But in reality it's the dataset and the environment based learning. Almost any model would do if you collected the data. It's not in the model, it's outside where we need to work on. |
| |
| ▲ | epistasis 7 hours ago | parent | prev [-] | | It may be a change in future models, perhaps. Here's one person's opinion: https://genomely.substack.com/p/simplefold-and-the-future-of... But as with anything in research, it will take months and years to see what the actual implications are. Predictions of future directions can only go so far! |
|
|
| ▲ | phoenicyan 6 hours ago | parent | prev | next [-] |
| Curious since AlphaFold got released: have classical molecular dynamics sims in this area become obsolete, at least for protein folding? How does the research coming out of venues like DESRES compare? Are they working on more specific problems in the same area or are they in a different business altogether? |
| |
| ▲ | the__alchemist 6 hours ago | parent | next [-] | | No. AlphaFold doesn't do dynamics; it does end-state snapshots only. It does not do anything about the motion of the atoms, which is the core functionality of MD. | |
| ▲ | tripplyons 6 hours ago | parent | prev | next [-] | | I was curious about what was released, and the parameters for AlphaFold V3 are only given to certain groups for non-commercial use: https://github.com/google-deepmind/alphafold3?tab=readme-ov-... However, it seems like anyone can download the parameters for AlphaFold V2: https://github.com/google-deepmind/alphafold?tab=readme-ov-f... | |
| ▲ | dekhn 6 hours ago | parent | prev [-] | | MD was never really a viable way to do structure prediction, so it didn't become obsolete with AlphaFold. Instead, MD is more useful for studying the physical process of protein folding (before the protein folds to its final structure, as well as once it has reached its final structure and sort of jiggles and wiggles around that). | | |
| ▲ | cowsandmilk 4 hours ago | parent [-] | | MD simulations typically aren’t run for time scales that tell you anything about the folding process. Most people are looking at motion after the protein has folded. |
|
|
|
| ▲ | GistNoesis 4 hours ago | parent | prev | next [-] |
| Intellectually, I don't like this approach. Predicting the end-result from the sequence of protein directly is prone to miss any new phenomenon and would just regurgitate/interpolate the training datasets. I would much prefer an approach based on first principles. In theory folding is easy, it's just running a simulation of your protein surrounded by some water molecules for the same number of nano-seconds nature do. The problem is that usually this take a long time because evolving a system needs to compute the energy of the system as a position of the atoms which is a complex problem involving Quantum Mechanics. It's mostly due to the behavior of the electrons, but because they are much lighter they operate on a faster timescale. You typically don't care about them, only the effect they have on your atoms. In the past, you would use various Lennard-Jones potentials for pairs of atoms when the pair of atoms are unbounded, and other potentials when they are bonded and it would get very complex very quickly. But now there are deep-learning based approach to compute the energy of the system by using a neural network. (See (Gromacs) Neural Network Potentials https://rowansci.com/publications/introduction-to-nnps ). So you train these networks so that they learn the local interactions between atoms based on trajectories generated from ab-initio theories. This allows you to have a faster simulator which approximate the more complex physics. It's in a sort just tabulating using a neural network the effect of the electrons would have in a specific atom arrangements according to the theory you have chosen. At any time if you have some doubt, you can always run the slower simulator in the small local neighborhood to check that the effective field neural network approximation holds. Only then once you have your simulator which is able to fold, you can generate some dataset of pairs "sequence of protein" to "end of trajectory", to learn the shortcut like Alpha/Simple/Fold do. And when in doubt you can go back to the slower more precise method. If you had enough data and can train perfectly a model with sufficient representation power, you could theoretically infer the correct physics just from the correspondence initial to final arrangements. But if you don't have enough data it will just learn some shortcut and accept that it will be wrong some times. |
| |
| ▲ | slashdave 4 hours ago | parent [-] | | > it's just running a simulation of your protein surrounded by some water molecules for the same number of nano-seconds nature do. No, the environment is important. Also, some proteins fold while being sequenced. Folding can also take minutes in some cases, which is the real problem. > which is a complex problem involving Quantum Mechanics Most MD simulations use classical approximations, and I don't see why folding is any different. |
|
|
| ▲ | kylehotchkiss 9 hours ago | parent | prev | next [-] |
| > Folding Proteins Is Simpler Than You Think Then why do we need customized LLM models, two of which seemed to require the resources of 2 of the wealthiest companies on earth (this and google's alphafold) to do it? |
| |
| ▲ | aDyslecticCrow 8 hours ago | parent | next [-] | | Its not an LLM, It's a transformer. I know the terms are really being butchered in media, but if we're gonna use the term LLM instead of AI, we better make sure it's actually a "large language model" that is being refereed to. If you're unsure, call it a neural net, or machine learning algorithm, or AI. It's indeed a large model. But if you knew the history of the field, it's a massive improvement. It has progressed from a almost "NP" problem only barely approachable with distributed cluster compute, to something that can run on a single server with some pricey hardware. The smallest model is only here is only 100M parameters and the largest is 3B parameters, that's very approachable to run locally with the right hardware, and easily within the range for a small biotech lab (compared to the cost of other biotech equipment) It's also (i'd argue) one of the only truly economically and sociably valuable AI technologies we've found over the past few years. Every simulated protein fold is saving a biotech company weeks of work for highly skilled biotech engineers and very expensive chemicals (In a way that that truly only supplement rather than replace the work). Any progress in the field is a huge win for society. | | |
| ▲ | kylehotchkiss 7 hours ago | parent [-] | | I'm more teasing the title than the tech :) I'm all for innovation in the field especially with so much bio funding cut! |
| |
| ▲ | wrsh07 8 hours ago | parent | prev | next [-] | | Folding proteins is pretty valuable and this model is comparably small This doesn't seem like particularly wasteful overinvestment. Granted, I'm more excited about the research coming out of arc | | | |
| ▲ | wrs 8 hours ago | parent | prev [-] | | How simple did you think it was before? | | |
| ▲ | kylehotchkiss 8 hours ago | parent [-] | | Not simple! Wasn't/Isn't X-ray crystallography what it usually takes to determine the structure? |
|
|
|
| ▲ | ziofill 5 hours ago | parent | prev | next [-] |
| In the plots in Fig. 4 it looks like they should have continued the training because the performance was still climbing, am I reading it incorrectly? |
|
| ▲ | kazinator 8 hours ago | parent | prev | next [-] |
| I'm satisfied with with folding roast beef onto a sandwich, or folding egg whites into batter. All the protein folding action I could ever want. |
|
| ▲ | turblety 8 hours ago | parent | prev | next [-] |
| I wonder why Apple can create a model to fold proteins, but still can't get Siri to control the phone competently? I'm not sure I agree with Apple's priorities. I guess these things are not synchronous and they can work on multiple things at a time. |
| |
| ▲ | tanelpoder 8 hours ago | parent | next [-] | | I guess it's because SimpleFold came from a research lab with different autonomy and less competing interests and internal politics... | |
| ▲ | frenchie4111 8 hours ago | parent | prev | next [-] | | I am genuinely interested where the strong negativity towards Siri has come from in recent culture. From what I gather it's likely due to the high expectations we have for Apple. But what I don't really get is why is there not a similar amount of negativity being directed at Google or Samsung, who both have equally shit phone AI assistants (obviously this is just from my perspective, I am a daily user of both iOS and a Samsung Android) I am not trying to defend Apple or Siri by any means. I think the product absolutely should (and will) improve. I am just curious to explore why there is such negativity being directed specifically at Apple's AI assistant. | | |
| ▲ | xp84 8 hours ago | parent | next [-] | | As a vocal critic of Siri, I can give you a number of reasons we hate it: 1. It seems to be actively getting worse. On a daily basis, I see it responding to queries nonsensically, like when i say “play (song) by (artist)” (I have Apple Music) by opening my Sirius app and putting on a random thing that isn’t even that artist. Other trivial commands are frequently just met with apologies or searching the web. 2. Over a year ago Apple conducted a flashy announcement full of promises about how Siri would not only do the things that it’s been marketed as being able to do for the last decade, but also things that no one has seen an assistant do. Many people believe that announcement was based on fantasy thinking and those people are looking more and more correct every day that Apple ships no actual improvements to Siri. 3. Apple also shipped a visual overhaul of how Siri looks, which gives the impression that work has been done, leading people to be even more disappointed when Siri continues to be a pile of trash. 4. The only competitor that makes sense to compare is Google, since no one else has access to do useful things on your device with your data. At least Google has a clear path to an LLM-based assistant, since they’ve built an LLM. It seems believable that android users will have access to a Gemini-based assistant, whereas it appears to most of us that Apple‘s internal dysfunction has rendered them unable to ship something of that caliber. | |
| ▲ | samuelg123 8 hours ago | parent | prev | next [-] | | I think Siri has always been criticized, likely because it has never worked super well and it has the most eyes (or ears) on it (iPhones still have 50% market share in the US). And now that we have ChatGPT with voice mode, Gemini Live, etc which have incredible speech recognition and reasoning comparatively, it's harder to argue that "every voice assistant is bad" still. | |
| ▲ | citizenpaul 8 hours ago | parent | prev | next [-] | | Is it just my rosie glasses or did siri work much better in the first couple of years and seem to decline continually since then. I actually used it a lot initially then eventually disabled it as it never worked anymore. | | |
| ▲ | devmor 8 hours ago | parent [-] | | I feel like the same is true of a lot of products that moved from being programmatically connected ML workflows to multi-modal AI. We, the consumer, have received inferior products because of the vague promise that the company might one day be able to make it cheaper if they invest now. |
| |
| ▲ | SoftTalker 7 hours ago | parent | prev | next [-] | | I've disabled Siri as much as I possibly can. I've never even tried to use it. I would do the same for any other AI assistant. I don't like that they are always listening, and I just don't like talking to computers. I find it unnatural, and I get irrationally angry when they don't understand what I want. If I could buy a phone without an assistant I would see that as a desirable feature. | |
| ▲ | Invictus0 8 hours ago | parent | prev [-] | | For the last three iOS major versions, Siri has been unable to execute the simple command "shuffle the playlist 'Jams'", or any variation, like "play the playlist Jams on shuffle". I am upset for that reason. |
| |
| ▲ | al_borland 8 hours ago | parent | prev | next [-] | | Something like this doesn’t actually have to work. There were no expectations at all in this space. Meanwhile, people expect perfection from Siri. At this point a new version of Siri will never live up to people’s expectations. Had they released something on-par with ChatGPT, people would hate it and probably file a class action lawsuit against Apple over it. The entire company isn’t going to work on Siri. In a large company there are a lot of priorities, and some things that happen on the side as well. For all we know this was one person’s weekend project to help learn something new that will later be applied to the priorities. I’ve made plenty of hobby projects related to work that weren’t important or priorities, but what I learned along the want proved extremely valuable to key deliverables down the road. | |
| ▲ | mapmeld 7 hours ago | parent | prev | next [-] | | As I understand it, Siri and Alexa could be plugged into an LLM, but changing it to an "open world" device that can tell your kid something disturbing, text all of your contacts, buy groceries, etc. comes with serious risk of reputational harm. While still falling short of people's expectations if it isn't ChatGPT-quality. OpenAI is new enough that they get to play by different rules. | |
| ▲ | EasyMark 4 hours ago | parent | prev [-] | | Fair point, even X was able to pump out a usable AI, grok. |
|
|
| ▲ | barbazoo 8 hours ago | parent | prev | next [-] |
| No folding here. Proteins go on the hanger or in the drawer. |
|
| ▲ | nextworddev 8 hours ago | parent | prev | next [-] |
| In industry Google practically dominates this field |
|
| ▲ | dyauspitr 7 hours ago | parent | prev | next [-] |
| Isn’t this a largely solved problem after Alphafold? |
| |
| ▲ | the__alchemist 37 minutes ago | parent | next [-] | | Should an entry in a field preclude other ones. I encourage you to apply reductio-ad-absurdum here. Should Pepsi exist if Coke does? Should C exist if Fortran does? | |
| ▲ | samfriedman 7 hours ago | parent | prev [-] | | Maybe they've been working on it, but got scooped? | | |
| ▲ | zamadatix 6 hours ago | parent [-] | | I don't think that's the case. The numbers in the paper suggest ~92% of the training data comes from pre-existing AI models, including AlphaFold, and they claim things like: > We largely adopt the data pipeline implemented in Boltz-11
1https://github.com/jwohlwend/boltz (Wohlwend et al., 2024), which is an open-source replication of AlphaFold3 I believe the story here is largely that they simplified the architecture and scaled it to 3B parameters while maintaining leading results. |
|
|
|
| ▲ | wild_pointer 9 hours ago | parent | prev | next [-] |
| Did you just assume what I think about protein folding simplicity?! |
|
| ▲ | underdeserver 8 hours ago | parent | prev | next [-] |
| So, how does this compare to AlphaFold? |
| |
| ▲ | mentalgear 7 hours ago | parent [-] | | seems like they use the normal transformer architecture versus deep fold's more specialised machine-learning approaches. |
|
|
| ▲ | 331c8c71 8 hours ago | parent | prev | next [-] |
| It is for structure prediction, not folding (rolleyes). |
| |
|
| ▲ | Invictus0 8 hours ago | parent | prev [-] |
| They'll do anything but fix Siri |
| |
| ▲ | mentalgear 7 hours ago | parent [-] | | They can keep on doing stuff like this that's open-source and beneficial to society. |
|