| ▲ | hmokiguess 11 hours ago | |
I always think about this, can someone with more knowledge than me help me understand the fragility of these operations? It sounds like the value of these very time-consuming, resource-intensive, and large scale operations is entirely self-contained in the weights produced at the end, right? Given that we have a lot of other players enabling this in other ways, like Open Sourcing weights (West vs East AI race), and even leaks, this play by Apple sounds really smart and the only opportunity window they are giving away here is "first to market" right? Is it safe to assume that eventually the weights will be out in the open for everyone? | ||
| ▲ | bayarearefugee 8 hours ago | parent | next [-] | |
> and the only opportunity window they are giving away here is "first to market" right? A lot of the hype in LLM economics is driven by speculation that eventually training these LLMs is going to lead to AGI and the first to get there will reap huge benefits. So if you believe that, being "first to market" is a pretty big deal. But in the real world there's no reason to believe LLMs lead to AGI, and given the fairly lock-step nature of the competition, there's also not really a reason to believe that even if LLMs did somehow lead to AGI that the same result wouldn't be achieved by everyone currently building "State of the Art" models at roughly the same time (like within days/months of each other). So... yeah, what Apple is doing is actually pretty smart, and I'm not particularly an Apple fan. | ||
| ▲ | pests 8 hours ago | parent | prev [-] | |
> is entirely self-contained in the weights produced at the end, right? Yes, and the knowledge gained along the way. For example, the new TPUv4 that Google uses requires rack and DC aware technologies (like optical switching fabric) for them to even work at all. The weights are important, and there is open weights, but only Google and the like are getting the experience and SOTA tech needed to operate cheaply at scale. | ||