| ▲ | LunaSea 4 hours ago |
| Unserious answer about a very serious event. I don't believe a word of Sam's "I believe" section. |
|
| ▲ | SOLAR_FIELDS 3 hours ago | parent | next [-] |
| Ha, I was giving an AI bootcamp to a room full of people and someone asked me my opinion of Altman. I hesitated for a second and replied that I would not trust Altman further than I could throw a rock about anything. If Graham says this guy will always stop at nothing to get whatever he wants, which I absolutely believe, then why would you trust anything that comes out of a person like that’s mouth? |
| |
| ▲ | dakolli 3 hours ago | parent [-] | | Who tf is dumb enough to pay for an AI bootcamp, genuinely curious. If you're selling AI bootcamps, or whoever is, they are just as much a scam artist as Sam. | | |
| ▲ | teaearlgraycold an hour ago | parent | next [-] | | You don’t even know what is covered. It could be anything from how to prompt to how to create your own models from numpy primitives. | |
| ▲ | moralestapia 3 hours ago | parent | prev | next [-] | | Who tf is dumb enough to not do it, though? If I was non-tech and owned a business, and someone (reputable) offers to teach me everything I need to get up to date with the most revolutionary technology of the decade (perhaps century?) for like ... 500 dollars? Why not? | | |
| ▲ | dakolli 3 hours ago | parent [-] | | Its neural network autocomplete that helps you write text a little faster, chill with "the most revolutionary technology of the last decade/century" talk. You're offending a lot of experts in way more important areas of research. | | |
| ▲ | snoman 21 minutes ago | parent | next [-] | | That’s so shockingly ignorant/reductive that you shouldn’t be surprised when people start ignoring you in technical conversations. | |
| ▲ | xvector 2 hours ago | parent | prev | next [-] | | You're cooked if this is actually how you see AI in 2026. | |
| ▲ | moralestapia 3 hours ago | parent | prev [-] | | >write text a little faster You might actually need to attend an AI bootcamp. This is not 2022's GPT, AI can deliver plenty of value for a business owner these days. | | |
|
| |
| ▲ | hungryhobbit 3 hours ago | parent | prev [-] | | Yeah, people learning new technology is terrible. /s |
|
|
|
| ▲ | probably_wrong 3 hours ago | parent | prev | next [-] |
| 10 hours ago a post made the frontpage here [0] about how OpenAI is backing a law that "would limit liability for AI-enabled mass deaths or financial disasters". Now he's here saying he believes that "working towards prosperity for everyone, empowering all people, and advancing science and technology are moral obligations for [him]". I know he doesn't believe a word of what he wrote in that post except, perhaps, that he cannot sleep and is pissed. I know I should be used to people openly lying with no consequence, but it still amazes me a bit. [0] https://news.ycombinator.com/item?id=47717587 |
| |
| ▲ | scruple 2 minutes ago | parent | next [-] | | > I know I should be used to people openly lying with no consequence, but it still amazes me a bit. Well that makes two of us. Character seems to mean nothing today. | |
| ▲ | SpicyLemonZest 3 hours ago | parent | prev | next [-] | | I think it's good for CEOs of powerful companies to make statements about how they don't want too much personal power and it's important to ensure everyone does well, even and perhaps especially if there's reason to suspect they don't believe it. Saying it doesn't solve the problem, but it helps create a permission structure for the rest of us to get it to actually happen. | | |
| ▲ | tyre 3 hours ago | parent [-] | | The reason he's saying that is because he doesn't want you to create that structure. He wants you to not create the laws or checks & balances on him because you "trust that he doesn't really want the power". It has worked for him, repeatedly. | | |
| ▲ | SpicyLemonZest 3 hours ago | parent [-] | | No, I don't think that's accurate. Altman has repeatedly and loudly demanded for these to be created, including a new detailed policy proposal just this month (https://cdn.openai.com/pdf/561e7512-253e-424b-9734-ef4098440...). | | |
| ▲ | tyre 3 hours ago | parent [-] | | OpenAI has also repeatedly and quietly lobbied against them. You linked a vague PDF whose promised actions are: > To help sustain momentum, OpenAI is: (1) welcoming and organizing feedback through newindustrialpolicy@openai.com; (2) establishing a pilot program of fellowships and focused research grants of up to $100,000 and up to $1 million in API credits for work that builds on these and related policy ideas; and (3) convening discussions at our new OpenAI Workshop opening in May in Washington, DC. Welcoming and organizing feedback! A pilot! Convening discussions! This "commitment" pales in comparison to the money they've spent lobbying against specific regulation that cedes power. Please don't fall for this stuff. |
|
|
| |
| ▲ | 0xy 2 hours ago | parent | prev [-] | | Incendiary and false headline aside, no sane person would suggest that a hardware store that sold an axe that was used by an axe murderer should be held liable unless that store knew what was about to unfold. Unless AI companies knowingly participate in murder plots, they should not be liable. Is Microsoft liable for providing Notepad, a product which can be used to write detailed and specific mass murder plots? Is Toyota liable for selling someone a car that is later used for vehicular manslaughter? Liability should depend on your participation in the event, of course. Otherwise you wouldn't be able to buy an axe, or a car, or use the internet at all. A closer analogy is ISPs not being liable for copyright infringement done by users, and subsequently not being required to police such activity for rights holders. | | |
| ▲ | probably_wrong 2 hours ago | parent | next [-] | | > Incendiary and false headline aside The text of the bill literally starts with "Creates the A.I. Safety Act. Provides that a
developer of a frontier AI model shall not be held liable for critical harms caused by the frontier model if (conditions)", and defines "critical harms" as "death or serious injury of 100 or more people or at least $1,000,000,000 of damages". The headline is, IMO, shockingly accurate. > Is Toyota liable for selling someone a car that is later used for vehicular manslaughter? No, but they are liable for selling a car with defective brakes, even if they don't know that the brakes are defective. And if the ex-Monsanto has to pay millions in compensation for causing cancer with a product that they tested to hell and back, then I don't see how that's different when the one causing cancer is an AI just because the developers pinky swear that it's safe. | |
| ▲ | saintfire 42 minutes ago | parent | prev [-] | | People championing the absolution of billionaires who create a chatbot that can't spell strawberry who then say it should be allowed to choose who lives and dies wasn't what I expected at the turn of the decade. Beautiful. |
|
|
|
| ▲ | mixtureoftakes 4 hours ago | parent | prev [-] |
| unpopular opinion but i think it's written quite well |
| |
| ▲ | ryan_n 3 hours ago | parent | next [-] | | I don't think that's unpopular, it is pretty well written. But the "I believe" section is extraordinarily hard to believe given Altman's history. > Working towards prosperity for everyone, empowering all people > We have to get safety right > AI has to be democratized; power cannot be too concentrated None of these statements, IMO, reflect his actions over the past 5 years. > we urgently need a society-wide response to be resilient to new threats. This includes things like new policy to help navigate through a difficult economic transition in order to get to a much better future I agree with this, but there is a near 0% chance of that happening anytime soon in the US. I think he probably is aware of this. Just my opinion, but it comes off as very insincere. To be clear, what happened is still awful and there's absolutely no justification for it. | |
| ▲ | daseiner1 33 minutes ago | parent | prev | next [-] | | it's "written well" but not at all a smart piece of writing. leading with a photo of a cute baby before engaging in an extended defense of one's own integrity is so obvious as to be insulting | |
| ▲ | kcatskcolbdi 3 hours ago | parent | prev | next [-] | | Yes, clearly not written with his own product. | | |
| ▲ | pesus 3 hours ago | parent [-] | | If that's the case, why doesn't he trust his own product enough to write this? | | |
| ▲ | alpaca128 2 hours ago | parent [-] | | He doesn't trust it for anything else either as far as I can tell. In an interview he's boasted about how he uses a paper notebook for everything all day. |
|
| |
| ▲ | kspacewalk2 4 hours ago | parent | prev [-] | | Perhaps by ChatGPT | | |
|