▲ | visarga 3 days ago | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
AI is used by students, teachers, researchers, software developers, marketers and other categories and the adoption rates are close to 90%. Even if it does not make us more productive we still like using it daily. But when used right, it does make us slightly more productive and I think it justifies its cost. So yes, in the long run it will be viable, we both like using it and it helps us work better. But I think the benefits of AI usage will accumulate with the person doing the prompting and their employers. Every AI usage is contextualized, every benefit or loss is also manifested in the local context of usage. Not at the AI provider. If I take a photo of my skin sore and put it on ChatGPT for advice, it is not OpenAI that is going to get its skin cured. They get a few cents per million tokens. So the AI providers are just utilities, benefits depend on who sets the prompts and and how skillfully they do it. Risks also go to the user, OpenAI assumes no liability. Users are like investors - they take on the cost, and support the outcomes, good or bad. AI company is like an employee, they don't really share in the profit, only get a fixed salary for work | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | okamiueru 3 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
I think that AI is a benefit for about 1% of what people think it is good for. The remaining 99% had become a significant challenge to the greatest human achievement in distribution of knowledge. If people used LLMs, knowing that all output is statistical garbage made to seem plausible (i.e. "hallusinations"), and that it just sometimes overlaps with reality, it would be a lot less dangerous. There is not a single case of using LLMs that has lead to a news story, that isn't handily explained by conflating a BS-generator with Fact-machine. Does this sound like I'm saying LLMs are bad? Well, in every single case where you need factual information, it's not only bad, it's dangerous and likely irresponsible. But there are a lot of great uses when you don't need facts, or by simply knowing it isn't producing facts, makes it useful. In most of these cases, you know the facts yourself, and the LLM is making the draft, the mundane statistically inferable glue/structure. So, what are these cases? - Directing attention in chaos: Suggest where focus needs attention from a human expert. (useful in a lot of areas, medicine, software development). - Media content: music, audio (fx, speech), 3d/2d art and assets and operations. - Text processing: drafting, contextual transformation, etc Don't trust AI if the mushroom you picked is safe to eat. But use its 100% confident sounding answer for which mushroom it is, as a starting point to look up the information. Just make sure that the book about mushrooms was written before LLMs took off.... | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | grues-dinner 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
> AI is used by students, teachers, researchers, software developers, marketers and other categories and the adoption rates are close to 90%. Even if it does not make us more productive we still like using it daily. Nearly everyone uses pens daily but almost no one really cares about them or says their company runs using pens. You might grumble when the pens that work keeps in the stationary cupboard are shit, perhaps. I imagine eventually "AI" services will be commoditised in the same way that pens are now. Loads of functional but faily low-quality stuff, some fairly nice but affordable stuff and some stratospheric gold plated bricks for the military and enthusiasts. In the middle is a large ecosystem of ink manufacturers, lathe makers, laser engravers, packaging companies and logistics and so on and on that are involved. The explosive, exponential winner-takes-all scenario where OpenAI and it's investors literally ascend to godhood and the rest of humanity lives forever under their divine bootheels doesn't seem to be the trajectory we're on. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | Tepix 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
We also know from studies that it makes us less capable, i.e. it rots our brains. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | sumanthvepa 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
This. Right now the consumer surplus created by improved productivity is being captured by users and to a small extent their employers. But that may not remain the case in future. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | RataNova 3 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||
Feels like we're shifting into a world where “AI fluency” becomes a core part of individual economic agency, more like financial literacy than software adoption |