| ▲ | citizenpaul a day ago | |||||||||||||||||||||||||||||||
Its really a high level bikeshed. Obviously we are all still using and experimenting with LLM's. However there is a huge gap of experiences and total usefulness depending on the exact task. The majority of HN's still reach for LLM's pretty regularly even if they fail horribly frequently. Thats really the pit the tech is stuck in. Sometimes it oneshots your answer perfectly, or pair programs with you perfectly for one task, or notices a bug you didn't. Sometimes it wastes hours of your time for various subtle reasons. Sometimes it adamantly insists 2 + 2 = 55 | ||||||||||||||||||||||||||||||||
| ▲ | nfw2 a day ago | parent [-] | |||||||||||||||||||||||||||||||
Latest reasoning models don't claim 2 + 2 = 55, and it's hard to find them making an sort of obviously false claims, or not admitting to being mistaken if you point out that they are | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||