| ▲ | rsynnott 8 hours ago | |
> How is a chatbot supposed to determine when a user fools even themselves about what they have experienced? And even if it _could_, note, from the article: > Overall, the participants deemed sycophantic responses more trustworthy and indicated they were more likely to return to the sycophant AI for similar questions, the researchers found. The vendors have a perverse incentive here; even if they _could_ fix it, they'd lose money by doing so. | ||