| ▲ | tantalor 4 hours ago | ||||||||||||||||
The human baseline seems flawed. 1. There is no initial screening that would filter out garbage responses. For example, users who just pick the first answer. 2. They don't ask for reasoning/rationale. | |||||||||||||||||
| ▲ | slongfield 4 hours ago | parent | next [-] | ||||||||||||||||
My favorite example of this was the Pew Research study: https://www.pewresearch.org/short-reads/2024/03/05/online-op... They found that ~15% of US adults under 30 claim to have been trained to operate a nuclear submarine. | |||||||||||||||||
| ▲ | mwigdahl 4 hours ago | parent | prev | next [-] | ||||||||||||||||
Lizardman's Constant is famously 4%. https://en.wikipedia.org/wiki/Slate_Star_Codex#Lizardman's_C... | |||||||||||||||||
| ▲ | felix089 4 hours ago | parent | prev | next [-] | ||||||||||||||||
RE 1, they actually do have a pre-screening screening of the participants in general, you can check how they do it in detail: https://www.rapidata.ai/ | |||||||||||||||||
| |||||||||||||||||
| ▲ | andreasgl 4 hours ago | parent | prev [-] | ||||||||||||||||
I agree. I wonder what the human baseline is for ”what is 1 + 1” on Rapidata. | |||||||||||||||||
| |||||||||||||||||