▲ | cma 6 days ago | |||||||
I would posit that most randomly selected AGIs (people) you ask this of with no tools allowed (pencil and paper) won't get closer on average. This doesn't prove people don't use logic And allowing python shells for both I think the randomly selected human would do worse on average. And I also think e.g. Google's IMO model would be able to pass but I have no way to verify that. | ||||||||
▲ | pron 6 days ago | parent | next [-] | |||||||
At work, you can trust people to either get the right answer or tell you they may not have the right answer. If someone is not trustworthy, you don't work with them again. The experience of trying to work with something that is completely not trustworthy on all fronts is novel and entirely dissimilar to working with either people or tools. | ||||||||
| ||||||||
▲ | foobarbecue 6 days ago | parent | prev [-] | |||||||
That's true about people with pencils and paper. But this is a computer, and my point is that with LLMs, you have a computer program that doesn't compute or do logic -- and for some reason people act like it does. I mean, this is a super basic python task that I could write in 2 minutes. The naiive implementation is a couple of basic for loops. I wouldn't hire a junior dev that couldn't write it. My point here is that for someone who can program, it's easier and better to write this in python than to write it in english and have an LLM translate it to python (and LLMs themselves are incapable of logic or reasoning, so the only option is for them to translate it to real code). | ||||||||
|