| ▲ | grog454 4 days ago |
| Ok, but then your "post" isn't scientific by definition since it cannot be verified. "Post" is in quotes because I don't know what you're trying to but you're implying some sort of public discourse. For fun: https://chatgpt.com/s/t_694361c12cec819185e9850d0cf0c629 |
|
| ▲ | eru 4 days ago | parent | next [-] |
| I didn't see anyone claiming any 'science'? Did I miss something? |
| |
| ▲ | grog454 4 days ago | parent [-] | | I guess there's two things I'm still stuck on: 1. What is the purpose of the benchmark? 2. What is the purpose of publicly discussing a benchmark's results but keeping the methodology secret? To me it's in the same spirit as claiming to have defeated alpha zero but refusing to share the game. | | |
| ▲ | nl 4 days ago | parent [-] | | 1. The purpose of the benchmark is to choose what models I use for my own system(s). This is extremely common practice in AI - I think every company I've worked with doing LLM work in the last 2 years has done this in some form. 2. I discussed that up-thread, but https://github.com/microsoft/private-benchmarking and https://arxiv.org/abs/2403.00393 discuss some further motivation for this if you are interested. > To me it's in the same spirit as claiming to have defeated alpha zero but refusing to share the game. This is an odd way of looking at it. There is no "winning" at benchmarks, it's simply that it is a better and more repeatable evaluation than the old "vibe test" that people did in 2024. | | |
| ▲ | grog454 4 days ago | parent [-] | | I see the potential value of private evaluations. They aren't scientific but you can certainly beat a "vibe test". I don't understand the value of a public post discussing their results beyond maybe entertainment. We have to trust you implicitly and have no way to validate your claims. > There is no "winning" at benchmarks, it's simply that it is a better and more repeatable evaluation than the old "vibe test" that people did in 2024. Then you must not be working in an environment where a better benchmark yields a competitive advantage. | | |
| ▲ | eru 4 days ago | parent [-] | | > I don't understand the value of a public post discussing their results beyond maybe entertainment. We have to trust you implicitly and have no way to validate your claims. In principle, we have ways: if nl's reports consistently predict how public benchmarks will turn out later, they can build up a reputation. Of course, that requires that we follow nl around for a while. |
|
|
|
|
|
| ▲ | nl 4 days ago | parent | prev [-] |
| As ChatGPT said to you: > A secret benchmark is: Useful for internal model selection That's what I'm doing. |
| |
| ▲ | grog454 4 days ago | parent [-] | | My question was "What's the value of a secret benchmark to anyone but the secret holder?" The root of this whole discussion was a post about how Gemini 3 outperformed other models on some presumably informal question benchmark (a"vibe test"?). When asked for the benchmark, the response from the op and and someone else was that secrecy was needed to protect the benchmark from contamination. I'm skeptical of the need in the op's cases and I'm skeptical of the effectiveness of the secrecy in general. In a case where secrecy has actual value, why even discuss the benchmark publicly at all? |
|