AIME2024 has 30 Tests - Cant score 80.96

#36
by fblgit - opened

Hi there... Im not sure how this AIME2024 evaluation has been done.. but AIME2024 has 15 & 15 tests.
R1 scored 79.8/80 meaning, 24 out of 30 samples were answered correctly.
If this model has outperformed R1, means it answered correctly 25 questions.. scoring 83.3..
What kind of AIME2024 has been used that is able to produce fractions of the questions?

I refer to:

It also feels very strange being able to improve a result on this test while degrading the majority of the known benchmarks..

Screenshot 2025-02-19 at 9.01.48 AM.png

What are you talking about 😂 ? Clearly this is not 79.8, maybe there's something off in your understanding somewhere.

How many right answers this model got on AIME to score 80.96 ?
Because R1 got 24 right answers to score his benchmark.

@fblgit I think the decimal depth comes from averaging multiple runs

@eugenhotaj-ppl Could anyone of your team address this massive issue about the closed source dataset and classifier? The whole community would be so grateful if you decide to make it publicly available, as this could serve the ones that need it the most.
You still can change the trajectory. But time window is shrinking. So if you need more to evaluate your move, you can also freeze everything, remove the model, apologize for all the trouble it caused. Then take the necessary time to evaluate the situation and take into consideration all the feedback you receive.

Feedback is a gift people offer by love. And it's precious. I hope you could learn this the easier way.
Beware recover from trahison is hard and not always happening.

Sign up or log in to comment