NeurIPS 2025 E2LMC: Early-stage Evaluation for LLM Training Competition

A blogpost that walks through the NeurIPS 2025 E2LMC competition: Early-stage Evaluation for LLM Training Competition - co-authored by the organizing and winning teams.

Affiliations

  1. TII
  2. Safran Tech
  3. EPFL
  4. Swisscom
  5. Universidade Estadual de Campinas
  6. IMT Atlantique
  7. EleutherAI

Published

Apr. 30, 2026

PDF

Table of Contents

Introduction

The evaluation of large language models has matured considerably in recent years, with benchmarks such as MMLU Hendrycks et al. (2020), GPQA Rein et al. (2023), MATH-Hard Hendrycks et al. (2021), and LiveCodeBench Jain et al. (2024) becoming standard tools for measuring capabilities across scientific knowledge, reasoning, and code generation. These benchmarks have proven effective at differentiating fully trained models, but they share a common blind spot: they were not designed with early training in mind. When applied to small language models (0.5B–3B parameters) during the first 200 billion tokens of training, these benchmarks consistently produce noisy, non-discriminative signals. Scores fluctuate near random baselines, model size orderings break down, and architectural differences become invisible. This is not merely an inconvenience, it represents a fundamental gap in our ability to make informed decisions about hyperparameters, data mixtures, and architectural choices at a stage where such decisions have the greatest downstream impact. The E2LM Competition (Early Training Evaluation of Language Models), hosted at NeurIPS 2025, was designed to address this gap. We provided participants with six pre-trained models across three scales, each with two architectural variants, along with intermediate checkpoints sampled throughout training. The challenge was to design evaluation tasks in the scientific knowledge domain that produce meaningful, monotonically improving signals during early training, while maintaining consistent model rankings at convergence. As a baseline, we demonstrated that even simple prompt reformulations, such as the cloze-style variant of MMLU, can recover informative learning curves where standard multiple-choice formats fail. This observation, inspired by the work of Muennighoff et al. Muennighoff et al. (2024), suggested that the problem lies not in what models know during early training, but in how we ask them to demonstrate it. In this post, we present a retrospective analysis of the competition: we describe the experimental setup proposed to participants, and briefly present the winning solutions.

Training signal comparison for 500M, 1B, and 3B dense models.
Noisy results obtained with state-of-the-art benchmarks with different models sizes.

Beyond the noise problem, there is also a question of timing. The earlier we can extract meaningful evaluation signals, the more useful they become. Decisions about data mixtures, architectures, and hyperparameters made in the first few hundred billion tokens have outsized impact on the final model. Yet most scientific benchmarks fail to provide discriminative signals precisely during this critical window, making it costly and difficult to derive conclusive insights from small-scale experiments. This challenge is not uniform across domains. Recent LLM releases have shown rapid improvements on STEM-related tasks such as mathematics and code generation, yet evaluating these capabilities early in training remains particularly difficult. By contrast, commonsense-related benchmarks like HellaSwag Zellers et al. (2019) tend to produce informative signals even at early stages and across model sizes. The gap between what we can measure early (general language understanding) and what we most want to measure early (scientific and reasoning capabilities) is at the heart of this competition. This gap becomes even more pronounced for small language models, which are inherently weaker than their larger counterparts and therefore harder to differentiate using benchmarks designed for frontier models.

Training signal comparison for 500M, 1B, and 3B dense models.
HellaSwag produces smooth, discriminative signals even at small scales and early in training.

MMLU benchmark

Evaluating STEM capabilities of a language model is a topic on its own – MMLU has been proven to be the de facto benchmark to evaluate general scientific knowledge of a language model. MMLU simply consists of MCQ-style questions from various topics, and the target model is prompted to answer a given MCQ question.

MMLU benchmark results for 500M, 1B, and 3B dense models.
In blue, 500M dense model, orange 1B dense model, green 3B dense model.

MMLU becomes a meaningful evaluation metric only after a model has been trained on a sufficient number of tokens, and this tends to hold true primarily for models larger than 1B parameters. For smaller models, the benchmark produces essentially random results. Since MMLU is a 4-choice multiple-choice task, a score of around 0.25 corresponds to random guessing, which is often what we observe in these smaller variants. A variant of MMLU, termed as MMLU-Var Muennighoff et al. (2024), is considered today to overcome this problem.

Why MMLU-Var is stable and consistent?

MMLU-Var, a variant of MMLU has been proven to give much stronger signals on Small Language Models. Let’s analyze in detail what is the simple “trick” used by MMLU-Var compared to classic MMLU.

MMLU-Var signal quality comparison for 500M, 1B, and 3B dense models.
In blue, 500M dense model, orange 1B dense model, green 3B dense model – MMLU-Var gives much higher signal quality than MMLU benchmark, even for smaller variants.

MMLU is a “log-likelihood” style MCQ benchmark. Given a prompt which states the question to be asked to the language model, 4 possible answers are given, and the log-likelihood of the tokens corresponding to the different choices (A, B, C, D) given the question are computed, and an answer is counted as being correct if the highest likelihood corresponds to the actual answer. In this case, as stated above, a random model would give a score of 0.25.

More concretely, assuming we have 4 choices per QA pair, 4 prompts are constructed from the QA pairs with each possible answer.

Example MMLU benchmark prompts highlighting answer token log-likelihoods.
Example prompts constructed in MMLU benchmark, tokens where log-likelihoods are extracted for answer prediction is represented in bold.

The log-likelihoods are then calculated on the answer tokens, and the answer which has the highest log-likelihood is considered the predicted answer. The answer from the model is then compared against the ground truth answer.

MMLU-Var converts the MCQ format into “completion-mode”. Instead of prompting the model in a MCQ style, 4 different prompts are constructed which represent the continuation version of the MCQ. The log-likelihood is computed on the answer tokens (in case multiple tokens, the sum of the log-likelihoods is considered), and the prompt with the highest log-likelihood on these tokens is considered as the predicted answer from the model.

Example MMLU-Var prompts highlighting answer token log-likelihoods.
Example prompts constructed in MMLU-Var benchmark, tokens where log-likelihoods are extracted for answer prediction is represented in bold.

For small language models, this “trick” converts MMLU benchmark from noisy signals to very meaningful ones. This could be explained by the fact that the trick helps the model to escape from the MCQ domain, which might be hard to grasp during the early training stage of Small Language Models, to a more natural next-token-prediction task.

Why This Matters for the LLM Community?

Given these observations, we believe there is substantial room to explore how existing benchmarks can be adapted to produce higher-quality signals on small language models, and how entirely new evaluation tasks might be designed to capture early training dynamics more effectively. Yet the scope of this problem extends well beyond what any single team can address. Meaningful progress requires the collective engagement of the broader LLM community, drawing on diverse perspectives from machine learning, the domain sciences, and evaluation methodology. This is what motivated us to frame the effort as an open competition. Concretely, we invited participants to investigate several open questions: Can existing popular benchmarks be adapted to yield more meaningful and discriminative signals? Can new benchmarks be designed from the ground up that give informative signal in the early stages of training? Are there alternatives to completion-style prompting that better capture early signals in small language models?

Competition format

The participants were provided multiple model checkpoints to facilitate their benchmark validation. These models were trained on two distinct data mixtures: Web-only data (random subset of FineWeb Penedo et al. (2024), a cleaned, deduplicated English web corpus from CommonCrawl) and Scientific Knowledge Data Mixture (FineWeb-edu Lozhkov et al. (2024) (50%), The Stack V1/V2 Kocetkov et al. (2022) (21.6%), InfiMM Han et al. (2024) (18.9%), TxT360 Tang et al. (2024) (9.5%)). We provided models at three scales each instantiated with two distinct architectural variants for the same size: a deep variant (arch1) and a wide variant (arch2), assuming deeper models reason better (more details can be found in our competition proposal Yagoubi et al. (2025)).

All submissions were independently evaluated across three distinct metrics: Signal Quality (SQ), Ranking Consistency (RC), and Compliance with Scientific Knowledge Domains (CS). The final score was computed as a weighted linear combination of these metrics. Additionally, submissions were validated for alignment with established scientific knowledge domains and screened for potential information leakage.

The overall score was calculated as:

Score=α1×ScoreSQ+α2×ScoreRC+α3×ScoreCS\htmlId{total_score}{\text{Score} = \alpha_{1} \times \mathbf{Score_{\mathrm{SQ}}} + \alpha_{2} \times \mathbf{Score_{\mathrm{RC}}} + \alpha_{3} \times \mathbf{Score_{\mathrm{CS}}}}

where α1\alpha_1, α2\alpha_2, and α3\alpha_3 represent the weight coefficients assigned to each respective metric. These weights were determined based on the relative importance of each evaluation criterion and initialized at α1=0.5\alpha_1 = 0.5, α2=0.1\alpha_2 = 0.1, and α3=0.4\alpha_3 = 0.4, thereby prioritizing signal quality, followed by compliance with scientific knowledge domains, and ranking consistency.

To enable participants to verify the correctness of their approaches locally, participants were provided with a scoring notebook as part of the starter kit, enabling them to compute ScoreSQ using openly available model checkpoints. To mitigate the risk of overfitting to these open checkpoints, a subset of evaluation checkpoints was withheld from participants. Upon submission to the Hugging Face (HF) evaluation space, the overall score was computed automatically using the complete set of checkpoints, and the results were made visible to participants through the platform interface.

Each sub-score offers a concrete metric for assessing whether the benchmark yields meaningful and reliable insights. Broadly, the ScoreSQ component rewards tasks that generate smooth, informative learning curves across the checkpoints, whereas the ScoreRC metric captures the stability, reproducibility, and robustness of model rankings over the full set of checkpoints. Finally, the ScoreCS score evaluates the degree to which the benchmark aligns with principles of sound reasoning and scientific knowledge.

A detailed breakdown of the computation methodologies and evaluation procedures for each sub-score is provided below.

Signal Quality

As mentioned above, this score rewards tasks that produce smooth and informative curves across the evaluated checkpoints. To ensure that only smooth and positive trend curves get rewarded, this score is broken down in two parts:

ScoreMonotonicity=max(0,16dj2n(n21))\htmlId{monotonicity}{\text{Score}_{\mathrm{Monotonicity}} = \max\left(0, 1 - \frac{6 \sum d_j^2}{n(n^2 - 1)}\right)} ρ=i=1n(xixˉ)(xi+xˉ)i=1n(xixˉ)2×i=1n(xi+xˉ)2\htmlId{autocorr_lag}{\rho_\ell = \frac{\sum_{i=1}^{n-\ell}(x_i - \bar{x})(x_{i+\ell} - \bar{x})}{\sqrt{\sum_{i=1}^{n-\ell}(x_i - \bar{x})^2} \times \sqrt{\sum_{i=1}^{n-\ell}(x_{i+\ell} - \bar{x})^2}}}

The autocorrelation score is the average of the absolute correlations across all lags:

ScoreAutoCorr=1L=1Lρ\htmlId{autocorr}{\text{Score}_{\mathrm{AutoCorr}} = \frac{1}{L} \sum_{\ell=1}^L |\rho_\ell|}

Finally, ScoreSQ can be calculated as the weighted combination of both while weighing both equally so that both directional learning progress as well as temporal stability are rewarded equally.

ScoreSQ=β1×ScoreMonotonicity+β2×ScoreAutoCorr\htmlId{score_sq}{\text{Score}_{\mathrm{SQ}} = \beta_1 \times \text{Score}_{\mathrm{Monotonicity}} + \beta_2 \times \text{Score}_{\mathrm{AutoCorr}}}

Ranking Consistency

For the metric to be stable, the ranking of models as training progresses needs to be consistent and reliable, specifically after processing a large number of tokens (1 trillion tokens). Ranking consistency needs to be evaluated for models of different sizes (0.5B, 1B, 3B) separately, and the final score is computed by averaging across these configurations. However, a question arises; how do we measure consistency? Prior works have used Kendall’s Tau Coefficient Kendall (1938) to evaluate the metric consistency across the training steps.

We implemented it as follows:

rank200BT(s)={1if r(s)>00otherwiser(s)=1KjKxjArch1(s)xjArch2(s)\htmlId{baseline_ranking}{\begin{aligned} \mathrm{rank}_{200\mathrm{BT}}(s) &= \begin{cases} 1 & \text{if } r(s) > 0 \\ 0 & \text{otherwise} \end{cases} \qquad & r(s) &= \frac{1}{|K|} \sum_{j \in K} x^{\mathrm{Arch1}}_j(s) - x^{\mathrm{Arch2}}_j(s) \end{aligned}}

where K={kN100k<200}K = \{k \in \mathbb{N} \mid 100 \leq k < 200\}, and xjArch1x^{\mathrm{Arch1}}_j and xjArch2x^{\mathrm{Arch2}}_j are the scores of models Arch1\mathrm{Arch1} and Arch2\mathrm{Arch2} at checkpoint jj respectively. If Arch1\mathrm{Arch1} is on average better than Arch2\mathrm{Arch2}, then rank200BT(s)=1\mathrm{rank}_{200\mathrm{BT}}(s) = 1, otherwise rank200BT(s)=0\mathrm{rank}_{200\mathrm{BT}}(s) = 0.

ScoreRC=τ(s)=1PpPτp(s)τp(s)={1if rank200BT(s)=rankp(s)0otherwise.\htmlId{ranking_consistency}{\begin{aligned} \text{Score}_{\mathrm{RC}} = \tau(s) &= \frac{1}{|P|} \sum_{p \in P} \tau_p(s) \qquad & \tau_p(s) &= \begin{cases} 1 & \text{if } \mathrm{rank}_{200\mathrm{BT}}(s) = \mathrm{rank}_p(s) \\ 0 & \text{otherwise.} \end{cases} \end{aligned}}

where rankp(s){0,1}\mathrm{rank}_p(s) \in \{0,1\}, and rankp(s)=1\mathrm{rank}_p(s) = 1 if xpArch1(s)>xpArch2(s)x^{\mathrm{Arch1}}_p(s) > x^{\mathrm{Arch2}}_p(s), otherwise rankp(s)=0\mathrm{rank}_p(s) = 0.

Compliance to Reasoning and Knowledge Domains (ScoreCS)

This metric measures whether an evaluation task tests scientific reasoning and knowledge rather than general language or commonsense abilities. From our analysis, science-focused tasks like MMLU-var clearly distinguish between models trained on curated scientific data versus web-only data. By contrast, commonsense benchmarks like HellaSwag produce similar results for both model types, making them unsuitable for this competition.

To quantify domain compliance, we compare two 1B models across training steps: one trained on scientific knowledge data.

ScoreCS=max(0,1ni=1n(xiSciKWDSxiWebDS))\htmlId{compliance_score}{\text{Score}_{\mathrm{CS}} = \max\left(0, \frac{1}{n} \sum_{i=1}^{n} \left(x_i^{\mathrm{SciKW-DS}} - x_i^{\mathrm{Web-DS}}\right)\right)}

This formula normalizes the performance gap, rewarding tasks that effectively measure knowledge-intensive learning regardless of difficulty level.

Metric Scores Analysis

MMLU-var ranks highest overall, followed by ARC-Easy and SciQ, which perform well across all metrics. ARC-Easy is particularly effective for early-stage training assessment and serves as a secondary baseline alongside MMLU-var. However, 82% of SciQ questions contain verbatim answers in their prompts, potentially inflating scores. We address this by applying leakage checks to all submissions and excluding such questions.

Scientific knowledge benchmarks consistently outperform HellaSwag despite its higher Consistency and Signal Quality scores. HellaSwag scores zero on the Compliance metric because it doesn’t test scientific knowledge. While commonsense tasks like HellaSwag and Winogrande appear in our full rankings

To validate our metric’s robustness, similar submissions won’t appear on the official leaderboard as they fail the scientific compliance requirement.

Competition infrastructure

Overall organization

We hosted our competition on a dedicated Hugging Face organization, where teams would join the organization and submit their solutions through a dedicated HuggingFace Space. The submission Space will push each submission bundle in a dedicated private HuggingFace storage. On our side, we fetch the private dataset during each fixed time period and run the series of evaluations. We have also provided the participants various Google Colab notebooks to get started quickly in integrating new benchmarks using the package.

Design of the overall competition infrastructure
Design of the overall competition infrastructure

Evaluation backend: lm-evaluation-harness

We asked the participants to develop their benchmarks using lm-evaluation-harness from EleutherAI Gao et al. (2023) - for practical reasons, we have ‘frozen’ the library to a specific commit and made it public under our GitHub organization, and asked the participants to submit git .patch files.

How to submit a benchmark?

Once the benchmark is ready, participants needed to generate the corresponding git .patch file, together with a very simple yaml file that contains few infiormation such as the benchmark name, the metric name as well as an optional HF fine-grained token to eventually access private datasets.

Key Statistics

We present below the key statistics of the competition:

MetricValue
HF organization members (total participants)156
Total submissions236 (136 passed, 100 failed)
Total registered teams128
Total active teams12

Active teams are defined as those with at least one passed submission. As a side note, we also open source all our models under this Hugging Face collection.

Presentation of Winning solutions

Team Morai

The solution by Team Morai extended the MMLU Var baseline by introducing a novel evaluation metric and a filtering procedure designed to select questions with a meaningful learning signal. Although results are shown on the MMLU benchmark, the team anticipates that these adaptations are applicable to other benchmarks and will yield similar performance improvements.

Proposed metric

MCQ example where a model assigns the highest probability to the correct answer.
Models output the highest probability to the correct answer.
MCQ example where models assign the same probability to the correct answer.
Models output the same probability of the correct answer.

Illustrative example with a MCQ in a completion format.

The proposed metric identifies models capable of differentiating between scientific truth and linguistically coherent but factually incorrect text. Although many models demonstrate high linguistic capabilities, that is, generating text that mimics the structure of scientific discourse, the main goal is to select models that prioritize scientific compliance over mere fluency. Given a question QQ, let PCP_C be the probability assigned by the model to the correct answer, and let PIC\mathbf{P}_{IC} be a vector containing the probabilities of the incorrect choices. For example, for a multiple-choice question (MCQ) with four options, PIC\mathbf{P}_{IC} has a length of three. Our metric is defined as:

s(PC,PIC)=log(PC)1PICpPIClog(p)\htmlId{morai-metric}{s(P_C,\mathbf{P}_{\mathrm{IC}})=\log(P_C)-\frac{1}{|\mathbf{P}_{\mathrm{IC}}|}\sum_{p\in\mathbf{P}_{\mathrm{IC}}}\log(p)}

Consider a biology MCQ presented to two models (see Fig. 1). Both models assign the highest probability to the correct answer, resulting in an identical accuracy score of 1. However, since the second model assigns a significantly higher probability to the correct answer, it demonstrates greater “confidence” in the underlying scientific knowledge. A metric defined as the log-likelihood of the correct answer would correctly rank the second model higher.

Now, consider a scenario where both models assign an identical probability of 0.6 to the correct answer, as in Fig. 2. A log-likelihood metric would treat them as equivalent. However, if the first model distributes the remaining probability among incorrect answers that are linguistically plausible within the domain, while the second model assigns much lower probabilities to those distractors, the second model exhibits better discriminative power. By applying the metric in Eq. 1, the second model is identified as superior. While alternative formulations, such as omitting the log operator or using max/min\max/\min operations instead of an average, were considered, preliminary experiments indicated that the current formulation provides the most stable learning signal.

Filtering procedure

Transitioning from a discrete to a continuous metric increases score variance during training. For instance, a model’s probability for a correct answer might shift from 0.06 to 0.062 between checkpoints. While accuracy remains unchanged (both 1), the proposed metrics could increase from 0.33 to 0.35. To distinguish genuine model improvement from random variation, it was implemented a filtering procedure to retain only questions that provide a meaningful learning signal. That is, to exclude those that are either trivial or excessively difficult. Each question QQ in the MMLU dataset was evaluated using the following score:

s(Q)=16arc(PCfPCi)\htmlId{morai-filter-score}{s(Q)=\frac{1}{6}\sum_{\mathrm{arc}}(P_C^f-P_C^i)}

where PCfP_C^f and PCiP_C^i represent the probability of the correct answer at the final and initial checkpoints, respectively. A score near 0 indicates a question was either answered correctly from the start (too easy) or remained incorrect throughout training (too hard). This value is averaged across six evaluated architectures. The filtering keep only the top 50% of questions. As shown in Fig. 3, the selected questions predominantly exhibit increasing learning curves. In contrast, the removed questions show a high density at 0, representing items where the model failed to assign any meaningful probability to the correct answer throughout the entire training process.

Learning curves comparing the top and bottom 50% of questions.
Learning curves of top 50% questions and bottom 50%.

We compared the proposed metric with and without the filtering procedure. Using the metric alone yielded a global score of 0.72. Implementing the filtering procedure increased this score to 0.801 and significantly improved scientific compliance, which rose from 0.4 to 0.6.

Future directions

The proposed solution still has certain limitations. The current filtering procedure is static, based on initial training iterations. Future work will explore adaptive filtering, where the question subset is updated dynamically as training progresses. Additionally, it was observed a negative correlation between question score and answer length; the filtering approach tends to remove questions with longer textual responses. Future iterations will aim to normalize the methodology to ensure fair evaluation across varying answer lengths.

Team Shaikespear

Introduction

Team Shaikespeare focuses on evaluating a broad range of tasks designed for the challenge. Given the components of the metric used to score the submissions, particular emphasis is placed on scientific compliance. The analysis evaluates the influence of various benchmark characteristics, including the subjects covered by the datasets, individual sample properties, and the evaluation metric itself. The following sections describe the different experiments conducted before reviewing their results. Overall, Team Shaikespeare’s study is organized around three main experiments: (1) evaluating a diverse set of datasets, (2) assessing the performance of a likelihood-based metric, and (3) exploring strategies for filtering large datasets.

Methods

Model prompting

Let θ\theta denote the parameters of an autoregressive language model defining a distribution pθp_{\theta} over token sequences. For any prompt (context) xx and continuation y=(y1,,ym)y = (y_1,\dots,y_m), the conditional log-likelihood is

logpθ(yx)=t=1mlogpθ(ytx,y<t)\htmlId{ar-ll}{\log p_{\theta}(y \mid x)=\sum_{t=1}^{m} \log p_{\theta}\left(y_t \mid x, y_{<t}\right)}

where y<t=(y1,,yt1)y_{<t}=(y_1,\dots,y_{t-1}).

Standard MCQA prompting. In the multiple-choice question answering (MCQA) setting, each instance consists of a question qq and a finite set of answer candidates C={ci}i=1KC=\{c_i\}_{i=1}^{K}, where each cic_i is a textual option. A common evaluation protocol formats the input by explicitly enumerating the candidates in the prompt, via a template πmc(q,C)\pi_{\mathrm{mc}}(q,C), and asks the model to output a choice identifier (e.g., a label i{A,B,C,D}\ell_i \in \{\texttt{A},\texttt{B},\texttt{C},\texttt{D}\}). The predicted index is then

i^mc=argmaxi{1,,K}logpθ(iπmc(q,C))\htmlId{mcqa-label}{\hat{i}_{\mathrm{mc}}=\underset{i \in \{1,\dots,K\}}{\text{argmax}}\log p_{\theta}\bigl(\ell_i \mid \pi_{\mathrm{mc}}(q,C)\bigr)}

This formulation evaluates the model’s ability to discriminate among options when they are jointly presented in-context.

Completion (continuation) mode prompting. Following the competition baseline (MMLU-var), a completion/continuation-based prompting strategy is adopted: the prompt contains the question but omits the answer options, via a template πcomp(q)\pi_{\mathrm{comp}}(q) ending with Answer:. Each candidate cic_i is then scored as a separate continuation of this prompt, and the selected answer is the one with the highest log-probability under the model. Formally, letting tok(ci)=(y1(i),,ymi(i))\mathrm{tok}(c_i) = (y^{(i)}_1,\dots,y^{(i)}_{m_i}) be the tokenization of cic_i, we compute

sθ(ci;q)=logpθ(tok(ci)πcomp(q))=t=1milogpθ(yt(i)πcomp(q),y<t(i))\htmlId{comp-score}{s_{\theta}(c_i; q)=\log p_{\theta}\bigl(\mathrm{tok}(c_i) \mid \pi_{\mathrm{comp}}(q)\bigr)=\sum_{t=1}^{m_i} \log p_{\theta}\left(y^{(i)}_t \mid \pi_{\mathrm{comp}}(q), y^{(i)}_{<t}\right)}

and predict

i^comp=argmaxi{1,,K}sθ(ci;q)\htmlId{comp-argmax}{\hat{i}_{\mathrm{comp}}=\arg\max_{i \in \{1,\dots,K\}} s_{\theta}(c_i; q)}
Datasets

A large range of scientifically oriented datasets was gathered in order to evaluate different dataset characteristics. This corpus comprises existing open datasets as well as a custom one.

Existing Datasets

Most of the datasets used for benchmarking come from publicly released sources. Sub-sampling keeps the number of samples smaller than 15k15k when source datasets are too large in order to respect the dataset size limitations required by the competition.

Custom Datasets

To complement the previous benchmarks, the team proposes a novel college-grade dataset comprising MCQA items drawn from course quizzes and exams at Ecole Polytechnique Federale de Lausanne (EPFL). Originally created in 2024 for Direct Preference Optimization (DPO) and Supervised Fine-Tuning (SFT) of large language models, MCQA questions were extracted from these materials to construct the EPFL_QA dataset.

This dataset comprises questions from STEM subjects ranging life science, chemistry, physics, mechanics, computer science, data science, cryptography, math, quantum-physics, machine-learning and information retrieval. Sample QA pairs are given below and show the diversity of questions that evaluate course knowledge or reasoning and may contain numerous symbols.

Metrics

All experiments are conducted using accuracy except when specified otherwise.

Conditional Log-Likelihood (CLL) metric

Because of the binary output of the accuracy metric, non-smooth patterns may appear in the scoring curves, even when large numbers of samples are aggregated. To address this issue, a likelihood-based metric was introduced, the Conditional Log-Likelihood, which computes, given a question/answer pair, the normalized conditional log likelihood of the answer with respect to the question:

CLL(q,a)=1a[logp(q,a)logp(q)]\htmlId{cll}{\mathrm{CLL}(q,a)=\frac{1}{|a|}\left[\log p(q,a)-\log p(q)\right]}

where (q,a)(q, a) is the question-answer pair and p(x)p(x) is the probability of the sequence xx.

Datasets Filtering

Although dataset size influences the scoring metric, large datasets also introduce computational overhead, especially for the competition setup that evaluates submitted tasks once for every model checkpoint.

Three different pruning strategies are presented to keep only the most relevant samples in the current context.

Category based pruning. The team first performs category-based pruning by leveraging the pre-existing partitions provided by the original dataset authors. Concretely, each dataset D\mathcal{D} is partitioned into KK splits (Dk)k=1K(\mathcal{D}_k)_{k=1}^K that group samples with shared semantic characteristics. Selected partitions DS\mathcal{D}_S, where S{1,,K}S \subseteq \{1, \ldots, K\}, are then evaluated separately.

The team uses this approach to diagnose split-specific contributions to the overall score. In particular, since many datasets are curated with semantically meaningful partitions (e.g., subject areas or skill types), the analysis investigates whether partitions emphasizing science and logic yield higher rewards than others.

Scientific pruning. A scientific score function is defined and used by the team to prune input datasets, retaining only samples that are most intrinsically scientific. This function relies on the previously defined CLL, together with two auxiliary models, msm_s and mgm_g, representing scientific and general-purpose models, respectively. Denoting their corresponding probability density functions as psp_s and pgp_g, the scientific score for a question-answer pair (q,a)(q, a) is defined as the difference between the two models’ CLLs:

SciScorems,mg(q,a)=CLLms(q,a)CLLmg(q,a)=1a[logps(q,a)logps(q)]1a[logpg(q,a)logpg(q)]SciScorems,mg(q,a)=1alogps(aq)pg(aq)\htmlId{sci-score}{\begin{aligned} \mathrm{SciScore}_{m_s,m_g}(q,a) &= \mathrm{CLL}_{m_s}(q, a) - \mathrm{CLL}_{m_g}(q, a)\\ &= \frac{1}{|a|}\left[\log p_s(q,a) - \log p_s(q)\right] - \frac{1}{|a|}\left[\log p_g(q,a) - \log p_g(q)\right]\\ \mathrm{SciScore}_{m_s,m_g}(q,a) &= \frac{1}{|a|}\log\frac{p_s(a|q)}{p_g(a|q)} \end{aligned}}

Samples with higher scores correspond to QA pairs that increase the confidence of the scientific model with respect to the general-purpose one. Because of the scale invariance of this score (doubling both models’ probabilities does not change the score), this function does not discriminate based on sample difficulty and instead evaluates scientific content.

Sample quality pruning. To filter out low-quality web samples, the team uses the GneissWeb data-preparation recipe as implemented in the Data Prep Kit (DPK) (Gohari et al., 2025; Wood et al., 2024). GneissWeb is selected because its recipe is explicitly designed to achieve a favorable quality-quantity trade-off and combines multiple complementary annotators to enable finer-grained filtering decisions.

Concretely, each document (a question-answer pair) is processed in two stages: exact line-level deduplication followed by quality filtering using a combination of learned annotators and heuristic checks. The quality signals computed for each document are described below.

TPC(d)=T(d)dchar\htmlId{tpc}{\mathrm{TPC}(d)=\frac{T(d)}{\lvert d \rvert_{\text{char}}}}

where T(d)T(d) is the number of tokenizer output tokens for document dd.

A document is retained only if it satisfies all of the following conditions:

  1. Quality gate: it exceeds at least one document-level quality threshold (DCLM or Cosmo).
  2. Category gates: it exceeds minimum score thresholds for all kCk \in \mathcal{C}.
  3. Readability or tokenization gate: it either has sufficiently good readability or falls within an acceptable range of token-per-character values.

Formally, a document dd is kept according to the following criterion:

1{d kept}=1 ⁣{(sdclm(d)>τdclm  scosmo(d)>τcosmo),(ctech(d)>τtech  cmed(d)>τmed  cedu(d)>τedu  csci(d)>τsci),(r(d)<τr  τTPCTPC(d)τTPC+).}\htmlId{doc-kept}{\mathbf{1}\{d\ \text{kept}\}=\mathbf{1}\!\left\{\begin{array}{l} \Bigl(s_{\mathrm{dclm}}(d)>\tau_{\mathrm{dclm}}\ \lor\ s_{\mathrm{cosmo}}(d)>\tau_{\mathrm{cosmo}}\Bigr),\\[3pt] \Bigl( c_{\mathrm{tech}}(d)>\tau_{\mathrm{tech}} \ \land\ c_{\mathrm{med}}(d)>\tau_{\mathrm{med}} \ \land\ c_{\mathrm{edu}}(d)>\tau_{\mathrm{edu}} \ \land\ c_{\mathrm{sci}}(d)>\tau_{\mathrm{sci}} \Bigr),\\[3pt] \Bigl( r(d)<\tau_r \ \lor\ \tau^{-}_{\mathrm{TPC}} \le \mathrm{TPC}(d) \le \tau^{+}_{\mathrm{TPC}} \Bigr). \end{array}\right\}}

where the thresholds τ\tau are user defined.

Results

Dataset Analysis of category based pruning

A comparison between STEM and non-STEM oriented datasets is provided in Table 1, which displays the results of the evaluation on different splits of the MMLU and MMLU Pro datasets. These results highlight that technical datasets tend to have better Scientific Compliance (SC).

DatasetSignal QualityRanking ConsistencyScientific ComplianceGlobal Score
MMLU0.9590.8370.4190.731
MMLU STEM0.9050.8510.4520.718
MMLU STEM0.9640.8000.4010.722
MMLU Pro0.9280.7300.5420.754
MMLU Pro STEM0.8610.8970.5740.749
MMLU Pro STEM0.8460.7370.4470.680
Table 1: STEM vs ~~STEM~~ scores.

Additional evaluations were performed on other datasets and confirm this observation, as seen in Table 2. More verbose datasets tend to have higher Signal Quality (SQ).

DatasetSignal QualityRanking ConsistencyScientific ComplianceGlobal Score
medmcqa0.9330.7350.440.716
tele_qna0.880.7590.4790.689
compsci0.7920.7610.4820.665
Table 2: Comparison of subject-focused datasets.
Influence of the number of tokens per sample

As seen in Figure 1, the number of tokens clearly discriminates the resulting curves, both in terms of accuracy and log-likelihood. One unusual result is how the sample curves corresponding to the higher number of tokens is higher than its counterpart for the MMLU dataset, while it is the opposite for the MMLU Pro dataset, even though both datasets display similar behaviors in log-likelihood.

Notably, the normalized log-likelihood curves exhibit a more consistent ordering across both benchmarks than accuracy does. This suggests that likelihood-based signals (even when normalized) are sensitive to sample length, but provide more stable early-training signals even when accuracy saturates or becomes noisy.

Accuracy and normalized log-likelihood curves for MMLU.Accuracy and normalized log-likelihood curves for MMLU Pro.
Figure 1: Analysis of the accuracy and normalized log-likelihood, for the number of tokens, against training checkpoints for the MMLU (top) and MMLU Pro (bottom) datasets.
CLL Metric

The CLL metric was compared against the standard accuracy metric on the MMLU Pro dataset, as seen in Table 3, and shows worse scores than the latter, especially for SC.

MetricSQRCSCScore
Accuracy0.930.730.540.75
CLL0.820.740.370.63
Table 3: Comparison of accuracy and CLL on MMLU Pro.
Custom Dataset: EPFL QA

The EPFL_QA dataset resulted in very low scores, seen in Table 5. This is likely due to the fact that many questions in the dataset lack context and may use symbols that are not introduced. In exam settings, the general context information and notations are usually defined once and are not reintroduced in every question.

Another possible reason is the difficulty blend of the dataset, as it contains both easy memorization questions from course quizzes and challenging reasoning and numerical questions. This may explain the early saturation of accuracy, as the model quickly learns how to answer the easy questions, while it never quite succeeds at answering the harder ones, as shown in the left plot of Figure 2. The right sub-figure also motivates this point as it highlights that the average normalized log-likelihood rewards richer and clearer contexts.

EPFL_QA accuracy and normalized log-likelihood curves.
Figure 2: Accuracy (left plot) and normalized log-likelihood (right plot) with respect to training checkpoints.
Sample quality pruning with GneissWeb

Using the GneissWeb prep kit and starting from the original MMLU collection DMMLU\mathcal{D}_{\mathrm{MMLU}} of MCQA items (each item is a document d=(q,a)d=(q,a)), the filtered subsets were constructed

DMMLU(τ)={dDMMLU  :  1{d kept}=1}\htmlId{gneissweb-dataset}{\mathcal{D}_{\mathrm{MMLU}}^{(\tau)}=\{\, d\in\mathcal{D}_{\mathrm{MMLU}} \;:\; \mathbf{1}\{d\ \text{kept}\}=1 \,\}}

where the team considers the default configuration τ(2.0)\tau^{(2.0)} provided with the recipe, and a strict configuration τ(3.0)\tau^{(3.0)} that tightens multiple filters such that

τ(3.0)={τdclm=0.9,τcosmo=0.9,τr=30,τTPC=0.10,τTPC+=0.50,τtech=0.5,τmed=0.5,τedu=0.5,τsci=0.95}\htmlId{gneissweb-tau}{\tau^{(3.0)}=\left\{\begin{array}{l} \tau_{\mathrm{dclm}}=0.9, \tau_{\mathrm{cosmo}}=0.9, \tau_r=30, \tau^{-}_{\mathrm{TPC}}=0.10, \tau^{+}_{\mathrm{TPC}}=0.50,\\ \tau_{\mathrm{tech}}=0.5, \tau_{\mathrm{med}}=0.5, \tau_{\mathrm{edu}}=0.5, \tau_{\mathrm{sci}}=0.95 \end{array}\right\}}

The metrics on the resulting benchmarks are reported in Table 4.

Overall, the default GneissWeb filtering τ(2.0)\tau^{(2.0)} produces results very similar to the original unfiltered MMLU benchmark. This suggests that moderate pruning can remove low-quality artifacts without changing the benchmark’s underlying signal. By contrast, the strict filtering setting τ(3.0)\tau^{(3.0)} hurts performance: it reduces both ranking consistency and scientific compliance.

To check whether filtering is still keeping the most useful questions, we look at the per-item average score ratio. This ratio captures, on average, how much each remaining item contributes to the total score compared to the unfiltered dataset:

per_item_avg_score=Global ScorefilterDfilterGlobal ScoreoriginalDoriginal=Global ScorefilterDoriginalDfilterGlobal Scoreoriginal\htmlId{per-item-avg-score}{\mathrm{per\_item\_avg\_score}=\frac{\frac{\mathrm{Global\ Score}_{\mathrm{filter}}}{|D_{\mathrm{filter}}|}}{\frac{\mathrm{Global\ Score}_{\mathrm{original}}}{|D_{\mathrm{original}}|}}=\frac{\mathrm{Global\ Score}_{\mathrm{filter}} \cdot |D_{\mathrm{original}}|}{|D_{\mathrm{filter}}| \cdot \mathrm{Global\ Score}_{\mathrm{original}}}}

If this ratio is close to 1, filtering kept items that are about as informative as the original average item. If it is much larger than 1, the filtered set is (on average) more informative per question than the original.

Dataset/MethodSQRCSCGlobal ScoreDataset SizePer Item Avg Score
MMLU0.9590.8370.4190.731158581.00
MMLU + τ(2.0)\tau^{(2.0)}0.9580.8230.4200.729149251.06
MMLU + τ(3.0)\tau^{(3.0)}0.9560.7370.3960.71061802.49
Table 4: Impact of GneissWeb sample-quality pruning on an MMLU-derived benchmark. The default recipe ($\tau^{(2.0)}$) preserves overall behavior, whereas stricter thresholds ($\tau^{(3.0)}$) reduce both ranking consistency and scientific compliance.
Science Score Filtering

The Science score can be used to prune large datasets. Since reasoning-oriented datasets consistently outperform others, as is the case when comparing MMLU Pro to MMLU in Table 1, the choice of auxiliary models can be made in order to retrieve the most reasoning-oriented samples. To this end, the team choose the following two models:

This filtering was performed to keep only 15k15k samples out of a dataset merging MMLU, MMLU Pro, and EPFL_QA. Subsequent results are shown in Table 5 with the SciScore pruning entry.

General Results

The general results in Table 5 crown the SciKnowEval benchmark. It is still worth mentioning that there is no clear winner in all categories.

Dataset/MethodSignal QualityRanking ConsistencyScientific ComplianceGlobal Score
EPFL_QA0.400.720.190.35
MMLU Pro (CLL)0.820.740.370.63
openbook_qa0.870.850.350.66
compsci0.790.760.480.66
tele_qna0.880.760.480.69
medmcqa0.930.730.440.72
SciScore pruning0.900.810.470.72
MMLU0.960.840.420.73
MMLU Pro0.930.730.540.75
SciKnowEval0.930.750.550.76
Table 5: General results. For each field, the best answer is in bold, the second best is underlined, the third best is italicized.

Conclusion and Perspectives

This work investigates early-stage evaluation and data curation strategies for scientific and logic-oriented language model training. The team’s results indicate that reasoning-focused datasets consistently improve scientific compliance, suggesting that the nature of the training signal matters already in the early regime. At the same time, the team deduced that likelihood-based metrics on their own do not reliably capture the behaviors of interest, and can be confounded by surface-form properties. In addition, sample length and verbosity substantially influence early signals, which motivates controlling for these factors when designing new benchmarks for early evaluation.

Looking ahead, an important direction is to incorporate explicit notions of sample difficulty into both pruning procedures and likelihood-based evaluation, so that filtering and scoring account for more than fluency and format effects. In parallel, benchmark design should better balance two competing objectives: producing smooth, stable curves at early checkpoints while still discriminating meaningfully between scientific reasoning capabilities. Together, these directions may lead to more faithful early proxies for downstream scientific performance and more effective data selection policies.

Team Noor

Evaluating Large Language Models (LLMs) at early stages of training remains a significant challenge. During this phase, models exhibit unstable behavior, and traditional metrics such as accuracy often produce noisy and volatile signals that obscure genuine learning progress. As a result, it becomes difficult to distinguish meaningful improvements from random fluctuations.

Rather than generating synthetic evaluation data, the team focuses on extracting a high-quality, high-signal subset from the existing MMLU-var benchmark. The analysis argues that many challenges associated with early-stage evaluation arise not from the models themselves, but from the inclusion of questions that either lack strong scientific grounding or fail to reflect progressive learning.

The objective is therefore twofold: (i) ensuring strict scientific relevance, and (ii) maximizing signal quality, defined as the smoothness and monotonicity of learning trajectories across training checkpoints.

Methodology

To enable rapid experimentation while remaining compute-efficient, the team adopted a compute once, filter many times strategy. All evaluation signals are precomputed in a single pass, enabling iteration exclusively on filtering logic without re-running expensive model evaluations.

1 - Comprehensive Data Generation

The team evaluated all three competition models (Dense-500M, Dense-1B, and Dense-3B) across all provided training checkpoints on the full MMLU-var dataset. For each question, the log-likelihood of all answer choices was computed, and the most probable option was selected as the model’s prediction. In addition to binary accuracy, the full log-probability distribution over choices was stored, enabling the computation of confidence-based metrics. This process resulted in a dense performance database containing accuracy and fine-grained confidence signals for every question-model-checkpoint triplet.

2 - Filtering Pipeline

From this comprehensive dataset, the team applied a two-stage filtering pipeline designed to retain only questions that are both scientifically grounded and informative of learning progress.

Scientific Compliance

The first stage enforces strict scientific relevance. Core Hard Science subjects, such as Mathematics, Physics, and Electrical Engineering, were automatically retained due to their well-defined and technical nature. For other scientific domains, including Biology, Medicine, and Computer Security, a rigorous validation step was applied using a Qwen2.5-7B LLM judge. Each question was independently prompted five times, and only those receiving a unanimous 5/55/5 Accept verdict were retained. This strict criterion prioritizes precision over coverage, ensuring high scientific alignment.

Signal Quality Optimization

The second stage targets the quality of the learning signal. For each remaining question, the evolution of a confidence-based score across training checkpoints was tracked and fit a linear regression line to this trajectory. Only questions exhibiting a positive slope—indicating consistent improvement over time—were retained. This step removes questions that are noisy, stagnant, or uninformative despite being scientifically valid.

Evaluation Metric: Confidence Margin

A central finding of the team’s work is the superiority of the Confidence Margin over standard accuracy for early-stage large language model evaluation. The Confidence Margin is defined as:

CM=logP(correct)maxlogP(other choices)\htmlId{confidence-margin}{\mathrm{CM}=\log P(\text{correct})-\max \log P(\text{other choices})}

This metric captures not only whether the model selects the correct answer, but also how strongly it prefers it relative to the most competitive distractor.

Accuracy versus confidence margin curves for the filtered benchmark.
Comparison of Accuracy (a) versus Confidence Margin (b) on the filtered benchmark.

Empirically, it is observed that accuracy curves at early checkpoints are highly volatile and non-monotonic, whereas Confidence Margin trajectories are significantly smoother. This allows the metric to reveal gradual learning dynamics that accuracy fails to capture, such as the steady promotion of the correct answer within the probability ranking before it becomes the top choice.

Conclusion

By combining a strict scientific compliance filter with a signal-quality filter based on Confidence Margin regression, the team constructed a benchmark that offers a clearer and more reliable lens for observing early-stage LLM development. The results suggest that improving early-stage evaluation does not necessarily require new data, but rather careful curation of existing benchmarks guided by metrics that reflect learning dynamics instead of final correctness alone.

Perspectives and Future Work

This work opens several promising directions for future research. First, signal modeling could be extended beyond linear trends to capture more complex learning dynamics. Second, subject-aware filtering strategies could account for differing learning behaviors across scientific domains. Finally, the proposed pipeline is benchmark-agnostic and could be applied to other evaluation datasets, contributing to more interpretable and reliable assessment of emerging LLMs.

Integration within lm-evaluation-harness

For winning solutions, we have decided to offer a native integration with the popular lm-evaluation-harness package in collaboration with the library authors and the participants. Anyone can use easily the benchmarks developed by the winning solutions starting from now, by making sure to install the latest version of lm-eval: pip install -U lm_eval. Below is an example command you can refer to:

lm_eval --model hf \
  --model_args pretrained=tiiuae/Falcon-H1-0.5B-Base \
  --tasks mmlu_early_training \ # `sciknoweval_mcqa` or `noor` 
  --device cuda:0 \
  --batch_size auto

Discussion

The E2LM competition asked whether we can design evaluation tasks that produce useful signals during early training. The results suggest that the challenge is twofold: it involves both what we evaluate and how we evaluate it. On one hand, prompt format plays a critical role: standard multiple-choice formats fail not because early-stage models lack scientific understanding, but because the format requires capabilities, such as comparing and choosing between options, that only develop later in training. On the other hand, the content and difficulty of the questions themselves matter just as much: benchmarks that are too hard, too easy, or poorly calibrated for the model’s developmental stage produce uninformative signals regardless of how they are prompted. Beyond prompt format, the competition highlighted several complementary strategies. Confidence-based metrics such as the Confidence Margin and log-probability gap consistently performed better than raw accuracy at detecting gradual learning progress. Careful selection of evaluation samples, through learning-curve-based filtering, scientific compliance scoring, or data quality pruning, can meaningfully improve benchmark quality without the need to create new data. Participants also went beyond adapting existing benchmarks: the EPFL_QA dataset, built from university-level STEM exams, represents a notable effort to create an entirely new evaluation resource from scratch. Building such a benchmark is far from straightforward: it requires collecting domain-specific content, calibrating difficulty appropriately, and handling challenges specific to academic material such as implicit context and notation. While the initial results exposed these difficulties, the dataset itself provides a solid starting point that the community can improve, expand, and adapt to other scientific fields.

What’s Next?

The E2LM competition has opened several promising research directions that we believe deserve continued attention from the community:

Building new evaluation resources The EPFL_QA dataset demonstrated both the value and the difficulty of creating benchmarks from scratch. While university-level exams are a natural source of calibrated scientific questions, adapting them for LLM evaluation requires careful handling of implicit context, notation, and difficulty distribution. We believe that extending it to other disciplines could yield a rich and diverse evaluation ecosystem for early training.

From early signals to scaling predictions. Our results offer encouraging evidence that early evaluation can serve as a reliable proxy for later training behavior: for benchmarks like MMLU-var and ARC-Easy, model rankings established at 200 billion tokens remained largely consistent through 1 trillion tokens. However, this consistency was not uniform across all configurations, and it remains to be seen whether these findings generalize beyond the 3B scale. Establishing a formal connection between early-stage signals and final model capabilities would be a significant step — it would allow practitioners to make confident architectural and data decisions based on a fraction of the total training compute.

Extending beyond scientific knowledge. This competition focused specifically on the scientific knowledge domain, but the underlying challenge (extracting meaningful evaluation signals during early training) applies broadly. Code generation, mathematical reasoning, and multilingual capabilities all face similar issues with noisy early benchmarks. Adapting the strategies and metrics explored here to these domains is a natural extension of this work.

Authors

TeamAuthors
FalconLLMYounes Belkada, Mugariya Farooq, Basma Boussaha, Mouadh Yagoubi, Yasser Dahou, Phuc Le Khac, Billel Mokeddem, Reda Alami
ShaikespearAnthony Kalaydjian, Eric Saikali
MoraiBeatriz Nascimento, Daniel Gardin, Caio Rhoden, Giovani Valdrighi
NoorMohammed Dahbani, Anas Ezzakri
EleutherAIBaber Abbasi
  1. Feng, K., Ding, K., Wang, W., Zhuang, X., Wang, Z., Qin, M., Zhao, Y., Yao, J., Zhang, Q., & Chen, H. (2024). Sciknoweval: Evaluating multi-level scientific knowledge of large language models. arXiv Preprint arXiv:2406.09098.
  2. Gao, L., Tow, J., Abbasi, B., Biderman, S., Black, S., DiPofi, A., Foster, C., Golding, L., Hsu, J., Le Noac’h, A., Li, H., McDonell, K., Muennighoff, N., Ociepa, C., Phang, J., Reynolds, L., Schoelkopf, H., Skowron, A., Sutawika, L., … Zou, A. (2023). A framework for few-shot language model evaluation (v0.4.0). Zenodo. 10.5281/zenodo.10256836
  3. Gohari, H. E., Kadhe, S. R., Shah, S. Y., Adam, C., Adebayo, A., Adusumilli, P., Ahmed, F., Angel, N. B., Borse, S., Chang, Y.-C., Dang, X.-H., Desai, N., Eres, R., Iwamoto, R., Karve, A., Koyfman, Y., Lee, W.-H., Liu, C., Lublinsky, B., … Bhattacharjee, B. (2025). GneissWeb: Preparing High Quality Data for LLMs at Scale. https://arxiv.org/abs/2502.14907
  4. Han, X., Jian, Y., Hu, X., Liu, H., Wang, Y., Fan, Q., Ai, Y., Huang, H., He, R., Yang, Z., & You, Q. (2024). InfiMM-WebMath-40B: Advancing Multimodal Pre-Training for Enhanced Mathematical Reasoning. https://arxiv.org/abs/2409.12568
  5. Hendrycks, D., Burns, C., Basart, S., Zou, A., Mazeika, M., Song, D., & Steinhardt, J. (2020). Measuring massive multitask language understanding. arXiv Preprint arXiv:2009.03300. back: 1, 2
  6. Hendrycks, D., Burns, C., Kadavath, S., Arora, A., Basart, S., Tang, E., Song, D., & Steinhardt, J. (2021). Measuring Mathematical Problem Solving With the MATH Dataset. arXiv Preprint arXiv:2103.03874.
  7. Jain, N., Han, K., Gu, A., Li, W.-D., Yan, F., Zhang, T., Wang, S., Solar-Lezama, A., Sen, K., & Stoica, I. (2024). LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code. https://arxiv.org/abs/2403.07974
  8. Kendall, M. G. (1938). A New Measure of Rank Correlation. Biometrika, 30(1–2), 81–93. 10.1093/biomet/30.1-2.81
  9. Kocetkov, D., Li, R., Ben Allal, L., Li, J., Mou, C., Muñoz Ferrandis, C., Jernite, Y., Mitchell, M., Hughes, S., Wolf, T., Bahdanau, D., von Werra, L., & de Vries, H. (2022). The Stack: 3 TB of permissively licensed source code. Preprint.
  10. Lozhkov, A., Ben Allal, L., von Werra, L., & Wolf, T. (2024). FineWeb-Edu: the Finest Collection of Educational Content . Hugging Face . https://doi.org/ 10.57967/hf/2497
  11. Maatouk, A., Ayed, F., Piovesan, N., Domenico, A. D., Debbah, M., & Luo, Z.-Q. (2023). TeleQnA: A Benchmark Dataset to Assess Large Language Models Telecommunications Knowledge. https://arxiv.org/abs/2310.15051
  12. Mihaylov, T., Clark, P., Khot, T., & Sabharwal, A. (2018). Can a suit of armor conduct electricity? a new dataset for open book question answering. arXiv Preprint arXiv:1809.02789.
  13. Muennighoff, N., Soldaini, L., Groeneveld, D., Lo, K., Morrison, J., Min, S., Shi, W., Walsh, P., Tafjord, O., Lambert, N., & others. (2024). Olmoe: Open mixture-of-experts language models. arXiv Preprint arXiv:2409.02060. back: 1, 2
  14. Pal, A., Umapathi, L. K., & Sankarasubbu, M. (2022). Medmcqa: A large-scale multi-subject multi-choice dataset for medical domain question answering. Conference on Health, Inference, and Learning, 248–260.
  15. Penedo, G., Kydlíček, H., allal, L. B., Lozhkov, A., Mitchell, M., Raffel, C., Werra, L. V., & Wolf, T. (2024). The FineWeb Datasets: Decanting the Web for the Finest Text Data at Scale. https://arxiv.org/abs/2406.17557
  16. Rein, D., Hou, B. L., Stickland, A. C., Petty, J., Pang, R. Y., Dirani, J., Michael, J., & Bowman, S. R. (2023). GPQA: A Graduate-Level Google-Proof Q&A Benchmark. https://arxiv.org/abs/2311.12022
  17. Song, X., Diao, M., Dong, G., Wang, Z., Fu, Y., Qiao, R., Wang, Z., Fu, D., Wu, H., Liang, B., Zeng, W., Wang, Y., GongQue, Z., Yu, J., Tan, Q., & Xu, W. (2025). CS-Bench: A Comprehensive Benchmark for Large Language Models towards Computer Science Mastery. https://arxiv.org/abs/2406.08587
  18. Tang, L., Ranjan, N., Pangarkar, O., Liang, X., Wang, Z., An, L., Rao, B., Jin, L., Wang, H., Cheng, Z., Sun, S., Mu, C., Miller, V., Ma, X., Peng, Y., Liu, Z., & Xing, E. P. (2024). TxT360: A Top-Quality LLM Pre-training Dataset Requires the Perfect Blend.
  19. Wang, Y., Ma, X., Zhang, G., Ni, Y., Chandra, A., Guo, S., Ren, W., Arulraj, A., He, X., Jiang, Z., Li, T., Ku, M., Wang, K., Zhuang, A., Fan, R., Yue, X., & Chen, W. (2024). MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark. https://arxiv.org/abs/2406.01574
  20. Wood, D., Lublinsky, B., Roytman, A., Singh, S., Adam, C., Adebayo, A., An, S., Chang, Y. C., Dang, X.-H., Desai, N., Dolfi, M., Emami-Gohari, H., Eres, R., Goto, T., Joshi, D., Koyfman, Y., Nassar, M., Patel, H., Selvam, P., … Daijavad, S. (2024). Data-Prep-Kit: getting your data ready for LLM application development. https://arxiv.org/abs/2409.18164
  21. Yagoubi, M., Dahou, Y., Mokeddem, B., Belkada, Y., Le-Khac, P. H., Boussaha, B. E. A., Alami, R., Zuo, J., Marsili, D., Farooq, M., Lalmas, M., Gkioxari, G., Gallinari, P., Torr, P., & Hacid, H. (2025). NeurIPS 2025 E2LM Competition: Early Training Evaluation of Language Models. https://arxiv.org/abs/2506.07731
  22. Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., & Choi, Y. (2019). HellaSwag: Can a Machine Really Finish Your Sentence? https://arxiv.org/abs/1905.07830