# Pretraining Language Models to Ponder in Continuous Space
> Zhouhan Lin is the corresponding author.
## Abstract
Humans ponder before articulating complex sentence elements, enabling deeper cognitive processing through focused effort. In this work, we introduce this pondering process into language models by repeatedly invoking the forward process within a single token generation step. During pondering, instead of generating an actual token sampled from the prediction distribution, the model ponders by yielding a weighted sum of all token embeddings according to the predicted token distribution. The generated embedding is then fed back as input for another forward pass. We show that the model can learn to ponder in this way through self-supervised learning, without any human annotations. Experiments across three widely used open-source architecturesâGPT-2, Pythia, and LLaMAâand extensive downstream task evaluations demonstrate the effectiveness and generality of our method. For language modeling tasks, pondering language models achieve performance comparable to vanilla models with twice the number of parameters. On 9 downstream benchmarks, our pondering-enhanced Pythia models significantly outperform the official Pythia models. Notably, PonderingPythia-2.8B surpasses Pythia-6.9B, and PonderingPythia-1B is comparable to TinyLlama-1.1B, which is trained on 10 times more data. The code is available at https://github.com/LUMIA-Group/PonderingLM.
## 1 Introduction
In the pursuit of improving model performance, scaling up model parameters and data sizes has long been the most widely adopted and accepted approach (Kaplan et al., 2020; Brown et al., 2020; Liu et al., 2024). However, this approach faces several bottlenecks, including the exhaustion of high-quality data (Villalobos et al., 2022; Muennighoff et al., 2023), the observed saturation in scaling laws (Hackenburg et al., 2025; Hoffmann et al., 2022a) and substantial communication overhead in distributed pre-training that grows super-linearly with model size (Narayanan et al., 2021; Pati et al., 2023; Li et al., 2024).
On the other hand, if we look at humans, the growth of human capabilities does not stem from simply increasing the number of neurons in the brain. Instead, when faced with complex problems, humans often enhance their problem-solving abilities by repeatedly pondering, engaging in deep cognitive processing before articulating their thoughts.
Analogously, in large language models, the most relevant research direction is test-time scaling. In particular, following the advancements in o1 and R1 (Jaech et al., 2024; DeepSeek-AI et al., 2025), generating long chains of thought (CoT) has emerged as the mainstream approach for scaling test-time computation. However, CoT-based methods also exhibit several drawbacks: they often require curated human-annotated datasets (Allen-Zhu & Li, 2023) and carefully designed reinforcement learning algorithms (Pang et al., 2025). Moreover, small models rarely benefit from CoT (Li et al., 2023), and the upper bound of performance remains constrained by the base pretrained model (Yue et al., 2025). Additionally, current language models employing CoT are still confined to discrete language spaces with fixed vocabularies, which, according to recent studies (Fedorenko et al., 2024; Hao et al., 2024; Pfau et al., 2024), are primarily optimized for communication rather than for internal computational thinking.
To overcome these challenges and inspired by human pondering, we introduce the Pondering Language Model (Pondering LM), which relies solely on self-supervised learning. Pondering LMs can be naturally learned through pretraining on large-scale general corpora, without the need for human-annotated datasets or reinforcement learning.
During pondering, instead of generating a discrete token sampled from the prediction distribution, the model produces a weighted sum of all token embeddings based on the predicted probabilities. This generated embedding is then fed back into the language model, allowing it to iteratively refine its predictions. As the weighted embedding is continuous, Pondering LMs overcome the expressive limitations of discrete token vocabularies and enable fully differentiable, end-to-end pretraining via gradient descent. Furthermore, by performing more computations per parameter, Pondering LMs achieve higher parameter knowledge density (Allen-Zhu & Li, 2024), potentially reducing communication costs at scale.
Experimentally, by introducing the pondering process during pretraining, our Pondering GPT-2, LLaMA, and Pythia models achieve pretraining perplexity comparable to that of vanilla models with twice as many parameters. Furthermore, Pondering Pythia models significantly outperform the official Pythia models across nine popular downstream tasks, PonderingPythia-2.8B surpasses Pythia-6.9B, and PonderingPythia-1B is comparable to TinyLlama-1.1B, which is trained on 10 times more data. Notably, increasing the number of pondering steps consistently enhances model performance, underscoring the substantial potential of this approach.
Moreover, our method is orthogonal to traditional scaling strategies, including parameter scaling and inference-time scaling via CoT, and thus can complement existing techniques, potentially introducing a third scaling axis to enhance model performance.
<details>
<summary>x1.png Details</summary>

### Visual Description
\n
## Diagram: Pondering Language Model Architecture
### Overview
The image depicts a diagram illustrating the architecture of a "Pondering Language Model". It shows a stacked arrangement of three identical "Pondering Language Model" blocks, each processing input embeddings and generating predicted probabilities. The diagram highlights the flow of data through these blocks, emphasizing the use of "Pondering Embeddings" and "Word Embeddings" at various stages.
### Components/Axes
The diagram consists of the following key components:
* **Pondering Language Model:** Represented as rectangular blocks, these are the core processing units. There are three stacked instances.
* **Input Embedding:** The initial layer, providing input to the first "Pondering Language Model" block.
* **Word Embed:** Represented by circles with a plus sign inside, these are used within each "Pondering Language Model" block.
* **Pondering Embedding:** Represented by light blue rectangles, these are also used within each "Pondering Language Model" block.
* **Predicted Probability:** The output layer, displaying probability values.
* **Labels:** "Pondering Language Model", "Pondering Embedding", "Word Embed", "Input Embedding", "Predicted Probability", "xN", "Vocab Size", "Hidden Size".
* **Numerical Values:** Probability values are displayed within green and yellow rectangles at various points in the diagram.
### Detailed Analysis or Content Details
The diagram shows a three-layer architecture. Let's analyze each layer:
**Layer 1 (Bottom):**
* **Input Embedding:** Four input embeddings are shown.
* **Pondering Language Model:** The input embeddings feed into a "Pondering Language Model" block.
* **Word Embed:** Each input embedding is connected to a "Word Embed" component.
* **Pondering Embedding:** Each "Word Embed" is connected to a "Pondering Embedding" component.
* **Output:** The output of this layer consists of five values: 0.1, 0.2, 0.1, 0.3, 0.3. These values are displayed within green and yellow rectangles.
**Layer 2 (Middle):**
* **Pondering Language Model:** This layer receives input from the previous layer's "Pondering Language Model" block.
* **Word Embed:** Similar to Layer 1, each input is connected to a "Word Embed" component.
* **Pondering Embedding:** Each "Word Embed" is connected to a "Pondering Embedding" component.
* **Output:** The output of this layer consists of five values: 0.1, 0.2, 0.5, 0.1, 0.2. These values are displayed within green and yellow rectangles.
**Layer 3 (Top):**
* **Pondering Language Model:** This layer receives input from the previous layer's "Pondering Language Model" block.
* **Output:** The output of this layer is the "Predicted Probability", consisting of five values: 0.6, 0.2, 0.7, 0.1, 0.8. These values are displayed within green and yellow rectangles.
* **Additional Probabilities:** Below the predicted probabilities, there are additional values: 0.2, 0.5, 0.1, 0.4, 0.1.
**Annotations:**
* **xN:** Located near the "Word Embed" component, likely indicating the number of word embeddings.
* **Vocab Size:** Located near the "Word Embed" component, indicating the vocabulary size.
* **Hidden Size:** Located near the "Pondering Language Model" block, indicating the hidden size.
The connections between components are indicated by arrows. The plus sign within the "Word Embed" circles suggests an addition operation. The circles with the plus sign are connected to the rectangles representing the "Pondering Embeddings".
### Key Observations
* The architecture is repetitive, with three identical "Pondering Language Model" blocks stacked on top of each other.
* The probability values at the output layer ("Predicted Probability") are generally higher than the intermediate values.
* The diagram emphasizes the interplay between "Word Embeddings" and "Pondering Embeddings" within each layer.
* The values within the green and yellow rectangles appear to represent probabilities or activations.
### Interpretation
The diagram illustrates a deep learning architecture designed for language modeling. The "Pondering Language Model" blocks likely represent recurrent neural network (RNN) or transformer layers. The "Word Embeddings" convert words into vector representations, while the "Pondering Embeddings" might represent attention mechanisms or other contextual information. The stacking of these layers allows the model to learn hierarchical representations of language, ultimately predicting the probability of the next word or sequence of words.
The use of "Pondering Embeddings" suggests a mechanism for the model to focus on relevant parts of the input sequence. The numerical values represent the model's internal state and predictions at different stages of processing. The diagram provides a high-level overview of the architecture, without specifying the exact details of the layer implementations. The "xN", "Vocab Size", and "Hidden Size" annotations provide information about the model's dimensions and capacity. The diagram suggests a model capable of processing sequential data and generating probabilistic outputs, making it suitable for tasks such as text generation, machine translation, or sentiment analysis.
</details>
âŹ
class PonderingLanguageModel (nn. Module):
def __init__ (self, lm, v, h, k):
self. lm = lm # language model
self. vocab_size = v
self. hidden_dim = h
self. pondering_steps = k
self. embedding = nn. Parameter (torch. randn (v, h), requires_grad = True)
def forward (self, input_tokens):
input_embedding =
self. embedding [input_tokens]
#Iterative pondering
for t in range (self. pondering_steps):
predicted_prob = self. lm (input_embedding)
pondering_embedding = torch. matmul (predicted_prob, self. embedding)
input_embedding = input_embedding + pondering_embedding
#Final forward pass
final_prob = self. lm (input_embedding)
return final_prob
Figure 1: Overview of the Pondering Language Model. Given input token embeddings, the base LM produces a probability distribution over the vocabulary, which is used to compute a continuous âpondering embeddingâ via a weighted sum of all token embeddings. This embedding is then added residually to the original input embeddings and fed back into the LM. By repeating this process for $k$ steps within a single token prediction, the model iteratively refines its output distributions. The pseudocode on the right illustrates the implementation details.
## 2 Pondering Language Model
In this section, we introduce our proposed Pondering Language Model (Figure 1), which integrates a pondering mechanism into language models via pretraining. Given that pretraining fundamentally constitutes a language modeling task, we briefly review this task before detailing our proposed model.
Language Modeling. Given a sequence of tokens $X=[x_{1},x_{2},\dots,x_{n}]$ , the primary objective of language modeling is typically to maximize the likelihood of predicting each token based on its preceding tokens. Formally, this is expressed through the joint probability factorization:
$$
P(x_{1},x_{2},\dots,x_{n})=\prod_{t=1}^{n}P(x_{t}\mid x_{<t}) \tag{1}
$$
Current language models first map tokens to input embeddings $\mathbf{E}^{0}=[\mathbf{e}^{0}_{1},\mathbf{e}^{0}_{2},\dots,\mathbf{e}^{0}_{n}]$ , where each embedding $\mathbf{e}^{0}_{i}\in\mathbb{R}^{d}$ is selected from a vocabulary embedding matrix $\mathbf{V}=[\mathbf{e}_{1},\mathbf{e}_{2},\dots,\mathbf{e}_{|V|}]$ , with vocabulary size $|V|$ and hidden dimension $d$ . The language model then generates output probabilities $\mathbf{P}$ for predicting the next token at each position:
$$
\mathbf{P}=\text{LM}(\mathbf{E}^{0}),\quad\mathbf{P}\in\mathbb{R}^{n\times|V|} \tag{2}
$$
The cross-entropy loss is computed directly from these predicted probabilities $\mathbf{P}$ to pretrain the language model.
Pondering Mechanism. In our proposed method, instead of directly using the predicted output probabilities $\mathbf{P}$ to compute the cross-entropy loss, we utilize these probabilities as weights to sum embeddings of all candidate tokens, forming what we call a "pondering embedding". Given the probability distribution $\mathbf{p}\in\mathbb{R}^{|V|}$ at each position, the pondering embedding $\mathbf{t}$ is:
$$
\mathbf{t}=\sum_{i=1}^{|V|}p_{i}\mathbf{e}_{i},\quad p_{i}\in\mathbb{R},
\mathbf{e}_{i}\in\mathbb{R}^{d} \tag{3}
$$
For computational efficiency, pondering embeddings $\mathbf{t}$ for all positions can be calculated simultaneously via matrix multiplication In practice, we use only the top-K tokens with highest probabilities at each position to compute the pondering embedding, reducing complexity from $\mathcal{O}(n|V|d)$ to $\mathcal{O}(nKd)$ . With $K=100\ll|V|$ , this does not degrade LM performance and makes the matrix multiplication overhead negligible within the overall LM computations. :
$$
\mathbf{T}=\mathbf{P}\mathbf{V},\quad\mathbf{T}\in\mathbb{R}^{n\times d} \tag{4}
$$
Through these pondering embeddings, we effectively map predicted probabilities back into the embedding space, preserving embedding information from all possible candidate tokens. To maintain the information from the original input embeddings, we integrate the pondering embeddings using a residual connection:
$$
\mathbf{E}^{1}=\mathbf{E}^{0}+\mathbf{T}=[\mathbf{e}^{0}_{1}+\mathbf{t}_{1},
\mathbf{e}^{0}_{2}+\mathbf{t}_{2},\dots,\mathbf{e}^{0}_{n}+\mathbf{t}_{n}] \tag{5}
$$
We then feed the updated embeddings $\mathbf{E}^{1}$ back into the same language model to obtain refined output probabilities:
$$
\mathbf{P}^{1}=\text{LM}(\mathbf{E}^{1}),\quad\mathbf{P}^{1}\in\mathbb{R}^{n
\times|V|} \tag{6}
$$
After obtaining $\mathbf{P}^{1}$ , we can iteratively repeat the previous process to achieve multi-step pondering. Specifically, given a predefined number of pondering steps $k$ Unless otherwise specified, we set $k=3$ for subsequent experiments., we iteratively compute new pondering embeddings and integrate them with the original input embeddings using residual connections, feeding the result back into the same language model until $k$ steps are reached:
$$
\displaystyle\mathbf{E}^{0} \displaystyle=[\mathbf{e}^{0}_{1},\mathbf{e}^{0}_{2},\dots,\mathbf{e}^{0}_{n}]
,\quad\mathbf{P}^{0}=\text{LM}(\mathbf{E}^{0}),\quad\mathbf{T}^{1}=\mathbf{P}^
{0}\mathbf{V} \displaystyle\mathbf{E}^{1} \displaystyle=\mathbf{E}^{0}+\mathbf{T}^{1}=[\mathbf{e}^{0}_{1}+\mathbf{t}^{1}
_{1},\mathbf{e}^{0}_{2}+\mathbf{t}^{1}_{2},\dots,\mathbf{e}^{0}_{n}+\mathbf{t}
^{1}_{n}],\quad\mathbf{P}^{1}=\text{LM}(\mathbf{E}^{1}),\quad\mathbf{T}^{2}=
\mathbf{P}^{1}\mathbf{V} \displaystyle\dots \displaystyle\mathbf{E}^{k} \displaystyle=\mathbf{E}^{0}+\sum_{i=1}^{k}\mathbf{T}^{i}=[\mathbf{e}^{0}_{1}+
\mathbf{t}^{1}_{1}+\dots+\mathbf{t}^{k}_{1},\dots,\mathbf{e}^{0}_{n}+\mathbf{t
}^{1}_{n}+\dots+\mathbf{t}^{k}_{n}],\quad\mathbf{P}^{k}=\text{LM}(\mathbf{E}^{
k}) \tag{7}
$$
This iterative pondering mechanism progressively refines the modelâs predictions. Finally, we can use the refined output probabilities $\mathbf{P}^{k}$ after $k$ pondering steps to compute the cross-entropy loss and optimize the language model to perform $k$ -step pondering.
## 3 Experiments
Our experiments consist of four parts. First, we validate the scaling curves of pondering models on widely used GPT-2 and LLaMA architectures. Second, we perform large-scale pretraining of PonderingPythia models on the Pile dataset and compare their scaling curves and language modeling capabilities with those of the official Pythia suite (Biderman et al., 2023). Third, we evaluate the downstream task performance of PonderingPythia models, including nine popular general tasks and an instruction-following task, and compare the results with official Pythia, OPT (Zhang et al., 2022), Bloom (Le Scao et al., 2023), GPT-Neo (Black et al., 2021), and TinyLLaMA (Zhang et al., 2024) models. Finally, we investigate the impact of the number of pondering steps on pretraining perplexity.
### 3.1 Small Scale Validation on GPT-2 and LLaMA
We apply our proposed method to two popular Transformer architectures, GPT-2 and LLaMA, to investigate its general applicability and effectiveness. Specifically, we plot and compare scaling curves between vanilla GPT-2, vanilla LLaMA, and their corresponding pondering versions, referred to as PonderingGPT and PonderingLLaMA, respectively.
Experimental Settings. We train all models from scratch on a subset of the Pile dataset with the same tokenizer. The model sizes range from 405M to 1.4B parameters, with a context length fixed at 2048 tokens. The number of training tokens for each model is chosen to approximately match the scaling laws proposed by Chinchilla (Hoffmann et al., 2022b).
The detailed configurations of the models, including sizes, learning rates, and batch sizes, are specified in Table 1. These hyperparameters primarily follow the GPT-3 specifications (Brown et al., 2020). However, unlike GPT-3, we untie the input and output embedding matrices.
Table 1: Model sizes and hyperparameters for scaling experiments.
| params | $\mathrm{n_{layers}}$ | $\mathrm{d_{model}}$ | $\mathrm{n_{heads}}$ | learning rate | batch size (in tokens) | tokens |
| --- | --- | --- | --- | --- | --- | --- |
| 405M | 24 | 1024 | 16 | 3e-4 | 0.5M | 7B |
| 834M | 24 | 1536 | 24 | 2.5e-4 | 0.5M | 15B |
| 1.4B | 24 | 2048 | 32 | 2e-4 | 0.5M | 26B |
<details>
<summary>x2.png Details</summary>

### Visual Description
\n
## Chart: Loss vs. Model Size
### Overview
The image presents a line chart comparing the loss values of different language models (GPT and LLaMA) with and without "Pondering" as a function of model size. The chart also includes a smaller subplot showing the change in loss (ÎLoss) for each model.
### Components/Axes
* **X-axis:** Model Size, labeled with values: 405M\*7B, 834M\*15B, 1.4B\*26B.
* **Y-axis (Main Chart):** Loss, ranging from approximately 2.1 to 2.7.
* **Y-axis (Subplot):** ÎLoss (Change in Loss), ranging from approximately 0 to 0.15.
* **Legend (Top-Right):**
* Blue Dashed Line: GPT fitted
* Orange Dashed Line: LLaMA fitted
* Blue Solid Line with Circle Markers: GPT
* Green Solid Line with Circle Markers: LLaMA
* Blue Solid Line with Star Markers: Pondering GPT
* Orange Solid Line with Star Markers: Pondering LLaMA
* **Text Annotations:** "2.01x" and "2.26x" are annotated on the chart, indicating relative scaling factors.
### Detailed Analysis or Content Details
**Main Chart:**
* **GPT (Blue Circles):** The line slopes downward, indicating decreasing loss with increasing model size.
* At 405M\*7B: Loss â 2.62
* At 834M\*15B: Loss â 2.38
* At 1.4B\*26B: Loss â 2.23
* **LLaMA (Green Circles):** The line also slopes downward, but is generally above the GPT line.
* At 405M\*7B: Loss â 2.52
* At 834M\*15B: Loss â 2.32
* At 1.4B\*26B: Loss â 2.18
* **Pondering GPT (Blue Stars):** The line slopes downward, starting above the GPT line and converging towards it at larger model sizes.
* At 405M\*7B: Loss â 2.48
* At 834M\*15B: Loss â 2.30
* At 1.4B\*26B: Loss â 2.20
* **Pondering LLaMA (Orange Stars):** The line slopes downward, starting above the LLaMA line and converging towards it at larger model sizes.
* At 405M\*7B: Loss â 2.43
* At 834M\*15B: Loss â 2.26
* At 1.4B\*26B: Loss â 2.15
* **Fitted Lines (Dashed):** The dashed lines represent the fitted curves for GPT (blue) and LLaMA (orange). They generally follow the trend of the corresponding solid lines.
**Subplot (ÎLoss):**
* L<sub>GPT</sub> (Black Dashed Line): Relatively flat, with a slight downward trend. ÎLoss â 0.05 - 0.07
* L<sub>LLaMA</sub> (Black Solid Line with Triangle Markers): Relatively flat, with a slight downward trend. ÎLoss â 0.05 - 0.08
* L<sub>PonderingGPT</sub> (Red Solid Line with Triangle Markers): Relatively flat, with a slight downward trend. ÎLoss â 0.08 - 0.10
* L<sub>PonderingLLaMA</sub> (Orange Solid Line with Triangle Markers): Relatively flat, with a slight downward trend. ÎLoss â 0.05 - 0.07
### Key Observations
* Both GPT and LLaMA exhibit decreasing loss as model size increases.
* GPT consistently has lower loss values than LLaMA across all model sizes.
* "Pondering" initially increases loss but appears to converge towards the non-pondering models as size increases.
* The ÎLoss subplot shows that the change in loss is relatively small across the tested model sizes for all configurations.
* The annotations "2.01x" and "2.26x" likely represent the scaling factor of loss reduction between model sizes for LLaMA and GPT respectively.
### Interpretation
The chart demonstrates the impact of model size on loss for GPT and LLaMA architectures, both with and without the "Pondering" mechanism. The consistent downward trend in loss for both models indicates that increasing model size generally improves performance (reduces loss). The lower loss values for GPT suggest that it is a more efficient model than LLaMA for the given task.
The "Pondering" mechanism appears to introduce a slight initial increase in loss, but its effect diminishes as the model size grows. This could indicate that "Pondering" is more beneficial for smaller models or requires larger models to fully realize its potential. The ÎLoss subplot confirms that the change in loss is relatively small, suggesting that the benefits of increasing model size may plateau beyond a certain point.
The annotations "2.01x" and "2.26x" provide a quantitative measure of the loss reduction achieved by increasing model size. These values suggest that GPT experiences a slightly larger reduction in loss per unit increase in model size compared to LLaMA.
The chart provides valuable insights into the trade-offs between model size, loss, and the effectiveness of the "Pondering" mechanism. This information can be used to guide the development and deployment of language models for specific applications.
</details>
Figure 2: (top) Scaling curves of GPT3 LLaMA and their corresponding pondering models. (bottom) Relative improvements of RoPE + RMSNorm + SwiGLU MLP and Pondering.
Results. Figure 2 (top) illustrates the scaling curves of the validation loss on the Pile subset for vanilla GPT-2, vanilla LLaMA, and their pondering counterparts. Our results show that incorporating pondering significantly improves the performance for both GPT-2 and LLaMA across the entire size range tested (405M to 1.4B parameters). For instance, by fitting scaling curves, we find that PonderingGPT-834M achieves a loss comparable to a vanilla GPT-2 model trained with approximately $2.01\times$ parameters $*$ tokens, while PonderingLLaMA-834M matches the loss of vanilla LLaMA trained with roughly $2.26\times$ parameters $*$ tokens.
Furthermore, Figure 2 (bottom) shows the relative validation loss improvements (denoted as $\Delta$ loss) of architectural modifications inherent in LLaMAânamely rotary positional embeddings (RoPE) (Su et al., 2024), RMSNorm (Zhang & Sennrich, 2019), and SwiGLU activationâin comparison to GPT-2, alongside the relative improvements obtained by our pondering method for both architectures. We observe that while the relative improvement due to RoPE, RMSNorm, and SwiGLU MLP diminishes as model size and data scale increase (reaching approximately 0.02 for the 1.4B parameter model), the relative improvement provided by pondering consistently remains around 0.1 across the entire tested scale.
<details>
<summary>x3.png Details</summary>

### Visual Description
\n
## Line Chart: Loss vs. Number of Parameters for Pythia and PonderingPythia
### Overview
This chart displays the relationship between the number of parameters in two language models, Pythia and PonderingPythia, and their corresponding loss values. The x-axis represents the number of parameters on a logarithmic scale, while the y-axis represents the loss. The chart uses lines and data points to illustrate how loss decreases as the number of parameters increases for both models.
### Components/Axes
* **X-axis Title:** "#Parameters (log scale)"
* **X-axis Markers:** 200M, 500M, 1B, 2B, 3B, 7B (representing millions and billions of parameters)
* **Y-axis Title:** "Loss"
* **Y-axis Scale:** Ranges from approximately 1.8 to 2.6.
* **Legend:** Located at the top-right of the chart.
* **Pythia:** Represented by a dark blue line and blue data points.
* **PonderingPythia:** Represented by a dark green line and green data points.
* **Annotation:** "37% params" is placed near the 1B parameter mark, likely indicating the parameter size of PonderingPythia relative to Pythia.
### Detailed Analysis
**Pythia (Dark Blue Line & Points):**
The Pythia line slopes downward, indicating that loss decreases as the number of parameters increases.
* At 200M parameters: Loss is approximately 2.55.
* At 500M parameters: Loss is approximately 2.20.
* At 1B parameters: Loss is approximately 2.00.
* At 2B parameters: Loss is approximately 1.92.
* At 3B parameters: Loss is approximately 1.89.
* At 7B parameters: Loss is approximately 1.86.
**PonderingPythia (Dark Green Line & Points):**
The PonderingPythia line also slopes downward, but the decrease in loss is less pronounced than for Pythia.
* At 200M parameters: Loss is approximately 2.30.
* At 500M parameters: Loss is approximately 2.10.
* At 1B parameters: Loss is approximately 1.95.
* At 2B parameters: Loss is approximately 1.89.
* At 3B parameters: Loss is approximately 1.87.
* At 7B parameters: Loss is approximately 1.86.
### Key Observations
* Both models exhibit a decreasing loss as the number of parameters increases, suggesting that larger models generally perform better (lower loss).
* Pythia consistently has a higher loss than PonderingPythia across all parameter sizes.
* The difference in loss between the two models appears to diminish as the number of parameters increases, converging around 7B parameters.
* The annotation "37% params" suggests that PonderingPythia uses approximately 37% of the parameters compared to Pythia at the 1B parameter level.
### Interpretation
The chart demonstrates the scaling behavior of two language models. Increasing the number of parameters generally leads to improved performance (lower loss). However, PonderingPythia achieves comparable or better performance than Pythia with significantly fewer parameters. This suggests that PonderingPythia is a more parameter-efficient model, potentially due to architectural differences or training strategies. The convergence of the loss curves at higher parameter counts indicates that the benefits of increasing parameters may diminish beyond a certain point. The "37% params" annotation is crucial; it highlights the efficiency of PonderingPythia, implying it achieves similar results to Pythia with a much smaller model size. This has implications for computational cost, memory requirements, and deployment feasibility. The logarithmic scale on the x-axis is important because it visually represents the exponential increase in parameters, making it easier to compare the models' performance across a wide range of sizes.
</details>
<details>
<summary>x4.png Details</summary>

### Visual Description
\n
## Line Chart: Loss vs. Training Tokens for Language Models
### Overview
This chart displays the loss function values for two language models, Pythia-1B and PonderingPythia-1B, as a function of the number of training tokens. The x-axis represents the number of training tokens on a logarithmic scale, while the y-axis represents the loss. The chart shows how the loss decreases as the models are trained with more tokens. A horizontal line at approximately 2.05 indicates a point of comparison, labeled "41% training tokens".
### Components/Axes
* **X-axis:** "#Training tokens (log scale)". The scale ranges from approximately 60B to 280B tokens. The axis markers are 60B, 100B, 200B, and 280B.
* **Y-axis:** "Loss". The scale ranges from approximately 1.95 to 2.25. The axis markers are 1.95, 2.00, 2.05, 2.10, 2.15, 2.20, and 2.25.
* **Legend:** Located in the top-right corner.
* Blue line: "Pythia-1B"
* Gray line: "PonderingPythia-1B"
* Blue circles: "Pythia-1B"
* Green circles: "PonderingPythia-1B"
* **Annotation:** "41% training tokens" is written horizontally near the gray line, at a loss value of approximately 2.05.
### Detailed Analysis
**Pythia-1B (Blue Line & Circles):**
The blue line representing Pythia-1B slopes downward, indicating a decrease in loss as the number of training tokens increases.
* At 60B tokens: Loss â 2.21
* At 100B tokens: Loss â 2.16
* At 200B tokens: Loss â 2.11
* At 280B tokens: Loss â 2.07
**PonderingPythia-1B (Green Line & Circles):**
The green line representing PonderingPythia-1B also slopes downward, but is steeper than the blue line, indicating a faster decrease in loss.
* At 60B tokens: Loss â 2.12
* At 100B tokens: Loss â 2.06
* At 200B tokens: Loss â 2.00
* At 280B tokens: Loss â 1.96
The horizontal line at approximately 2.05 serves as a reference point. PonderingPythia-1B crosses this line at approximately 100B tokens, while Pythia-1B crosses it at approximately 200B tokens.
### Key Observations
* PonderingPythia-1B consistently exhibits lower loss values than Pythia-1B across all training token counts.
* The rate of loss reduction is greater for PonderingPythia-1B, suggesting it learns more efficiently from the training data.
* Both models demonstrate diminishing returns in loss reduction as the number of training tokens increases, as the slope of the lines becomes less steep at higher token counts.
* The "41% training tokens" annotation suggests a specific point of comparison or evaluation within the training process.
### Interpretation
The chart demonstrates the impact of training data size on the performance of two language models. PonderingPythia-1B consistently outperforms Pythia-1B, indicating that the "Pondering" mechanism (whatever that may be) contributes to more effective learning. The logarithmic scale on the x-axis highlights the importance of initial training stages, where loss reduction is most significant. The diminishing returns observed at higher token counts suggest that there is a point beyond which adding more training data yields progressively smaller improvements in model performance. The annotation "41% training tokens" could represent a point where the models have reached a certain level of convergence or a specific evaluation milestone. The difference in loss between the two models suggests that PonderingPythia-1B may be more robust or generalize better to unseen data.
</details>
<details>
<summary>x5.png Details</summary>

### Visual Description
\n
## Line Chart: Loss vs. FLOPS for Pythia Models
### Overview
This image presents a line chart comparing the loss values of two Pythia models â "Vanilla Pythia-70M" and "PonderingPythia-70M" â as a function of FLOPS (EFLOPs). The chart illustrates how the loss decreases for both models as the computational effort (FLOPS) increases.
### Components/Axes
* **X-axis:** FLOPS (EFLOPs), ranging from approximately 80 to 420.
* **Y-axis:** Loss, ranging from approximately 2.54 to 2.80.
* **Data Series 1:** Vanilla Pythia-70M (represented by a blue line with circular markers).
* **Data Series 2:** PonderingPythia-70M (represented by a green line with circular markers).
* **Legend:** Located in the top-right corner, clearly labeling each data series with its corresponding color.
### Detailed Analysis
**Vanilla Pythia-70M (Blue Line):**
The blue line exhibits a downward trend, indicating decreasing loss with increasing FLOPS.
* At approximately 80 FLOPS, the loss is around 2.79.
* At approximately 160 FLOPS, the loss is around 2.76.
* At approximately 240 FLOPS, the loss is around 2.74.
* At approximately 320 FLOPS, the loss is around 2.73.
* At approximately 400 FLOPS, the loss is around 2.72.
**PonderingPythia-70M (Green Line):**
The green line also shows a downward trend, but the loss decreases more rapidly than that of the Vanilla Pythia-70M.
* At approximately 80 FLOPS, the loss is around 2.67.
* At approximately 160 FLOPS, the loss is around 2.61.
* At approximately 240 FLOPS, the loss is around 2.59.
* At approximately 320 FLOPS, the loss is around 2.57.
* At approximately 400 FLOPS, the loss is around 2.55.
### Key Observations
* PonderingPythia-70M consistently exhibits lower loss values than Vanilla Pythia-70M across the entire range of FLOPS.
* The rate of loss reduction is steeper for PonderingPythia-70M, especially at lower FLOPS values.
* Both models demonstrate diminishing returns in loss reduction as FLOPS increase beyond 320.
### Interpretation
The chart suggests that PonderingPythia-70M is a more efficient model than Vanilla Pythia-70M, achieving lower loss values for a given amount of computational effort (FLOPS). The steeper initial decline in loss for PonderingPythia-70M indicates that it benefits more from increased computational resources, particularly at lower FLOPS levels. The diminishing returns observed at higher FLOPS suggest that there's a point where further increasing FLOPS yields only marginal improvements in model performance (loss reduction). This could be due to factors like model saturation or the limitations of the training data. The data implies that PonderingPythia-70M is a better choice when computational resources are limited, or when maximizing performance within a specific FLOPS budget is crucial.
</details>
Figure 3: Language modeling loss when scaling parameter count and training tokens. PonderingPythia achieves comparable performance to the official Pythia-6.9B while using only 37% of the parameters, and matches the performance of the official Pythia-1B with just 41% of the training tokens.
Table 2: Language Modeling Perplexity (ppl). Lower is better. The values in parentheses indicate the improvement ( $\downarrow$ ) compared to the corresponding Pythia model. For simplicity, we refer to PonderingPythia as Ponder.
| Model | Pile | Wikitext | Lambada Openai | Lambada Standard |
| --- | --- | --- | --- | --- |
| Pythia-70M | $18.36$ | $57.03$ | $141.94$ | $967.72$ |
| Ponder-70M | $12.68$ ( $\downarrow$ 5.68) | $34.07$ ( $\downarrow$ 22.96) | $39.19$ ( $\downarrow$ 102.75) | $145.51$ ( $\downarrow$ 822.21) |
| Pythia-160M | $12.55$ | $33.43$ | $38.15$ | $186.51$ |
| Ponder-160M | $9.87$ ( $\downarrow$ 2.68) | $23.69$ ( $\downarrow$ 9.74) | $15.96$ ( $\downarrow$ 22.19) | $41.24$ ( $\downarrow$ 145.27) |
| Pythia-410M | $8.85$ | $20.11$ | $10.86$ | $31.54$ |
| Ponder-410M | $8.15$ ( $\downarrow$ 0.70) | $17.78$ ( $\downarrow$ 2.33) | $8.55$ ( $\downarrow$ 2.31) | $17.48$ ( $\downarrow$ 14.06) |
| Pythia-1B | $7.85$ | $16.45$ | $7.91$ | $17.41$ |
| Ponder-1B | $7.15$ ( $\downarrow$ 0.70) | $14.61$ ( $\downarrow$ 1.84) | $6.02$ ( $\downarrow$ 1.89) | $10.64$ ( $\downarrow$ 6.77) |
| Pythia-1.4B | $7.24$ | $14.72$ | $6.08$ | $10.87$ |
| Ponder-1.4B | $6.82$ ( $\downarrow$ 0.42) | $13.59$ ( $\downarrow$ 1.13) | $5.37$ ( $\downarrow$ 0.71) | $8.88$ ( $\downarrow$ 1.99) |
| Pythia-2.8B | $6.63$ | $12.69$ | $5.04$ | $8.23$ |
| Ponder-2.8B | $6.21$ ( $\downarrow$ 0.42) | $11.74$ ( $\downarrow$ 0.95) | $4.15$ ( $\downarrow$ 0.89) | $6.17$ ( $\downarrow$ 2.06) |
| Pythia-6.9B | $6.29$ | $11.41$ | $4.45$ | $6.95$ |
### 3.2 Large-Scale Pretraining on Pile
We further validate the effectiveness of our pondering method by conducting large-scale pretraining experiments on the entire Pile dataset (300B tokens) (Gao et al., 2020). We train a new model, named PonderingPythia, from scratch using exactly the same architectural components (parallel attention and MLP layers, rotary embeddings with 1/4 head dimensions), same tokenizer and training hyperparameters (optimizer settings, learning rate schedule, batch size, and context length) as the original Pythia models. We then compare PonderingPythia modelsâ scaling curves and language modeling capabilities with the official Pythia model suite.
#### 3.2.1 Scaling Curves
We plot the PonderingPythia and official Pythia modelâs scaling curves along parameters size, training tokens and training flops. As depicted in Figure 3 (left), the fitted curves show that a 2.5B-parameter PonderingPythia model achieves a validation loss comparable to the 6.9B-parameter official Pythia model, while requiring only 37% of the parameters. In Figure 3 (middle), we evaluated the 1B PonderingPythia and official Pythia models at intervals of 20B training tokens. The fitted curves indicate that PonderingPythia trained with 115B tokens achieves performance comparable to the official Pythia model trained with 280B tokens, consuming only 41% of the training tokens. In Figure 3 (right), we report the language modeling loss of vanilla Pythia-70M To match the training FLOPs of vanilla Pythia-70M with PonderingPythia-70M, we trained the vanilla model for 4 epochs. and PonderingPythia-70M under the same computational budget during pretraining. It can be observed that PonderingPythia-70M consistently outperforms vanilla Pythia-70M when trained with the same number of FLOPs.
#### 3.2.2 Language Modeling Ability Evaluation
We measure the perplexity on several language modeling datasets to reflect general language modeling capabilities. Specifically, we report perplexity scores on the Pile validation set, Wikitext, and the Lambada dataset (both OpenAI and standard versions), detailed in Table 2. The results demonstrate significant perplexity improvements across all datasets and model sizes. Notably, the perplexity achieved by the PonderingPythia-2.8B model is even better than that of the official Pythia-6.8B model, which is nearly 2.5 times larger.
### 3.3 Downstream Tasks Evaluation
We use the PonderingPythia models pretrained in the previous subsection to conduct downstream task evaluations.
<details>
<summary>x6.png Details</summary>

### Visual Description
\n
## Radar Chart: Model Performance Comparison
### Overview
This image presents a radar chart comparing the performance of two models, "Pythia-1B" and "PonderingPythia-1B", across seven different categories. The chart uses a radial layout with each axis representing a category, and the distance from the center indicating the score.
### Components/Axes
* **Axes/Categories:** The radar chart has seven axes, each labeled with a different skill or domain:
* Writing
* Roleplay
* Reasoning
* Math
* Coding
* Extraction
* STEM
* Humanities
* **Radial Scale:** The radial scale ranges from 0 to 3, with markers at 0, 0.5, 1, 1.5, 2, 2.5, and 3.
* **Legend:** Located at the bottom-right of the chart, the legend identifies the two data series:
* **Pythia-1B:** Represented by a blue line, with an average score of 1.89.
* **PonderingPythia-1B:** Represented by a green line, with an average score of 2.22.
### Detailed Analysis
The chart displays two polygonal lines representing the performance of each model across the seven categories.
**Pythia-1B (Blue Line):**
The line generally fluctuates between approximately 1.2 and 2.5.
* **Writing:** Approximately 2.3
* **Roleplay:** Approximately 2.1
* **Reasoning:** Approximately 1.7
* **Math:** Approximately 1.5
* **Coding:** Approximately 1.2
* **Extraction:** Approximately 1.3
* **STEM:** Approximately 1.6
* **Humanities:** Approximately 1.8
**PonderingPythia-1B (Green Line):**
The line generally fluctuates between approximately 1.7 and 2.8.
* **Writing:** Approximately 2.8
* **Roleplay:** Approximately 2.6
* **Reasoning:** Approximately 2.3
* **Math:** Approximately 2.0
* **Coding:** Approximately 1.7
* **Extraction:** Approximately 2.1
* **STEM:** Approximately 2.2
* **Humanities:** Approximately 2.4
### Key Observations
* **Overall Performance:** PonderingPythia-1B consistently outperforms Pythia-1B across all categories, as indicated by the higher average score (2.22 vs. 1.89) and the generally outward positioning of the green line.
* **Strengths:** PonderingPythia-1B demonstrates particularly strong performance in Writing (approximately 2.8) and Roleplay (approximately 2.6).
* **Weaknesses:** Pythia-1B shows relatively lower performance in Coding (approximately 1.2) and Math (approximately 1.5).
* **Shape:** Both lines have a somewhat irregular shape, indicating varying levels of performance across the different categories.
### Interpretation
The radar chart effectively visualizes the comparative strengths and weaknesses of the two models. PonderingPythia-1B appears to be a more versatile model, achieving higher scores across a broader range of tasks. The chart suggests that PonderingPythia-1B has been trained or fine-tuned to excel in areas like creative writing and interactive role-playing, while Pythia-1B may be relatively weaker in these domains. The differences in performance could be attributed to variations in the training data, model architecture, or training methodology. The chart provides a clear and concise overview of the models' capabilities, allowing for informed decision-making regarding their application in specific tasks. The average scores provide a single metric for overall comparison, while the detailed category-specific scores offer a more nuanced understanding of their respective strengths and weaknesses.
</details>
<details>
<summary>x7.png Details</summary>

### Visual Description
\n
## Radar Chart: Model Performance Comparison
### Overview
This image presents a radar chart comparing the performance of two language models, "Pythia-1.4B" and "PonderingPythia-1.4B", across eight different categories. The chart uses a radial layout with each axis representing a specific skill or domain. The performance score ranges from 0 to 3.5.
### Components/Axes
* **Axes Labels (Categories):** Writing, Roleplay, Reasoning, Math, Coding, Extraction, STEM, Humanities. These are arranged clockwise around the chart.
* **Radial Scale:** The chart features a radial scale ranging from 0 to 3.5, with increments of 0.5. The scale is positioned at the center of the chart.
* **Legend:** Located at the bottom-right of the chart.
* **Pythia-1.4B:** Represented by a solid grey line. Average score: 2.2
* **PonderingPythia-1.4B:** Represented by a solid teal line. Average score: 2.75
### Detailed Analysis
The chart displays two polygonal lines representing the performance of each model across the eight categories.
**Pythia-1.4B (Grey Line):**
* **Writing:** Approximately 2.5
* **Roleplay:** Approximately 2.0
* **Reasoning:** Approximately 2.2
* **Math:** Approximately 1.8
* **Coding:** Approximately 2.0
* **Extraction:** Approximately 1.5
* **STEM:** Approximately 1.2
* **Humanities:** Approximately 2.8
The grey line shows a relatively consistent performance across categories, with a slight dip in STEM and Extraction, and a peak in Humanities.
**PonderingPythia-1.4B (Teal Line):**
* **Writing:** Approximately 3.2
* **Roleplay:** Approximately 3.0
* **Reasoning:** Approximately 3.0
* **Math:** Approximately 2.5
* **Coding:** Approximately 2.7
* **Extraction:** Approximately 2.2
* **STEM:** Approximately 2.5
* **Humanities:** Approximately 2.8
The teal line generally exhibits higher scores than the grey line across all categories. It shows a strong performance in Writing, Roleplay, and Reasoning, with a slight dip in Humanities.
### Key Observations
* PonderingPythia-1.4B consistently outperforms Pythia-1.4B across all evaluated categories.
* Both models show relatively lower performance in STEM and Extraction compared to other categories.
* Pythia-1.4B has its highest score in Humanities, while PonderingPythia-1.4B excels in Writing, Roleplay, and Reasoning.
* The average score for Pythia-1.4B is 2.2, while the average score for PonderingPythia-1.4B is 2.75, indicating an overall performance advantage for the latter.
### Interpretation
The radar chart effectively visualizes the comparative strengths and weaknesses of the two language models. The data suggests that PonderingPythia-1.4B is a more capable model overall, demonstrating superior performance in a wide range of tasks. The lower scores in STEM and Extraction for both models might indicate areas where further training or architectural improvements are needed. The difference in performance between the two models is particularly noticeable in areas like Writing, Roleplay, and Reasoning, suggesting that the modifications made to create PonderingPythia-1.4B have significantly enhanced its capabilities in these domains. The chart provides a clear and concise overview of the models' relative performance, facilitating informed decision-making regarding their application in various NLP tasks.
</details>
Figure 4: Instruction-following abilities evaluated on MT-Bench. PonderingPythia-1B and 1.4B consistently outperform their corresponding official Pythia models across all subtasks.
#### 3.3.1 General Downstream Tasks
We consider various widely-used benchmarks, including the tasks originally used by Pythia (LAMBADA (Paperno et al., 2016), PIQA (Bisk et al., 2020), WinoGrande (Sakaguchi et al., 2021), ARC-Easy and ARC-Challenge (Clark et al., 2018), SciQ (Welbl et al., 2017). Additionally, we include HellaSwag (Zellers et al., 2019) for commonsense reasoning and RACE (Lai et al., 2017) for reading comprehension.
We evaluate both zero-shot and five-shot learning performance using the LM evaluation harness (Gao et al., 2023). Detailed results are shown in Table 3. Across all evaluated model sizes, PonderingPythia consistently and significantly outperforms the official Pythia models, as well as comparable OPT, GPT-Neo, and Bloom models. Remarkably, with only 1/10 of the training data (300B tokens) and fewer parameters (1B vs. 1.1B), our PonderingPythia-1B achieves results comparable to, or even surpassing, TinyLlamaâwhich uses a more advanced LLaMA architecture and 3T tokens. PonderingPythia-2.8B also surpasses Pythia-6.9B, which is nearly 2.5 times its size.
Table 3: Five-shot and zero-shot evaluations on downstream NLP tasks. All pretrained model weights used for comparison are obtained from their official repositories. We refer to PonderingPythia as Ponder for simplicity. $\Delta$ acc is compared to the official Pythia models.
| Model (#training tokens) | Lambada OpenAI | ARC -E | Lambada Standard | ARC -C | Wino Grande | PIQA | Hella Swag | SciQ | RACE | Avg acc / $\Delta$ acc $\uparrow$ |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
| 5-shot | | | | | | | | | | |
| Pythia-70M (300B) | 11.9 | 36.7 | 9.2 | 17.1 | 50.5 | 58.7 | 26.7 | 57.8 | 25.1 | 32.6 |
| Ponder-70M (300B) | 28.4 | 44.6 | 21.1 | 18.6 | 52.2 | 62.4 | 28.4 | 75.7 | 26.9 | 39.8 / +7.2 |
| Pythia-160M (300B) | 24.9 | 44.7 | 19.0 | 18.4 | 50.4 | 63.5 | 28.6 | 76.4 | 27.8 | 39.3 |
| OPT-125M (300B) | 29.8 | 42.3 | 26.4 | 19.7 | 51.3 | 63.4 | 29.0 | 78.8 | 30.1 | 41.2 |
| GPTneo-125M (300B) | 29.3 | 43.9 | 23.0 | 19.1 | 51.9 | 63.9 | 28.5 | 79.9 | 27.9 | 40.8 |
| Ponder-160M (300B) | 39.9 | 50.1 | 31.2 | 22.1 | 51.2 | 65.8 | 31.9 | 85.4 | 30.1 | 45.3 / +6.0 |
| Pythia-410M (300B) | 43.9 | 54.7 | 32.8 | 22.3 | 53.4 | 68.0 | 33.8 | 88.9 | 30.4 | 47.6 |
| OPT-350M (300B) | 38.3 | 45.4 | 32.1 | 20.5 | 53.0 | 65.8 | 31.9 | 85.7 | 29.5 | 44.7 |
| Bloom-560M (366B) | 29.4 | 50.2 | 29.7 | 21.9 | 52.7 | 64.2 | 31.4 | 88.0 | 30.0 | 44.2 |
| Ponder-410M (300B) | 48.9 | 58.7 | 43.7 | 26.1 | 54.0 | 70.5 | 37.3 | 91.0 | 32.4 | 51.4 /+3.8 |
| Pythia-1B (300B) | 48.3 | 58.6 | 35.8 | 25.4 | 52.8 | 71.3 | 37.7 | 91.6 | 31.7 | 50.4 |
| OPT-1.3B (300B) | 54.0 | 60.4 | 49.0 | 26.9 | 56.9 | 72.4 | 38.5 | 91.8 | 35.4 | 52.7 |
| GPTneo-1.3B (300B) | 49.9 | 59.9 | 44.5 | 25.9 | 56.6 | 71.6 | 38.6 | 86.1 | 34.5 | 51.9 |
| Bloom-1.1B (366B) | 36.3 | 54.9 | 37.4 | 24.9 | 53.4 | 67.6 | 34.8 | 88.7 | 33.0 | 47.9 |
| Ponder-1B (300B) | 57.7 | 63.2 | 52.5 | 28.6 | 58.6 | 73.3 | 41.9 | 93.4 | 36.3 | 56.2 / +5.8 |
| Pythia-1.4B (300B) | 54.5 | 63.1 | 44.5 | 28.8 | 57.1 | 71.0 | 40.5 | 92.4 | 34.6 | 54.1 |
| Bloom-1.7B (366B) | 42.5 | 58.8 | 41.5 | 26.2 | 57.7 | 68.7 | 37.6 | 91.9 | 33.5 | 50.9 |
| Ponder-1.4B (300B) | 59.2 | 67.5 | 49.9 | 32.4 | 60.4 | 73.5 | 44.2 | 94.3 | 37.1 | 57.6 / +3.5 |
| Tinyllama-1.1B (3T) | 53.8 | 64.8 | 45.0 | 31.1 | 59.4 | 73.8 | 44.9 | 94.0 | 36.4 | 55.9 |
| OPT-2.7B (300B) | 60.2 | 64.7 | 55.0 | 29.8 | 62.2 | 75.1 | 46.1 | 93.0 | 37.5 | 58.2 |
| GPTneo-2.7B (300B) | 56.0 | 64.0 | 51.6 | 30.1 | 59.6 | 73.9 | 42.4 | 93.3 | 35.5 | 56.3 |
| Bloom-3B (366B) | 46.2 | 63.8 | 47.1 | 31.7 | 57.8 | 70.8 | 41.4 | 93.4 | 34.6 | 54.1 |
| Pythia-2.8B (300B) | 59.0 | 67.0 | 50.7 | 31.0 | 61.1 | 74.4 | 45.3 | 93.7 | 35.9 | 57.6 |
| Pythia-6.9B (300B) | 62.5 | 69.6 | 54.8 | 35.6 | 62.9 | 76.6 | 48.0 | 94.6 | 36.7 | 60.1 |
| Ponder-2.8B (300B) | 64.2 | 70.6 | 58.7 | 35.8 | 65.3 | 76.7 | 49.0 | 94.3 | 39.0 | 61.5 / +3.9 |
| 0-shot | | | | | | | | | | |
| Pythia-70M (300B) | 18.7 | 36.7 | 13.5 | 18.8 | 51.1 | 59.9 | 26.6 | 59.9 | 23.9 | 34.3 |
| Ponder-70M (300B) | 33.5 | 42.5 | 24.1 | 19.1 | 51.1 | 61.5 | 28.3 | 70.3 | 27.8 | 39.8 / +5.5 |
| Pythia-160M (300B) | 33.0 | 43.8 | 21.4 | 18.9 | 52.0 | 62.2 | 28.4 | 73.7 | 27.9 | 40.1 |
| OPT-125M (300B) | 37.9 | 43.5 | 29.0 | 19.0 | 50.3 | 62.8 | 29.2 | 75.2 | 30.1 | 41.9 |
| GPTneo-125M (300B) | 37.5 | 43.9 | 26.0 | 19.1 | 50.4 | 63.0 | 28.7 | 76.5 | 27.5 | 41.4 |
| Ponder-160M (300B) | 44.3 | 46.8 | 33.8 | 20.5 | 52.3 | 63.9 | 31.9 | 76.8 | 29.6 | 44.4 / +4.3 |
| Pythia-410M (300B) | 51.4 | 52.2 | 36.4 | 21.4 | 53.8 | 66.9 | 33.7 | 81.5 | 30.9 | 47.6 |
| OPT-350M (300B) | 45.2 | 44.0 | 35.8 | 20.7 | 52.3 | 64.5 | 32.0 | 74.9 | 29.8 | 44.4 |
| Bloom-560M (366B) | 34.3 | 47.5 | 33.3 | 22.4 | 51.5 | 63.8 | 31.5 | 80.3 | 30.5 | 43.9 |
| Ponder-410M (300B) | 56.9 | 51.9 | 45.3 | 22.6 | 56.0 | 68.7 | 37.0 | 81.4 | 33.8 | 50.4 / +2.8 |
| Pythia-1B (300B) | 55.9 | 56.8 | 42.0 | 24.2 | 52.5 | 70.5 | 37.7 | 83.3 | 32.7 | 50.6 |
| OPT-1.3B (300B) | 57.9 | 57.1 | 52.5 | 23.4 | 59.7 | 71.8 | 41.6 | 84.3 | 34.3 | 53.6 |
| GPTneo-1.3B (300B) | 57.1 | 56.2 | 45.3 | 23.2 | 55.0 | 71.2 | 38.6 | 86.1 | 34.5 | 51.9 |
| Bloom-1.1B (366B) | 42.6 | 51.5 | 42.9 | 23.6 | 54.9 | 67.3 | 34.5 | 83.6 | 32.6 | 48.2 |
| Ponder-1B (300B) | 62.3 | 60.5 | 51.9 | 27.0 | 56.5 | 72.2 | 41.8 | 87.4 | 35.4 | 55.0 / +4.4 |
| Pythia-1.4B (300B) | 61.6 | 60.4 | 49.7 | 25.9 | 57.5 | 70.8 | 40.4 | 86.4 | 34.1 | 54.1 |
| Bloom-1.7B (366B) | 46.2 | 56.4 | 44.5 | 23.7 | 56.8 | 68.5 | 37.5 | 85.0 | 33.2 | 50.2 |
| Ponder-1.4B (300B) | 65.2 | 62.0 | 53.8 | 27.0 | 60.1 | 72.6 | 44.0 | 89.0 | 35.2 | 56.5 / +2.4 |
| Tinyllama-1.1B (3T) | 58.8 | 60.3 | 49.3 | 28.0 | 59.0 | 73.3 | 45.0 | 88.9 | 36.4 | 55.4 |
| OPT-2.7B (300B) | 63.5 | 60.8 | 56.0 | 26.8 | 61.2 | 73.8 | 45.9 | 85.8 | 36.2 | 56.7 |
| GPTneo-2.7B (300B) | 62.1 | 61.2 | 51.6 | 27.5 | 57.8 | 72.3 | 42.7 | 89.3 | 35.1 | 55.5 |
| Bloom-3B (366B) | 51.7 | 59.4 | 50.9 | 28.0 | 58.7 | 70.8 | 41.4 | 88.8 | 35.2 | 53.9 |
| Pythia-2.8B (300B) | 64.6 | 64.4 | 54.3 | 29.5 | 60.2 | 73.8 | 45.4 | 88.5 | 34.9 | 57.3 |
| Pythia-6.9B (300B) | 67.2 | 67.3 | 55.9 | 31.4 | 61.0 | 75.2 | 48.1 | 89.3 | 36.9 | 59.1 |
| Ponder-2.8B (300B) | 68.9 | 66.5 | 60.8 | 32.5 | 63.6 | 75.0 | 48.6 | 91.0 | 36.5 | 60.4 / +3.1 |
#### 3.3.2 Instruction-following Ability Evaluation
To assess the instruction-following capability of our model, we further fine-tuned PonderingPythia-1B and 1.4B, as well as the corresponding official Pythia models, on the Alpaca dataset using the same settings (Taori et al., 2023). The fine-tuned models were evaluated with MT-Bench (Zheng et al., 2023), a popular multi-turn question benchmark. The experimental results are shown in Figure 4. As illustrated, both PonderingPythia-1B and 1.4B consistently outperform their official Pythia counterparts across all subtasks, achieving average improvements of 0.33 and 0.55, respectively. The marginal gains on Coding and Math tasks may be attributed to the limited coding and mathematical abilities of the Pythia models.
<details>
<summary>x8.png Details</summary>

### Visual Description
\n
## Line Chart: Loss vs. Pondering Steps
### Overview
The image presents a line chart illustrating the relationship between "Pondering Steps" and "Loss". The chart shows a decreasing trend, indicating that as the number of pondering steps increases, the loss value decreases.
### Components/Axes
* **X-axis:** "Pondering Steps (0 = Baseline)". The axis is marked with values 0, 1, 3, 5, and 10.
* **Y-axis:** "Loss". The axis ranges from approximately 2.58 to 2.72.
* **Data Series:** A single green line representing the loss value at each pondering step.
* **Data Points:** Circular markers are placed at each data point on the line.
### Detailed Analysis
The green line slopes downward from left to right, indicating a negative correlation between pondering steps and loss.
* **Pondering Steps = 0:** Loss â 2.715
* **Pondering Steps = 1:** Loss â 2.685
* **Pondering Steps = 3:** Loss â 2.66
* **Pondering Steps = 5:** Loss â 2.615
* **Pondering Steps = 10:** Loss â 2.59
The decrease in loss is most significant between 0 and 3 pondering steps. The rate of decrease slows down as the number of pondering steps increases beyond 5.
### Key Observations
The chart demonstrates a clear diminishing return effect. The initial increase in pondering steps leads to a substantial reduction in loss, but the benefit of additional pondering steps diminishes as the loss approaches a plateau.
### Interpretation
This data suggests that the "pondering" process is effective in reducing loss, but there's a point of diminishing returns. The initial stages of pondering (0 to 3 steps) yield the most significant improvements. Beyond 5 steps, the reduction in loss becomes marginal. This could indicate that the model reaches a point where further pondering doesn't provide substantial new information or refinement. The baseline (0 pondering steps) represents the initial loss value before any pondering is applied. The chart is likely evaluating the effectiveness of a reasoning or decision-making process where "pondering steps" represent iterations of analysis or refinement.
</details>
Figure 5: Increasing the number of pondering steps consistently reduces loss on the Pile validation set.
### 3.4 Impact of Pondering Steps
To further investigate the effect of pondering steps on model performance, we trained several 160M-parameter Pythia models from scratch with different numbers of pondering steps: 0 (baseline), 1, 3, 5, and 10. These models were pretrained on a 30B-token subset of the Pile dataset. The results, presented in Figure 5, clearly show that increasing the number of pondering steps consistently reduces the language modeling loss on the Pile validation set. This demonstrates the significant potential and scalability of our method, suggesting that model performance can be further improved by increasing the depth of pondering steps.
## 4 Limitations and Future Work
### 4.1 Limitations
There are two limitations to our work. Firstly, due to computational constraints, we only scaled our method up to a 2.8B-parameter model trained on 300B tokens from the Pile dataset. It would be interesting to extend our approach to larger models and larger-scale datasets in the future. Secondly, although our results demonstrate that the proposed method scales better than vanilla models under the same training FLOPs (Figure 3), it also introduces additional inference overhead (increasing roughly linearly with the number of pondering steps), similar to test-time scaling methods.
### 4.2 Future Work
There are several promising directions for future work. Firstly, our proposed method is not limited to decoder-only architectures or language modeling; it has the potential to be applied to a wide range of model types and domains. For example, it could be adapted to state-space models such as Mamba (Gu & Dao, 2023), encoder models, or RWKV (Peng et al., 2023), as well as extended to other areas. Another promising direction is the introduction of token-adaptive pondering, which may significantly reduce computation and further enhance our method. It would also be interesting to investigate the interpretability of the pondering process, such as how the model "thinks" during pondering, the semantics of the pondering embedding, and whether the model learns to reflect on its predictions through pondering. Finally, exploring the combination of our method with orthogonal approaches such as CoT and other test-time scaling methods could also be an interesting direction.
## 5 Related Work
### 5.1 Test-Time Compute Scaling
Scaling test-time computation has emerged as an influential approach, providing an effective alternative to merely increasing model parameters (Snell et al., 2024). Current methods for enhancing test-time computation primarily fall into two categories: parallel scaling and sequential scaling (Zeng et al., 2024; Muennighoff et al., 2025).
Parallel scaling methods involve generating multiple candidate solutions simultaneously and selecting the best solution based on certain evaluation criteria. Prominent examples of this paradigm include Best-of-N search (Cobbe et al., 2021; Sun et al., 2024; Gui et al., 2024; Amini et al., 2024; Sessa et al., 2024) and Majority Vote (Wang et al., 2022). The primary difference among these parallel approaches lies in their strategy for selecting the optimal outcome. Nevertheless, parallel scaling approaches often encounter difficulty accurately identifying the best candidate solution among multiple generated possibilities (Stroebl et al., 2024; Hassid et al., 2024).
Sequential scaling methods, conversely, focus on progressively refining the modelâs reasoning through multiple iterative steps. Representative methods include CoT (Wei et al., 2022; Nye et al., 2021), iterative rethinking and revision (Huang et al., 2022; Min et al., 2024; Madaan et al., 2024; Wang et al., 2024b, b; Lee et al., 2025; Hou et al., 2025; Muennighoff et al., 2025; Li et al., 2025). Following the emergence of powerful OpenAI o1 and DeepSeek R1, leveraging extensive chain-of-thought prompting to scale test-time computation has become increasingly popular. Nonetheless, existing methods often come with limitations, including reliance on specially curated datasets (Allen-Zhu & Li, 2023), extended context windows (Zhu et al., 2025), and complex reinforcement learning strategies (Pang et al., 2025). In contrast, our method effectively circumvents these limitations.
### 5.2 Latent Thinking in Language Models
Latent thinking in LMs typically refers to the hidden computational processes within transformer architectures (Yang et al., 2024; Biran et al., 2024). Prior works exploring latent thinking can generally be divided into two categories based on their implementation methods:
Adding Additional Tokens: Wang et al. (2024a) proposed predicting a planning token as a discrete latent variable prior to generating reasoning steps. Similarly, Pfau et al. (2024) investigated filler tokens (e.g., ââŠâ) and found them effective for parallelizable problems, although they also highlighted limitations compared to CoT methods, potentially constraining scalability to more complex reasoning tasks. Zhou et al. (2024) pretrained models by randomly inserting a learnable token into the training corpus, demonstrating improvements across various tasks. Furthermore, Quiet-STaR employs reinforcement learning to train models to generate intermediate rationales at each token, thereby enhancing reasoning capabilities (Zelikman et al., 2024).
Reusing Hidden States: Early Universal Transformers (CsordĂĄs et al., ) reused hidden states through a transition network, aiming to enhance the encoder-decoder transformer architecture toward Turing completeness. A more recent method (Geiping et al., 2025) iterates over multiple layers of a language model to refine intermediate hidden states, treating this process as a form of latent thinking. Giannou et al. (2023) proposed looped transformer, recycling output hidden states back into input embeddings for algorithmic tasks. Similarly, Hao et al. (2024) fine-tuned models to reason directly within continuous latent spaces, using final hidden states as embeddings to achieve reasoning without explicit CoT. In contrast, our method relies on pondering embeddings derived from the predicted probabilities of LMs.
## 6 Conclusion
In this paper, we introduce the pondering process into language models through solely self-supervised learning. Pondering LM can be naturally learned through pretraining on large-scale general corpora. Our extensive experiments across three widely adopted architecturesâGPT-2, Pythia, and LLaMAâhighlight the effectiveness and generality of our proposed method. Notably, our PonderingPythia consistently outperforms the official Pythia model on language modeling tasks, scaling curves, downstream tasks, and instruction-following abilities when pretrained on the large-scale Pile dataset. As increasing the number of pondering steps further improves language model performance, we posit that our approach introduces a promising new dimension along which language model capabilities can be scaled.
## References
- Allen-Zhu & Li (2023) Allen-Zhu, Z. and Li, Y. Physics of language models: Part 3.2, knowledge manipulation. arXiv preprint arXiv:2309.14402, 2023.
- Allen-Zhu & Li (2024) Allen-Zhu, Z. and Li, Y. Physics of language models: Part 3.3, knowledge capacity scaling laws. arXiv preprint arXiv:2404.05405, 2024.
- Amini et al. (2024) Amini, A., Vieira, T., Ash, E., and Cotterell, R. Variational best-of-n alignment. arXiv preprint arXiv:2407.06057, 2024.
- Biderman et al. (2023) Biderman, S., Schoelkopf, H., Anthony, Q. G., Bradley, H., OâBrien, K., Hallahan, E., Khan, M. A., Purohit, S., Prashanth, U. S., Raff, E., et al. Pythia: A suite for analyzing large language models across training and scaling. In International Conference on Machine Learning, pp. 2397â2430. PMLR, 2023.
- Biran et al. (2024) Biran, E., Gottesman, D., Yang, S., Geva, M., and Globerson, A. Hopping too late: Exploring the limitations of large language models on multi-hop queries. arXiv preprint arXiv:2406.12775, 2024.
- Bisk et al. (2020) Bisk, Y., Zellers, R., Gao, J., Choi, Y., et al. Piqa: Reasoning about physical commonsense in natural language. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pp. 7432â7439, 2020.
- Black et al. (2021) Black, S., Gao, L., Wang, P., Leahy, C., and Biderman, S. GPT-Neo: Large Scale Autoregressive Language Modeling with Mesh-Tensorflow, March 2021. URL https://doi.org/10.5281/zenodo.5297715. If you use this software, please cite it using these metadata.
- Brown et al. (2020) Brown, T., Mann, B., Ryder, N., Subbiah, M., Kaplan, J. D., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877â1901, 2020.
- Clark et al. (2018) Clark, P., Cowhey, I., Etzioni, O., Khot, T., Sabharwal, A., Schoenick, C., and Tafjord, O. Think you have solved question answering? try arc, the ai2 reasoning challenge. arXiv preprint arXiv:1803.05457, 2018.
- Cobbe et al. (2021) Cobbe, K., Kosaraju, V., Bavarian, M., Chen, M., Jun, H., Kaiser, L., Plappert, M., Tworek, J., Hilton, J., Nakano, R., et al. Training verifiers to solve math word problems, 2021. URL https://arxiv. org/abs/2110.14168, 9, 2021.
- (11) CsordĂĄs, R., Irie, K., Schmidhuber, J., Potts, C., and Manning, C. D. Moeut: Mixture-of-experts universal transformers. In The Thirty-eighth Annual Conference on Neural Information Processing Systems.
- DeepSeek-AI et al. (2025) DeepSeek-AI, Guo, D., Yang, D., and et al., H. Z. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. 2025. URL https://arxiv.org/abs/2501.12948.
- Fedorenko et al. (2024) Fedorenko, E., Piantadosi, S. T., and Gibson, E. A. Language is primarily a tool for communication rather than thought. Nature, 630(8017):575â586, 2024.
- Gao et al. (2020) Gao, L., Biderman, S., Black, S., Golding, L., Hoppe, T., Foster, C., Phang, J., He, H., Thite, A., Nabeshima, N., et al. The pile: An 800gb dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027, 2020.
- Gao et al. (2023) Gao, L., Tow, J., Abbasi, B., Biderman, S., Black, S., DiPofi, A., Foster, C., Golding, L., Hsu, J., Le Noacâh, A., Li, H., McDonell, K., Muennighoff, N., Ociepa, C., Phang, J., Reynolds, L., Schoelkopf, H., Skowron, A., Sutawika, L., Tang, E., Thite, A., Wang, B., Wang, K., and Zou, A. A framework for few-shot language model evaluation, 12 2023. URL https://zenodo.org/records/10256836.
- Geiping et al. (2025) Geiping, J., McLeish, S., Jain, N., Kirchenbauer, J., Singh, S., Bartoldson, B. R., Kailkhura, B., Bhatele, A., and Goldstein, T. Scaling up test-time compute with latent reasoning: A recurrent depth approach. arXiv preprint arXiv:2502.05171, 2025.
- Giannou et al. (2023) Giannou, A., Rajput, S., Sohn, J.-y., Lee, K., Lee, J. D., and Papailiopoulos, D. Looped transformers as programmable computers. In International Conference on Machine Learning, pp. 11398â11442. PMLR, 2023.
- Gu & Dao (2023) Gu, A. and Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752, 2023.
- Gui et al. (2024) Gui, L., GĂąrbacea, C., and Veitch, V. Bonbon alignment for large language models and the sweetness of best-of-n sampling. arXiv preprint arXiv:2406.00832, 2024.
- Hackenburg et al. (2025) Hackenburg, K., Tappin, B. M., Röttger, P., Hale, S. A., Bright, J., and Margetts, H. Scaling language model size yields diminishing returns for single-message political persuasion. Proceedings of the National Academy of Sciences, 122(10):e2413443122, 2025.
- Hao et al. (2024) Hao, S., Sukhbaatar, S., Su, D., Li, X., Hu, Z., Weston, J., and Tian, Y. Training large language models to reason in a continuous latent space. arXiv preprint arXiv:2412.06769, 2024.
- Hassid et al. (2024) Hassid, M., Remez, T., Gehring, J., Schwartz, R., and Adi, Y. The larger the better? improved llm code-generation via budget reallocation. arXiv preprint arXiv:2404.00725, 2024.
- Hoffmann et al. (2022a) Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E., Casas, D. d. L., Hendricks, L. A., Welbl, J., Clark, A., et al. Training compute-optimal large language models. arXiv preprint arXiv:2203.15556, 2022a.
- Hoffmann et al. (2022b) Hoffmann, J., Borgeaud, S., Mensch, A., Buchatskaya, E., Cai, T., Rutherford, E., de Las Casas, D., Hendricks, L. A., Welbl, J., Clark, A., et al. An empirical analysis of compute-optimal large language model training. Advances in Neural Information Processing Systems, 35:30016â30030, 2022b.
- Hou et al. (2025) Hou, Z., Lv, X., Lu, R., Zhang, J., Li, Y., Yao, Z., Li, J., Tang, J., and Dong, Y. Advancing language model reasoning through reinforcement learning and inference scaling. arXiv preprint arXiv:2501.11651, 2025.
- Huang et al. (2022) Huang, J., Gu, S. S., Hou, L., Wu, Y., Wang, X., Yu, H., and Han, J. Large language models can self-improve. arXiv preprint arXiv:2210.11610, 2022.
- Jaech et al. (2024) Jaech, A., Kalai, A., Lerer, A., Richardson, A., El-Kishky, A., Low, A., Helyar, A., Madry, A., Beutel, A., Carney, A., et al. Openai o1 system card. arXiv preprint arXiv:2412.16720, 2024.
- Kaplan et al. (2020) Kaplan, J., McCandlish, S., Henighan, T., Brown, T. B., Chess, B., Child, R., Gray, S., Radford, A., Wu, J., and Amodei, D. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361, 2020.
- Lai et al. (2017) Lai, G., Xie, Q., Liu, H., Yang, Y., and Hovy, E. Race: Large-scale reading comprehension dataset from examinations. arXiv preprint arXiv:1704.04683, 2017.
- Le Scao et al. (2023) Le Scao, T., Fan, A., Akiki, C., Pavlick, E., IliÄ, S., Hesslow, D., CastagnĂ©, R., Luccioni, A. S., Yvon, F., GallĂ©, M., et al. Bloom: A 176b-parameter open-access multilingual language model. 2023.
- Lee et al. (2025) Lee, K.-H., Fischer, I., Wu, Y.-H., Marwood, D., Baluja, S., Schuurmans, D., and Chen, X. Evolving deeper llm thinking. arXiv preprint arXiv:2501.09891, 2025.
- Li et al. (2025) Li, D., Cao, S., Griggs, T., Liu, S., Mo, X., Patil, S. G., Zaharia, M., Gonzalez, J. E., and Stoica, I. Llms can easily learn to reason from demonstrations structure, not content, is what matters! arXiv preprint arXiv:2502.07374, 2025.
- Li et al. (2023) Li, L. H., Hessel, J., Yu, Y., Ren, X., Chang, K.-W., and Choi, Y. Symbolic chain-of-thought distillation: Small models can also "think" step-by-step. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2665â2679, Toronto, Canada, jul 2023. Association for Computational Linguistics.
- Li et al. (2024) Li, W., Liu, X., Li, Y., Jin, Y., Tian, H., Zhong, Z., Liu, G., Zhang, Y., and Chen, K. Understanding communication characteristics of distributed training. In Proceedings of the 8th Asia-Pacific Workshop on Networking, pp. 1â8, 2024.
- Liu et al. (2024) Liu, A., Feng, B., Xue, B., Wang, B., Wu, B., Lu, C., Zhao, C., Deng, C., Zhang, C., Ruan, C., et al. Deepseek-v3 technical report. arXiv preprint arXiv:2412.19437, 2024.
- Madaan et al. (2024) Madaan, A., Tandon, N., Gupta, P., Hallinan, S., Gao, L., Wiegreffe, S., Alon, U., Dziri, N., Prabhumoye, S., Yang, Y., et al. Self-refine: Iterative refinement with self-feedback. Advances in Neural Information Processing Systems, 36, 2024.
- Min et al. (2024) Min, Y., Chen, Z., Jiang, J., Chen, J., Deng, J., Hu, Y., Tang, Y., Wang, J., Cheng, X., Song, H., et al. Imitate, explore, and self-improve: A reproduction report on slow-thinking reasoning systems. arXiv preprint arXiv:2412.09413, 2024.
- Muennighoff et al. (2023) Muennighoff, N., Rush, A., Barak, B., Le Scao, T., Tazi, N., Piktus, A., Pyysalo, S., Wolf, T., and Raffel, C. A. Scaling data-constrained language models. Advances in Neural Information Processing Systems, 36:50358â50376, 2023.
- Muennighoff et al. (2025) Muennighoff, N., Yang, Z., Shi, W., Li, X. L., Fei-Fei, L., Hajishirzi, H., Zettlemoyer, L., Liang, P., CandĂšs, E., and Hashimoto, T. s1: Simple test-time scaling. arXiv preprint arXiv:2501.19393, 2025.
- Narayanan et al. (2021) Narayanan, D., Shoeybi, M., Casper, J., LeGresley, P., Patwary, M., Korthikanti, V., Vainbrand, D., Kashinkunti, P., Bernauer, J., Catanzaro, B., et al. Efficient large-scale language model training on gpu clusters using megatron-lm. In Proceedings of the international conference for high performance computing, networking, storage and analysis, pp. 1â15, 2021.
- Nye et al. (2021) Nye, M., Andreassen, A. J., Gur-Ari, G., Michalewski, H., Austin, J., Bieber, D., Dohan, D., Lewkowycz, A., Bosma, M., Luan, D., et al. Show your work: Scratchpads for intermediate computation with language models. arXiv preprint arXiv:2112.00114, 2021.
- Pang et al. (2025) Pang, B., Dong, H., Xu, J., Savarese, S., Zhou, Y., and Xiong, C. Bolt: Bootstrap long chain-of-thought in language models without distillation. arXiv preprint arXiv:2502.03860, 2025.
- Paperno et al. (2016) Paperno, D., Kruszewski, G., Lazaridou, A., Pham, Q. N., Bernardi, R., Pezzelle, S., Baroni, M., Boleda, G., and FernĂĄndez, R. The lambada dataset: Word prediction requiring a broad discourse context. arXiv preprint arXiv:1606.06031, 2016.
- Pati et al. (2023) Pati, S., Aga, S., Islam, M., Jayasena, N., and Sinclair, M. D. Computation vs. communication scaling for future transformers on future hardware. arXiv preprint arXiv:2302.02825, 2023.
- Peng et al. (2023) Peng, B., Alcaide, E., Anthony, Q., Albalak, A., Arcadinho, S., Biderman, S., Cao, H., Cheng, X., Chung, M., Grella, M., et al. Rwkv: Reinventing rnns for the transformer era. arXiv preprint arXiv:2305.13048, 2023.
- Pfau et al. (2024) Pfau, J., Merrill, W., and Bowman, S. R. Letâs think dot by dot: Hidden computation in transformer language models. arXiv preprint arXiv:2404.15758, 2024.
- Sakaguchi et al. (2021) Sakaguchi, K., Bras, R. L., Bhagavatula, C., and Choi, Y. Winogrande: An adversarial winograd schema challenge at scale. Communications of the ACM, 64(9):99â106, 2021.
- Sessa et al. (2024) Sessa, P. G., Dadashi, R., Hussenot, L., Ferret, J., Vieillard, N., Ramé, A., Shariari, B., Perrin, S., Friesen, A., Cideron, G., et al. Bond: Aligning llms with best-of-n distillation. arXiv preprint arXiv:2407.14622, 2024.
- Snell et al. (2024) Snell, C., Lee, J., Xu, K., and Kumar, A. Scaling llm test-time compute optimally can be more effective than scaling model parameters. arXiv preprint arXiv:2408.03314, 2024.
- Stroebl et al. (2024) Stroebl, B., Kapoor, S., and Narayanan, A. Inference scaling flaws: The limits of llm resampling with imperfect verifiers. arXiv preprint arXiv:2411.17501, 2024.
- Su et al. (2024) Su, J., Ahmed, M., Lu, Y., Pan, S., Bo, W., and Liu, Y. Roformer: Enhanced transformer with rotary position embedding. Neurocomputing, 568:127063, 2024.
- Sun et al. (2024) Sun, H., Haider, M., Zhang, R., Yang, H., Qiu, J., Yin, M., Wang, M., Bartlett, P., and Zanette, A. Fast best-of-n decoding via speculative rejection. arXiv preprint arXiv:2410.20290, 2024.
- Taori et al. (2023) Taori, R., Gulrajani, I., Zhang, T., Dubois, Y., Li, X., Guestrin, C., Liang, P., and Hashimoto, T. B. Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca, 2023.
- Villalobos et al. (2022) Villalobos, P., Sevilla, J., Heim, L., Besiroglu, T., Hobbhahn, M., and Ho, A. Will we run out of data? an analysis of the limits of scaling datasets in machine learning. arXiv preprint arXiv:2211.04325, 1, 2022.
- Wang et al. (2022) Wang, X., Wei, J., Schuurmans, D., Le, Q., Chi, E., Narang, S., Chowdhery, A., and Zhou, D. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171, 2022.
- Wang et al. (2024a) Wang, X., Caccia, L., Ostapenko, O., Yuan, X., Wang, W. Y., and Sordoni, A. Guiding language model reasoning with planning tokens, 2024a. URL https://arxiv.org/abs/2310.05707.
- Wang et al. (2024b) Wang, Y., Wu, Y., Wei, Z., Jegelka, S., and Wang, Y. A theoretical understanding of self-correction through in-context alignment. arXiv preprint arXiv:2405.18634, 2024b.
- Wei et al. (2022) Wei, J., Wang, X., Schuurmans, D., Bosma, M., Xia, F., Chi, E., Le, Q. V., Zhou, D., et al. Chain-of-thought prompting elicits reasoning in large language models. Advances in neural information processing systems, 35:24824â24837, 2022.
- Welbl et al. (2017) Welbl, J., Liu, N. F., and Gardner, M. Crowdsourcing multiple choice science questions. arXiv preprint arXiv:1707.06209, 2017.
- Yang et al. (2024) Yang, S., Gribovskaya, E., Kassner, N., Geva, M., and Riedel, S. Do large language models latently perform multi-hop reasoning? arXiv preprint arXiv:2402.16837, 2024.
- Yue et al. (2025) Yue, Y., Chen, Z., Lu, R., Zhao, A., Wang, Z., Song, S., and Huang, G. Does reinforcement learning really incentivize reasoning capacity in llms beyond the base model? arXiv preprint arXiv:2504.13837, 2025.
- Zelikman et al. (2024) Zelikman, E., Harik, G., Shao, Y., Jayasiri, V., Haber, N., and Goodman, N. D. Quiet-star: Language models can teach themselves to think before speaking, 2024. URL https://arxiv.org/abs/2403.09629.
- Zellers et al. (2019) Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., and Choi, Y. Hellaswag: Can a machine really finish your sentence? arXiv preprint arXiv:1905.07830, 2019.
- Zeng et al. (2024) Zeng, Z., Cheng, Q., Yin, Z., Wang, B., Li, S., Zhou, Y., Guo, Q., Huang, X., and Qiu, X. Scaling of search and learning: A roadmap to reproduce o1 from reinforcement learning perspective. arXiv preprint arXiv:2412.14135, 2024.
- Zhang & Sennrich (2019) Zhang, B. and Sennrich, R. Root mean square layer normalization. Advances in Neural Information Processing Systems, 32, 2019.
- Zhang et al. (2024) Zhang, P., Zeng, G., Wang, T., and Lu, W. Tinyllama: An open-source small language model. arXiv preprint arXiv:2401.02385, 2024.
- Zhang et al. (2022) Zhang, S., Roller, S., Goyal, N., Artetxe, M., Chen, M., Chen, S., Dewan, C., Diab, M., Li, X., Lin, X. V., et al. Opt: Open pre-trained transformer language models. arXiv preprint arXiv:2205.01068, 2022.
- Zheng et al. (2023) Zheng, L., Chiang, W.-L., Sheng, Y., Zhuang, S., Wu, Z., Zhuang, Y., Lin, Z., Li, Z., Li, D., Xing, E., et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36:46595â46623, 2023.
- Zhou et al. (2024) Zhou, J., Pang, L., Shen, H., and Cheng, X. Think before you speak: Cultivating communication skills of large language models via inner monologue, 2024. URL https://arxiv.org/abs/2311.07445.
- Zhu et al. (2025) Zhu, D., Wei, X., Zhao, G., Wu, W., Zou, H., Ran, J., Wang, X., Sun, L., Zhang, X., and Li, S. Chain-of-thought matters: Improving long-context language models with reasoning path supervision. arXiv preprint arXiv:2502.20790, 2025.