<details>
<summary>Image 1 Details</summary>

### Visual Description
Icon/Small Image (202x49)
</details>
## Training Compute-Optimal Large Language Models
Jordan Hoffmann β
, Sebastian Borgeaud β
, Arthur Mensch β
, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, Tom Hennigan, Eric Noland, Katie Millican, George van den Driessche, Bogdan Damoc, Aurelia Guy, Simon Osindero, Karen Simonyan, Erich Elsen, Jack W. Rae, Oriol Vinyals and Laurent Sifre β
β
Equal contributions
We investigate the optimal model size and number of tokens for training a transformer language model under a given compute budget. We find that current large language models are significantly undertrained, a consequence of the recent focus on scaling language models whilst keeping the amount of training data constant. By training over 400 language models ranging from 70 million to over 16 billion parameters on 5 to 500 billion tokens, we find that for compute-optimal training, the model size and the number of training tokens should be scaled equally: for every doubling of model size the number of training tokens should also be doubled. We test this hypothesis by training a predicted computeoptimal model, Chinchilla , that uses the same compute budget as Gopher but with 70B parameters and 4 more more data. Chinchilla uniformly and significantly outperforms Gopher (280B), GPT-3 (175B), Jurassic-1 (178B), and Megatron-Turing NLG (530B) on a large range of downstream evaluation tasks. This also means that Chinchilla uses substantially less compute for fine-tuning and inference, greatly facilitating downstream usage. As a highlight, Chinchilla reaches a state-of-the-art average accuracy of 67.5% on the MMLU benchmark, greater than a 7% improvement over Gopher .
## 1. Introduction
Recently a series of Large Language Models (LLMs) have been introduced (Brown et al., 2020; Lieber et al., 2021; Rae et al., 2021; Smith et al., 2022; Thoppilan et al., 2022), with the largest dense language models now having over 500 billion parameters. These large autoregressive transformers (Vaswani et al., 2017) have demonstrated impressive performance on many tasks using a variety of evaluation protocols such as zero-shot, few-shot, and fine-tuning.
The compute and energy cost for training large language models is substantial (Rae et al., 2021; Thoppilan et al., 2022) and rises with increasing model size. In practice, the allocated training compute budget is often known in advance: how many accelerators are available and for how long we want to use them. Since it is typically only feasible to train these large models once, accurately estimating the best model hyperparameters for a given compute budget is critical (Tay et al., 2021).
Kaplan et al. (2020) showed that there is a power law relationship between the number of parameters in an autoregressive language model (LM) and its performance. As a result, the field has been training larger and larger models, expecting performance improvements. One notable conclusion in Kaplan et al. (2020) is that large models should not be trained to their lowest possible loss to be compute optimal. Whilst we reach the same conclusion, we estimate that large models should be trained for many more training tokens than recommended by the authors. Specifically, given a 10 increase computational budget, they suggests that the size of the model should increase 5 5 while the number of training tokens should only increase 1.8 . Instead, we find that model size and the number of training tokens should be scaled in equal proportions.
Following Kaplan et al. (2020) and the training setup of GPT-3 (Brown et al., 2020), many of the recently trained large models have been trained for approximately 300 billion tokens (Table 1), in line with the approach of predominantly increasing model size when increasing compute.
Figure 1 j Overlaid predictions. We overlay the predictions from our three different approaches, along with projections from Kaplan et al. (2020). We find that all three methods predict that current large models should be substantially smaller and therefore trained much longer than is currently done. In Figure A3, we show the results with the predicted optimal tokens plotted against the optimal number of parameters for fixed FLOP budgets. Chinchilla outperforms Gopher and the other large models (see Section 4.2).
<details>
<summary>Image 2 Details</summary>

### Visual Description
## Chart: Parameter Count vs. Computational Cost
### Overview
This chart depicts the relationship between the number of parameters in a language model and the computational cost (measured in FLOPS) required to train it. It compares three different approaches to scaling language models with a theoretical scaling law proposed by Kaplan et al. (2020). Several specific models are plotted as data points to illustrate their parameter count and FLOPS.
### Components/Axes
* **X-axis:** FLOPS (Floating Point Operations Per Second), logarithmic scale from 10<sup>17</sup> to 10<sup>25</sup>.
* **Y-axis:** Parameters, logarithmic scale from 10<sup>7</sup> to 1 T (10<sup>12</sup>).
* **Lines:**
* Approach 1 (Blue)
* Approach 2 (Orange)
* Approach 3 (Green)
* **Theoretical Line:** Kaplan et al. (2020) (Black dashed line)
* **Data Points (Models):**
* Chinchilla (70B parameters) (Teal star)
* Gopher (280B parameters) (Yellow star)
* GPT-3 (175B parameters) (Red star)
* Megatron-Turing NLG (530B parameters) (Purple star)
* **Legend:** Located in the top-right corner, associating colors with the different approaches and the Kaplan et al. line.
### Detailed Analysis
The chart shows a strong positive correlation between parameters and FLOPS. All three approaches demonstrate that increasing the number of parameters requires a significant increase in computational cost.
* **Kaplan et al. (2020):** The black dashed line represents a theoretical scaling law. It slopes upward from approximately 10<sup>17</sup> FLOPS and 10<sup>7</sup> parameters to 10<sup>25</sup> FLOPS and 10<sup>12</sup> parameters.
* **Approach 1 (Blue):** This line starts at approximately 10<sup>18</sup> FLOPS and 10<sup>8</sup> parameters and slopes upward, remaining below the Kaplan et al. line.
* **Approach 2 (Orange):** This line starts at approximately 10<sup>19</sup> FLOPS and 10<sup>8</sup> parameters and slopes upward, intersecting the Kaplan et al. line around 10<sup>23</sup> FLOPS.
* **Approach 3 (Green):** This line starts at approximately 10<sup>19</sup> FLOPS and 10<sup>8</sup> parameters and slopes upward, remaining above the Kaplan et al. line.
**Data Point Values (Approximate):**
* **Chinchilla (70B):** Approximately 10<sup>21</sup> FLOPS and 7 x 10<sup>10</sup> parameters.
* **Gopher (280B):** Approximately 10<sup>23</sup> FLOPS and 2.8 x 10<sup>11</sup> parameters.
* **GPT-3 (175B):** Approximately 10<sup>23</sup> FLOPS and 1.75 x 10<sup>11</sup> parameters.
* **Megatron-Turing NLG (530B):** Approximately 10<sup>24</sup> FLOPS and 5.3 x 10<sup>11</sup> parameters.
### Key Observations
* The models generally fall along the trend lines, but there is some deviation.
* Megatron-Turing NLG (530B) requires the highest FLOPS and has the largest number of parameters.
* Chinchilla (70B) has the lowest FLOPS and parameter count among the plotted models.
* Approach 3 consistently requires the most FLOPS for a given number of parameters.
* Approach 1 consistently requires the least FLOPS for a given number of parameters.
### Interpretation
The chart demonstrates the computational cost associated with scaling language models. The different approaches suggest trade-offs between parameter count and computational efficiency. The Kaplan et al. line provides a theoretical benchmark, and the plotted models show how real-world models compare to this benchmark. The deviations from the theoretical line could be due to various factors, such as differences in model architecture, training data, and optimization techniques. The chart highlights the significant computational resources required to train large language models, and the need for efficient scaling strategies. The positioning of the models relative to the theoretical line and the different approaches suggests that some models may be over- or under-trained for their parameter count, or that different training strategies are more or less efficient. The data suggests that simply increasing parameters does not guarantee improved performance and that optimizing the training process is crucial.
</details>
In this work, we revisit the question: Given a fixed FLOPs budget, 1 how should one trade-off model size and the number of training tokens? To answer this question, we model the final pre-training loss 2 πΏ ' π π· ' as a function of the number of model parameters π , and the number of training tokens, π· . Since the computational budget πΆ is a deterministic function FLOPs ' π π· ' of the number of seen training tokens and model parameters, we are interested in minimizing πΏ under the constraint FLOPs ' π π· ' = πΆ :
<!-- formula-not-decoded -->
The functions ππππ‘ ' πΆ ' , and π·πππ‘ ' πΆ ' describe the optimal allocation of a computational budget πΆ . We empirically estimate these functions based on the losses of over 400 models, ranging from under 70M to over 16B parameters, and trained on 5B to over 400B tokens - with each model configuration trained for several different training horizons. Our approach leads to considerably different results than that of Kaplan et al. (2020). We highlight our results in Figure 1 and how our approaches differ in Section 2.
Based on our estimated compute-optimal frontier, we predict that for the compute budget used to train Gopher , an optimal model should be 4 times smaller, while being training on 4 times more tokens. We verify this by training a more compute-optimal 70B model, called Chinchilla , on 1.4 trillion tokens. Not only does Chinchilla outperform its much larger counterpart, Gopher , but its reduced model size reduces inference cost considerably and greatly facilitates downstream uses on smaller hardware. The energy cost of a large language model is amortized through its usage for inference an fine-tuning. The benefits of a more optimally trained smaller model, therefore, extend beyond the immediate benefits of its improved performance.
1 For example, knowing the number of accelerators and a target training duration.
2 For simplicity , we perform our analysis on the smoothed training loss which is an unbiased estimate of the test loss, as we are in the infinite data regime (the number of training tokens is less than the number of tokens in the entire corpus).
Table 1 j Current LLMs . We show five of the current largest dense transformer models, their size, and the number of training tokens. Other than LaMDA (Thoppilan et al., 2022), most models are trained for approximately 300 billion tokens. We introduce Chinchilla , a substantially smaller model, trained for much longer than 300B tokens.
| Model | Size (# Parameters) | Training Tokens |
|----------------------------------|-----------------------|-------------------|
| LaMDA (Thoppilan et al., 2022) | 137 Billion | 168 Billion |
| GPT-3 (Brown et al., 2020) | 175 Billion | 300 Billion |
| Jurassic (Lieber et al., 2021) | 178 Billion | 300 Billion |
| Gopher (Rae et al., 2021) | 280 Billion | 300 Billion |
| MT-NLG 530B (Smith et al., 2022) | 530 Billion | 270 Billion |
| Chinchilla | 70 Billion | 1.4 Trillion |
## 2. Related Work
Large language models. A variety of large language models have been introduced in the last few years. These include both dense transformer models (Brown et al., 2020; Lieber et al., 2021; Rae et al., 2021; Smith et al., 2022; Thoppilan et al., 2022) and mixture-of-expert (MoE) models (Du et al., 2021; Fedus et al., 2021; Zoph et al., 2022). The largest dense transformers have passed 500 billion parameters (Smith et al., 2022). The drive to train larger and larger models is clear-so far increasing the size of language models has been responsible for improving the state-of-the-art in many language modelling tasks. Nonetheless, large language models face several challenges, including their overwhelming computational requirements (the cost of training and inference increase with model size) (Rae et al., 2021; Thoppilan et al., 2022) and the need for acquiring more high-quality training data. In fact, in this work we find that larger, high quality datasets will play a key role in any further scaling of language models.
Modelling the scaling behavior. Understanding the scaling behaviour of language models and their transfer properties has been important in the development of recent large models (Hernandez et al., 2021; Kaplan et al., 2020). Kaplan et al. (2020) first showed a predictable relationship between model size and loss over many orders of magnitude. The authors investigate the question of choosing the optimal model size to train for a given compute budget. Similar to us, they address this question by training various models. Our work differs from Kaplan et al. (2020) in several important ways. First, the authors use a fixed number of training tokens and learning rate schedule for all models; this prevents them from modelling the impact of these hyperparameters on the loss. In contrast, we find that setting the learning rate schedule to approximately match the number of training tokens results in the best final loss regardless of model size-see Figure A1. For a fixed learning rate cosine schedule to 130B tokens, the intermediate loss estimates (for π· 0 130B) are therefore overestimates of the loss of a model trained with a schedule length matching π· 0 . Using these intermediate losses results in underestimating the effectiveness of training models on less data than 130B tokens, and eventually contributes to the conclusion that model size should increase faster than training data size as compute budget increases. In contrast, our analysis predicts that both quantities should scale at roughly the same rate. Secondly, we include models with up to 16B parameters, as we observe that there is slight curvature in the FLOP-loss frontier (see Appendix E)-in fact, the majority of the models used in our analysis have more than 500 million parameters, in contrast the majority of runs in Kaplan et al. (2020) are significantly smaller-many being less than 100M parameters.
Recently, Clark et al. (2022) specifically looked in to the scaling properties of Mixture of Expert
language models, showing that the scaling with number of experts diminishes as the model size increases-their approach models the loss as a function of two variables: the model size and the number of experts. However, the analysis is done with a fixed number of training tokens, as in Kaplan et al. (2020), potentially underestimating the improvements of branching.
Estimating hyperparameters for large models. The model size and the number of training tokens are not the only two parameters to chose when selecting a language model and a procedure to train it. Other important factors include learning rate, learning rate schedule, batch size, optimiser, and width-to-depth ratio. In this work, we focus on model size and the number of training steps, and we rely on existing work and provided experimental heuristics to determine the other necessary hyperparameters. Yang et al. (2021) investigates how to choose a variety of these parameters for training an autoregressive transformer, including the learning rate and batch size. McCandlish et al. (2018) finds only a weak dependence between optimal batch size and model size. Shallue et al. (2018); Zhang et al. (2019) suggest that using larger batch-sizes than those we use is possible. Levine et al. (2020) investigates the optimal depth-to-width ratio for a variety of standard model sizes. We use slightly less deep models than proposed as this translates to better wall-clock performance on our hardware.
Improved model architectures. Recently, various promising alternatives to traditional dense transformers have been proposed. For example, through the use of conditional computation large MoE models like the 1.7 trillion parameter Switch transformer (Fedus et al., 2021), the 1.2 Trillion parameter GLaM model (Du et al., 2021), and others (Artetxe et al., 2021; Zoph et al., 2022) are able to provide a large effective model size despite using relatively fewer training and inference FLOPs. However, for very large models the computational benefits of routed models seems to diminish (Clark et al., 2022). An orthogonal approach to improving language models is to augment transformers with explicit retrieval mechanisms, as done by Borgeaud et al. (2021); Guu et al. (2020); Lewis et al. (2020). This approach effectively increases the number of data tokens seen during training (by a factor of 10 in Borgeaud et al. (2021)). This suggests that the performance of language models may be more dependant on the size of the training data than previously thought.
## 3. Estimating the optimal parameter/training tokens allocation
We present three different approaches to answer the question driving our research: Given a fixed FLOPs budget, how should one trade-off model size and the number of training tokens? In all three cases we start by training a range of models varying both model size and the number of training tokens and use the resulting training curves to fit an empirical estimator of how they should scale. We assume a power-law relationship between compute and model size as done in Clark et al. (2022); Kaplan et al. (2020), though future work may want to include potential curvature in this relationship for large model sizes. The resulting predictions are similar for all three methods and suggest that parameter count and number of training tokens should be increased equally with more compute 3 with proportions reported in Table 2. This is in clear contrast to previous work on this topic and warrants further investigation.
3 We compute FLOPs as described in Appendix F.
Figure 2 j Training curve envelope. On the left we show all of our different runs. We launched a range of model sizes going from 70M to 10B, each for four different cosine cycle lengths. From these curves, we extracted the envelope of minimal loss per FLOP, and we used these points to estimate the optimal model size ( center ) for a given compute budget and the optimal number of training tokens ( right ). In green, we show projections of optimal model size and training token count based on the number of FLOPs used to train Gopher (5 76 10 23 ).
<details>
<summary>Image 3 Details</summary>

### Visual Description
\n
## Chart: Training Performance and Scaling Laws
### Overview
The image presents three charts arranged horizontally. The first chart is a heatmap-style line plot showing Training Loss vs. FLOPS, colored by model size (Parameters). The second chart shows Parameters vs. FLOPS, and the third chart shows Tokens vs. FLOPS. All charts use logarithmic scales on both axes. The charts appear to investigate the relationship between computational cost (FLOPS), model size (Parameters), training data size (Tokens), and training performance (Loss).
### Components/Axes
* **Chart 1 (Left):**
* X-axis: FLOPS (Floating Point Operations Per Second), scale is logarithmic from 10<sup>17</sup> to 10<sup>23</sup>.
* Y-axis: Training Loss, scale is linear from 2.0 to 6.0.
* Color Scale (Legend): Represents model size in Parameters, ranging from 75M (dark purple) to 10B (yellow). Intermediate values are 250M (green), 500M (orange), 1B (light orange), 2.5B (red), 5B (yellow-orange), and 10B (yellow). The legend is positioned on the top-right.
* **Chart 2 (Center):**
* X-axis: FLOPS, scale is logarithmic from 10<sup>17</sup> to 10<sup>25</sup>.
* Y-axis: Parameters, scale is logarithmic from 10<sup>7</sup> (10M) to 10<sup>10</sup> (10B).
* Trend Line: A red line indicating a linear relationship.
* Annotation: "678" is placed near a data point.
* Highlighted Region: A light blue rectangle in the top-right corner.
* **Chart 3 (Right):**
* X-axis: FLOPS, scale is logarithmic from 10<sup>17</sup> to 10<sup>25</sup>.
* Y-axis: Tokens, scale is logarithmic from 10<sup>8</sup> to 10<sup>12</sup>.
* Trend Line: A red line indicating a linear relationship.
* Annotation: "1.5T" is placed near a data point.
* Highlighted Region: A light blue rectangle in the top-right corner.
### Detailed Analysis or Content Details
* **Chart 1:** The lines generally slope downwards, indicating that as FLOPS increase (more computation), training loss decreases. The color gradient shows that larger models (yellow) tend to have lower loss values at higher FLOPS, but also exhibit more variability. The lines are densely packed at lower FLOPS values, and spread out as FLOPS increase.
* Approximate data points (reading from the chart, with uncertainty):
* At FLOPS = 10<sup>17</sup>, Loss ranges from approximately 5.5 to 6.0 across all model sizes.
* At FLOPS = 10<sup>22</sup>, Loss ranges from approximately 2.2 to 3.5, with the lowest losses associated with the largest models (yellow).
* **Chart 2:** The data points show a strong positive correlation between Parameters and FLOPS. The red trend line appears to fit the data well.
* Approximate data points:
* At FLOPS = 10<sup>17</sup>, Parameters are approximately 10M (10<sup>7</sup>).
* At FLOPS = 10<sup>25</sup>, Parameters are approximately 10B (10<sup>10</sup>).
* The annotated point "678" appears to be near FLOPS = 10<sup>21</sup> and Parameters = 100M (10<sup>8</sup>).
* **Chart 3:** The data points also show a strong positive correlation between Tokens and FLOPS. The red trend line appears to fit the data well.
* Approximate data points:
* At FLOPS = 10<sup>17</sup>, Tokens are approximately 10<sup>8</sup>.
* At FLOPS = 10<sup>25</sup>, Tokens are approximately 10<sup>12</sup>.
* The annotated point "1.5T" appears to be near FLOPS = 10<sup>23</sup> and Tokens = 1.5 x 10<sup>12</sup>.
### Key Observations
* **Scaling Laws:** The charts suggest adherence to scaling laws, where increasing model size (Parameters) and training data (Tokens) generally leads to improved performance (lower Loss), but at a computational cost (FLOPS).
* **Linear Relationships:** The relationships between Parameters/Tokens and FLOPS appear approximately linear on the logarithmic scales.
* **Model Size Impact:** Larger models consistently achieve lower loss values at higher FLOPS.
* **Data Density:** The data is more concentrated at lower FLOPS values, suggesting that fewer experiments were conducted at very high computational costs.
### Interpretation
The data presented strongly supports the idea of scaling laws in deep learning. Increasing the number of parameters in a model and the amount of training data (measured in tokens) generally leads to improved performance, as indicated by the decreasing training loss. However, this improvement comes at a significant computational cost, measured in FLOPS. The approximately linear relationships observed on the logarithmic scales suggest that the computational cost scales predictably with model size and data size. The annotations ("678", "1.5T") likely represent specific experimental configurations or milestones. The highlighted regions in the top-right corners of the second and third charts may indicate the limits of the data or areas where the scaling laws begin to deviate. The heatmap in the first chart provides a visual representation of how model size influences the training loss curve, demonstrating that larger models require more computation to achieve optimal performance. The spread of the lines in the first chart also suggests that there is some variance in training outcomes even for models of the same size, potentially due to factors such as initialization or optimization algorithms.
</details>
## 3.1. Approach 1: Fix model sizes and vary number of training tokens
In our first approach we vary the number of training steps for a fixed family of models (ranging from 70M to over 10B parameters), training each model for 4 different number of training sequences. From these runs, we are able to directly extract an estimate of the minimum loss achieved for a given number of training FLOPs. Training details for this approach can be found in Appendix D.
For each parameter count π we train 4 different models, decaying the learning rate by a factor of 10 over a horizon (measured in number of training tokens) that ranges by a factor of 16 . Then, for each run, we smooth and then interpolate the training loss curve. From this, we obtain a continuous mapping from FLOP count to training loss for each run. Then, for each FLOP count, we determine which run achieves the lowest loss. Using these interpolants, we obtain a mapping from any FLOP count πΆ , to the most efficient choice of model size π and number of training tokens π· such that FLOPs ' π π· ' = πΆ . 4 At 1500 logarithmically spaced FLOP values, we find which model size achieves the lowest loss of all models along with the required number of training tokens. Finally, we fit power laws to estimate the optimal model size and number of training tokens for any given amount of compute (see the center and right panels of Figure 2), obtaining a relationship ππππ‘ / πΆ π and π·πππ‘ / πΆ π . We find that π = 0 50 and π = 0 50-as summarized in Table 2. In Section D.4, we show a head-to-head comparison at 10 21 FLOPs, using the model size recommended by our analysis and by the analysis of Kaplan et al. (2020)-using the model size we predict has a clear advantage.
## 3.2. Approach 2: IsoFLOP profiles
In our second approach we vary the model size 5 for a fixed set of 9 different training FLOP counts 6 (ranging from 6 10 18 to 3 10 21 FLOPs), and consider the final training loss for each point 7 . in contrast with Approach 1 that considered points ' π π· πΏ ' along the entire training runs. This allows us to directly answer the question: For a given FLOP budget, what is the optimal parameter count?
4 Note that all selected points are within the last 15% of training. This suggests that when training a model over π· tokens, we should pick a cosine cycle length that decays 10 over approximately π· tokens-see further details in Appendix B. 5 In approach 2, model size varies up to 16B as opposed to approach 1 where we only used models up to 10B. 6 The number of training tokens is determined by the model size and training FLOPs.
7 We set the cosine schedule length to match the number of tokens, which is optimal according to the analysis presented in Appendix B.
Figure 3 j IsoFLOP curves. For various model sizes, we choose the number of training tokens such that the final FLOPs is a constant. The cosine cycle length is set to match the target FLOP count. We find a clear valley in loss, meaning that for a given FLOP budget there is an optimal model to train ( left ). Using the location of these valleys, we project optimal model size and number of tokens for larger models ( center and right ). In green, we show the estimated number of parameters and tokens for an optimal model trained with the compute budget of Gopher .
<details>
<summary>Image 4 Details</summary>

### Visual Description
## Scatter Plot: Training Loss vs. Parameters & Tokens vs. FLOPs
### Overview
The image presents two scatter plots. The left plot shows Training Loss as a function of Parameters, with different curves representing different training step counts. The right plot shows Tokens processed as a function of FLOPs (Floating Point Operations). A horizontal line is present in the left plot, and a diagonal line is present in the right plot.
### Components/Axes
**Left Plot:**
* **X-axis:** Parameters (log scale, from 10B to 6B). Markers are at approximately 100M, 300M, 1B, 3B, 6B.
* **Y-axis:** Training Loss (linear scale, from 2.0 to 3.2).
* **Legend:** Represents training step counts (6e18, 1e19, 3e19, 6e19, 1e20, 3e20, 6e20). Each step count is associated with a different color.
* **Horizontal Line:** A turquoise horizontal line is present at approximately 638.
**Right Plot:**
* **X-axis:** FLOPs (log scale, from 10^7 to 10^15). Markers are at approximately 10^7, 10^9, 10^11, 10^13, 10^15.
* **Y-axis:** Tokens (log scale, from 10B to 1T). Markers are at approximately 10B, 100B, 1B, 10B, 100B, 1T.
* **Diagonal Line:** A red dashed diagonal line is present.
* **Horizontal Line:** A turquoise horizontal line is present at approximately 1.4T.
### Detailed Analysis or Content Details
**Left Plot:**
* **6e18 (Green):** Starts at approximately 3.15, decreases to around 2.6, then plateaus.
* **1e19 (Light Blue):** Starts at approximately 3.0, decreases to around 2.5, then plateaus.
* **3e19 (Blue):** Starts at approximately 2.9, decreases to around 2.4, then plateaus.
* **6e19 (Dark Blue):** Starts at approximately 2.8, decreases to around 2.3, then plateaus.
* **1e20 (Purple):** Starts at approximately 2.7, decreases to around 2.2, then plateaus.
* **3e20 (Dark Purple):** Starts at approximately 2.6, decreases to around 2.1, then plateaus.
* **6e20 (Black):** Starts at approximately 2.5, decreases to around 2.0, then plateaus.
* All curves exhibit a decreasing trend in Training Loss as Parameters increase, then level off. The curves for higher step counts generally reach lower Training Loss values.
**Right Plot:**
* The data points form an approximately linear relationship.
* The line slopes upward, indicating that as FLOPs increase, the number of Tokens processed also increases.
* The data points are relatively closely clustered around the red dashed line.
* The first data point is at approximately (10^7 FLOPs, 10^10 Tokens).
* The last data point is at approximately (10^15 FLOPs, 10^12 Tokens).
### Key Observations
* The left plot demonstrates that increasing the number of parameters generally leads to lower training loss, up to a point of diminishing returns.
* The right plot shows a strong correlation between FLOPs and the number of tokens processed. The relationship appears to be roughly linear.
* The horizontal lines in both plots may represent a threshold or limit.
### Interpretation
The data suggests a scaling relationship between model size (Parameters), training effort (FLOPs), and performance (Training Loss). Increasing model size and training compute generally improves performance, but there are diminishing returns. The linear relationship between FLOPs and Tokens suggests that the computational cost of training scales predictably with the amount of data processed. The horizontal lines could indicate a saturation point where further increases in parameters or FLOPs do not yield significant improvements in training loss or token processing. The data points are relatively well-behaved, suggesting a consistent and predictable relationship between these variables. The curves in the left plot show that the training loss decreases with increasing parameters, but the rate of decrease slows down as the number of parameters increases. This suggests that there is a point of diminishing returns where adding more parameters does not significantly improve the training loss. The right plot shows that the number of tokens processed increases linearly with the number of FLOPs. This suggests that the computational cost of training is proportional to the amount of data processed.
</details>
For each FLOP budget, we plot the final loss (after smoothing) against the parameter count in Figure 3 (left). In all cases, we ensure that we have trained a diverse enough set of model sizes to see a clear minimum in the loss. We fit a parabola to each IsoFLOPs curve to directly estimate at what model size the minimum loss is achieved (Figure 3 (left)). As with the previous approach, we then fit a power law between FLOPs and loss-optimal model size and number of training tokens, shown in Figure 3 (center, right). Again, we fit exponents of the form ππππ‘ / πΆ π and π·πππ‘ / πΆ π and we find that π = 0 49 and π = 0 51-as summarized in Table 2.
## 3.3. Approach 3: Fitting a parametric loss function
Lastly, we model all final losses from experiments in Approach 1 & 2 as a parametric function of model parameter count and the number of seen tokens. Following a classical risk decomposition (see Section D.2), we propose the following functional form
<!-- formula-not-decoded -->
The first term captures the loss for an ideal generative process on the data distribution, and should correspond to the entropy of natural text. The second term captures the fact that a perfectly trained transformer with π parameters underperforms the ideal generative process. The final term captures the fact that the transformer is not trained to convergence, as we only make a finite number of optimisation steps, on a sample of the dataset distribution.
Model fitting. To estimate ' π΄ π΅ πΈ πΌ π½ ' , we minimize the Huber loss (Huber, 1964) between the predicted and observed log loss using the L-BFGS algorithm (Nocedal, 1980):
<!-- formula-not-decoded -->
We account for possible local minima by selecting the best fit from a grid of initialisations. The Huber loss ( πΏ = 10 3 ) is robust to outliers, which we find important for good predictive performance over held-out data points. Section D.2 details the fitting procedure and the loss decomposition.
Figure 4 j Parametric fit. We fit a parametric modelling of the loss Λ πΏ ' π π· ' and display contour ( left ) and isoFLOP slices ( right ). For each isoFLOP slice, we include a corresponding dashed line in the left plot. In the left plot, we show the efficient frontier in blue, which is a line in log-log space. Specifically , the curve goes through each iso-loss contour at the point with the fewest FLOPs. We project the optimal model size given the Gopher FLOP budget to be 40B parameters.
<details>
<summary>Image 5 Details</summary>

### Visual Description
## Chart: IsoLoss Contours and IsoFLOPs Slices
### Overview
The image presents a two-part chart exploring the relationship between model size, training FLOPs (Floating Point Operations), and loss. The left side displays IsoLoss contours, showing lines of equal loss for different combinations of model size and training FLOPs. The right side shows IsoFLOPs slices, illustrating how loss varies with model size for specific training FLOPs budgets. Both charts are overlaid with empirical data points.
### Components/Axes
**Left Chart (IsoLoss Contours):**
* **X-axis:** Training FLOPs (logarithmic scale, from 10<sup>18</sup> to 10<sup>23</sup>). Labeled "Training FLOPs".
* **Y-axis:** Model Size (logarithmic scale, from 100M to 100B). Labeled "Model size".
* **Contours:** Lines representing constant loss values.
* **Data Points:** Scatter plot of "Empirical data" in dark purple.
* **Line:** "Efficient frontier" in blue.
* **Line:** "IsoFLOPs slice" in black.
* **Vertical Line:** "Gopher budget" in teal.
**Right Chart (IsoFLOPs Slices):**
* **X-axis:** Model Size (logarithmic scale, from 100M to 40B). Labeled "Model size".
* **Y-axis:** Loss (linear scale, from 2.00 to 5.00). Labeled "Loss".
* **Lines:** Curves representing different training FLOPs budgets.
* **Data Points:** Scatter plot of "IsoFLOPs slice" in light green.
* **Line:** "Gopher" in black.
**Legend (Right Chart):**
* 6e+18 (cyan)
* 1e+19 (light blue)
* 3e+19 (teal)
* 6e+19 (dark teal)
* 1e+20 (purple)
* 3e+20 (dark purple)
* 6e+20 (navy)
* 1e+21 (dark navy)
* 3e+21 (black)
**Colorbar (Right Chart):**
* Represents Loss, ranging from dark gray (low loss ~2.0) to red (high loss ~5.0).
### Detailed Analysis or Content Details
**Left Chart (IsoLoss Contours):**
* The "Efficient frontier" line slopes upward, indicating that achieving lower loss generally requires either a larger model size or more training FLOPs.
* The "Empirical data" points are clustered around the lower-left region of the chart, with a dense concentration between 10<sup>19</sup> and 10<sup>21</sup> FLOPs and 1B to 10B model size.
* The "IsoFLOPs slice" line appears to be a vertical line at approximately 10<sup>22</sup> FLOPs.
* The "Gopher budget" vertical line is positioned around 10<sup>21</sup> FLOPs.
* IsoLoss contours are densely packed in the lower-left, indicating a steep gradient in loss. Contours become more spaced out as FLOPs and model size increase.
**Right Chart (IsoFLOPs Slices):**
* For each FLOPs budget, the loss generally decreases as model size increases, then plateaus.
* The 6e+18 line shows a rapid decrease in loss for model sizes up to approximately 1B, then levels off.
* The 1e+19 line shows a similar trend, but with a lower minimum loss.
* The 3e+19 line shows a further decrease in minimum loss.
* The 6e+19 line shows a further decrease in minimum loss.
* The 1e+20 line shows a further decrease in minimum loss.
* The 3e+20 line shows a further decrease in minimum loss.
* The 6e+20 line shows a further decrease in minimum loss.
* The 1e+21 line shows a further decrease in minimum loss.
* The 3e+21 line shows a further decrease in minimum loss.
* The "Gopher" line is relatively flat, indicating that increasing model size beyond a certain point does not significantly reduce loss for that FLOPs budget.
* The IsoFLOPs slice data points are scattered along the curves, with a concentration around the lower loss values.
### Key Observations
* There's a clear trade-off between model size, training FLOPs, and loss.
* The "Efficient frontier" represents the optimal combination of model size and FLOPs for a given loss target.
* The empirical data suggests that current models are often operating below the efficient frontier.
* The "Gopher" model appears to be operating near its optimal model size for its FLOPs budget.
* Higher FLOPs budgets generally allow for lower loss values, but with diminishing returns as model size increases.
### Interpretation
The charts demonstrate the scaling laws governing the performance of large language models. They illustrate that increasing model size and training FLOPs generally leads to lower loss, but there are diminishing returns. The "Efficient frontier" represents the theoretical limit of performance for a given computational budget. The empirical data suggests that there is room for improvement in current model training strategies. The IsoFLOPs slices provide a more detailed view of the loss landscape for different FLOPs budgets, revealing the optimal model size for each budget. The "Gopher" model serves as a benchmark, showing the performance achievable with a specific FLOPs budget. The colorbar on the right chart provides a visual representation of the loss values, allowing for a quick assessment of model performance. The logarithmic scales on both axes are crucial for visualizing the wide range of values involved in large language model training. The positioning of the legend in the top-right corner of the right chart allows for easy comparison of the different FLOPs budgets. The overall message is that careful consideration of model size, training FLOPs, and the efficient frontier is essential for maximizing the performance of large language models.
</details>
Efficient frontier. We can approximate the functions ππππ‘ and π·πππ‘ by minimizing the parametric loss Λ πΏ under the constraint FLOPs ' π π· ' 6 ππ· (Kaplan et al., 2020). The resulting ππππ‘ and π·πππ‘ balance the two terms in Equation (3) that depend on model size and data. By construction, they have a power-law form:
<!-- formula-not-decoded -->
Weshow contours of the fitted function Λ πΏ in Figure 4 (left), and the closed-form efficient computational frontier in blue. From this approach, we find that π = 0 46 and π = 0 54-as summarized in Table 2.
## 3.4. Optimal model scaling
We find that the three approaches, despite using different fitting methodologies and different trained models, yield comparable predictions for the optimal scaling in parameters and tokens with FLOPs (shown in Table 2). All three approaches suggest that as compute budget increases, model size and the amount of training data should be increased in approximately equal proportions. The first and second approaches yield very similar predictions for optimal model sizes, as shown in Figure 1 and Figure A3. The third approach predicts even smaller models being optimal at larger compute budgets. We note that the observed points ' πΏ π π· ' for low training FLOPs ( πΆ 6 1 π 21) have larger residuals k πΏ Λ πΏ ' π π· 'k 2 2 than points with higher computational budgets. The fitted model places increased weight on the points with more FLOPs-automatically considering the low-computational budget points as outliers due to the Huber loss. As a consequence of the empirically observed negative curvature in the frontier πΆ ! ππππ‘ (see Appendix E), this results in predicting a lower ππππ‘ than the two other approaches.
In Table 3 we show the estimated number of FLOPs and tokens that would ensure that a model of a given size lies on the compute-optimal frontier. Our findings suggests that the current generation of
Table 2 j Estimated parameter and data scaling with increased training compute. The listed values are the exponents, π and π , on the relationship ππππ‘ / πΆ π and π·πππ‘ / πΆ π . Our analysis suggests a near equal scaling in parameters and data with increasing compute which is in clear contrast to previous work on the scaling of large models. The 10 th and 90 th percentiles are estimated via bootstrapping data (80% of the dataset is sampled 100 times) and are shown in parenthesis.
| Approach | Coeff. π where π πππ‘ / πΆ π | Coeff. π where π· πππ‘ / πΆ π |
|-------------------------------------|------------------------------------------------------------------|------------------------------------------------------------------|
| 1. Minimum over training curves | 0 50 ' 0 488 0 502 ' | 0 50 ' 0 501 0 512 ' |
| 2. IsoFLOP profiles | 0 49 ' 0 462 0 534 ' | 0 51 ' 0 483 0 529 ' |
| 3. Parametric modelling of the loss | 0 46 ' 0 454 0 455 ' | 0 54 ' 0 542 0 543 ' |
| Kaplan et al. (2020) | 0.73 | 0.27 |
Table 3 j Estimated optimal training FLOPs and training tokens for various model sizes. For various model sizes, we show the projections from Approach 1 of how many FLOPs and training tokens would be needed to train compute-optimal models. The estimates for Approach 2 & 3 are similar (shown in Section D.3)
.
| Parameters | FLOPs | FLOPs (in Gopher unit) | Tokens |
|--------------|----------|--------------------------------|----------------|
| 400 Million | 1.92e+19 | 1 29 968 | 8.0 Billion |
| 1 Billion | 1.21e+20 | 1 4 761 | 20.2 Billion |
| 10 Billion | 1.23e+22 | 1 46 | 205.1 Billion |
| 67 Billion | 5.76e+23 | 1 | 1.5 Trillion |
| 175 Billion | 3.85e+24 | 6 7 | 3.7 Trillion |
| 280 Billion | 9.9e+24 | 17 2 | 5.9 Trillion |
| 520 Billion | 3.43e+25 | 59 5 | 11.0 Trillion |
| 1 Trillion | 1.27e+26 | 221 3 | 21.2 Trillion |
| 10 Trillion | 1.3e+28 | 22515 9 | 216.2 Trillion |
large language models are considerably over-sized, given their respective compute budgets, as shown in Figure 1. For example, we find that a 175 billion parameter model should be trained with a compute budget of 4 41 10 24 FLOPs and on over 4.2 trillion tokens. A 280 billion Gopher -like model is the optimal model to train given a compute budget of approximately 10 25 FLOPs and should be trained on 6.8 trillion tokens. Unless one has a compute budget of 10 26 FLOPs (over 250 the compute used to train Gopher ), a 1 trillion parameter model is unlikely to be the optimal model to train. Furthermore, the amount of training data that is projected to be needed is far beyond what is currently used to train large models, and underscores the importance of dataset collection in addition to engineering improvements that allow for model scale. While there is significant uncertainty extrapolating out many orders of magnitude, our analysis clearly suggests that given the training compute budget for many current LLMs, smaller models should have been trained on more tokens to achieve the most performant model.
In Appendix C, we reproduce the IsoFLOP analysis on two additional datasets: C4 (Raffel et al., 2020a) and GitHub code (Rae et al., 2021). In both cases we reach the similar conclusion that model size and number of training tokens should be scaled in equal proportions.
## 4. Chinchilla
Based on our analysis in Section 3, the optimal model size for the Gopher compute budget is somewhere between 40 and 70 billion parameters. We test this hypothesis by training a model on the larger end of this range-70B parameters-for 1.4T tokens, due to both dataset and computational efficiency considerations. In this section we compare this model, which we call Chinchilla , to Gopher and other LLMs. Both Chinchilla and Gopher have been trained for the same number of FLOPs but differ in the size of the model and the number of training tokens.
While pre-training a large language model has a considerable compute cost, downstream finetuning and inference also make up substantial compute usage (Rae et al., 2021). Due to being 4 smaller than Gopher , both the memory footprint and inference cost of Chinchilla are also smaller.
## 4.1. Model and training details
The full set of hyperparameters used to train Chinchilla are given in Table 4. Chinchilla uses the same model architecture and training setup as Gopher with the exception of the differences listed below.
- We train Chinchilla on MassiveText (the same dataset as Gopher ) but use a slightly different subset distribution (shown in Table A1) to account for the increased number of training tokens.
- We train Chinchilla with a slightly modified SentencePiece (Kudo and Richardson, 2018) tokenizer that does not apply NFKC normalisation. The vocabulary is very similar- 94.15% of tokens are the same as those used for training Gopher . We find that this particularly helps with the representation of mathematics and chemistry, for example.
- We use AdamW (Loshchilov and Hutter, 2019) for Chinchilla rather than Adam (Kingma and Ba, 2014) as this improves the language modelling loss and the downstream task performance after finetuning. 8
- Whilst the forward and backward pass are computed in bfloat16 , we store a float32 copy of the weights in the distributed optimiser state (Rajbhandari et al., 2020). See Lessons Learned from Rae et al. (2021) for additional details.
In Appendix G we show the impact of the various optimiser related changes between Chinchilla and Gopher . All models in this analysis have been trained on TPUv3/TPUv4 (Jouppi et al., 2017) with JAX (Bradbury et al., 2018) and Haiku (Hennigan et al., 2020). We include a Chinchilla model card (Mitchell et al., 2019) in Table A8.
| Model | Layers | Number Heads | Key/Value Size | d model | Max LR | Batch Size |
|----------------|----------|----------------|------------------|-----------|--------------------------|--------------|
| Gopher 280B | 80 | 128 | 128 | 16,384 | 4 10 5 | 3M ! 6M |
| Chinchilla 70B | 80 | 64 | 128 | 8,192 | 1 10 4 | 1.5M ! 3M |
Table 4 j Chinchilla architecture details. We list the number of layers, the key/value size, the bottleneck activation size d model , the maximum learning rate, and the training batch size (# tokens). The feed-forward size is always set to 4 dmodel . Note that we double the batch size midway through training for both Chinchilla and Gopher .
8 Interestingly , a model trained with AdamW only passes the training performance of a model trained with Adam around 80% of the way through the cosine cycle, though the ending performance is notably better- see Figure A7
Table 5 j All evaluation tasks. We evaluate Chinchilla on a collection of language modelling along with downstream tasks. We evaluate on largely the same tasks as in Rae et al. (2021), to allow for direct comparison.
| | # Tasks | Examples |
|-----------------------|-----------|---------------------------------------------------------------------------------------------|
| Language Modelling | 20 | WikiText-103, The Pile: PG-19, arXiv, FreeLaw, |
| Reading Comprehension | 3 | RACE-m, RACE-h, LAMBADA |
| Question Answering | 3 | Natural Questions, TriviaQA, TruthfulQA |
| Common Sense | 5 | HellaSwag, Winogrande, PIQA, SIQA, BoolQ |
| MMLU | 57 | High School Chemistry, Astronomy, Clinical Knowledge, |
| BIG-bench | 62 | Causal Judgement, Epistemic Reasoning, Temporal Sequences, |
## 4.2. Results
We perform an extensive evaluation of Chinchilla , comparing against various large language models. We evaluate on a large subset of the tasks presented in Rae et al. (2021), shown in Table 5. As the focus of this work is on optimal model scaling, we included a large representative subset, and introduce a few new evaluations to allow for better comparison to other existing large models. The evaluation details for all tasks are the same as described in Rae et al. (2021).
## 4.2.1. Language modelling
Figure 5 j Pile Evaluation. For the different evaluation sets in The Pile (Gao et al., 2020), we show the bits-per-byte (bpb) improvement (decrease) of Chinchilla compared to Gopher . On all subsets, Chinchilla outperforms Gopher .
<details>
<summary>Image 6 Details</summary>

### Visual Description
\n
## Bar Chart: Decrease in bpb Compared to Gopher
### Overview
This is a bar chart displaying the decrease in bits per byte (bpb) for various datasets when compared to the Gopher language model. The x-axis represents the dataset name, and the y-axis represents the decrease in bpb. The bars are all blue and arranged in ascending order of decrease in bpb.
### Components/Axes
* **X-axis Label:** Dataset Name
* **Y-axis Label:** Decrease in bpb compared to Gopher
* **Y-axis Scale:** 0.00 to 0.10 (approximately)
* **Datasets (X-axis):** pubmed_abstracts, nih_exporter, uspto_backgrounds, pubmed_central, pile_cc, bookcorpus2, stackexchange, opensubtitles, openwebtext2, hackernews, dm_mathematics, arxiv, freelaw, books3, philpapers, github, ubuntu_irc, europarl, gutenberg_pg_19
### Detailed Analysis
The chart shows a clear trend of increasing decrease in bpb as we move from left to right across the datasets.
Here's a breakdown of approximate values, reading from left to right:
* **pubmed_abstracts:** ~0.012
* **nih_exporter:** ~0.014
* **uspto_backgrounds:** ~0.016
* **pubmed_central:** ~0.018
* **pile_cc:** ~0.021
* **bookcorpus2:** ~0.022
* **stackexchange:** ~0.023
* **opensubtitles:** ~0.025
* **openwebtext2:** ~0.028
* **hackernews:** ~0.030
* **dm_mathematics:** ~0.032
* **arxiv:** ~0.034
* **freelaw:** ~0.036
* **books3:** ~0.038
* **philpapers:** ~0.040
* **github:** ~0.044
* **ubuntu_irc:** ~0.062
* **europarl:** ~0.075
* **gutenberg_pg_19:** ~0.095
The largest decrease in bpb is observed for the "gutenberg_pg_19" dataset (~0.095), while the smallest decrease is seen for "pubmed_abstracts" (~0.012).
### Key Observations
* The decrease in bpb is relatively small for the first several datasets (pubmed_abstracts through bookcorpus2), ranging from approximately 0.012 to 0.022.
* There's a noticeable jump in the decrease in bpb between "books3" (~0.038) and "github" (~0.044).
* "gutenberg_pg_19" stands out significantly with the highest decrease in bpb.
* The data appears to be consistently increasing, with no major dips or plateaus.
### Interpretation
The chart demonstrates the effectiveness of the Gopher language model in compressing different types of data. A larger decrease in bpb indicates better compression performance. The fact that "gutenberg_pg_19" (Project Gutenberg, 19th-century literature) exhibits the largest decrease suggests that Gopher is particularly well-suited for compressing text from this domain. Conversely, the smaller decreases for datasets like "pubmed_abstracts" and "nih_exporter" might indicate that these datasets have characteristics that make them less compressible by Gopher, or that Gopher's architecture is less optimized for biomedical text.
The consistent upward trend suggests that Gopher generally performs better on datasets that are more amenable to its compression algorithms. The differences in compression performance across datasets could be due to variations in vocabulary, sentence structure, and the presence of specialized terminology. This information is valuable for understanding the strengths and weaknesses of the Gopher model and for selecting appropriate datasets for training and evaluation.
</details>
Chinchilla significantly outperforms Gopher on all evaluation subsets of The Pile (Gao et al., 2020), as shown in Figure 5. Compared to Jurassic-1 (178B) Lieber et al. (2021), Chinchilla is more performant on all but two subsets-dm\_mathematics and ubuntu\_irc - see Table A5 for a raw bits-per-byte comparison. On Wikitext103 (Merity et al., 2017), Chinchilla achieves a perplexity of 7.16 compared to 7.75 for Gopher . Some caution is needed when comparing Chinchilla with Gopher on these language modelling benchmarks as Chinchilla is trained on 4 more data than Gopher and thus train/test set leakage may artificially enhance the results. We thus place more emphasis on other
Table 6 j Massive Multitask Language Understanding (MMLU). We report the average 5-shot accuracy over 57 tasks with model and human accuracy comparisons taken from Hendrycks et al. (2020). We also include the average prediction for state of the art accuracy in June 2022/2023 made by 73 competitive human forecasters in Steinhardt (2021).
| Random | 25.0% |
|----------------------------------|---------|
| Average human rater | 34.5% |
| GPT-3 5-shot | 43.9% |
| Gopher 5-shot | 60.0% |
| Chinchilla 5-shot | 67.6% |
| Average human expert performance | 89.8% |
| June 2022 Forecast | 57.1% |
| June 2023 Forecast | 63.4% |
tasks for which leakage is less of a concern, such as MMLU (Hendrycks et al., 2020) and BIG-bench (BIG-bench collaboration, 2021) along with various closed-book question answering and common sense analyses.
## 4.2.2. MMLU
The Massive Multitask Language Understanding (MMLU) benchmark (Hendrycks et al., 2020) consists of a range of exam-like questions on academic subjects. In Table 6, we report Chinchilla 's average 5-shot performance on MMLU (the full breakdown of results is shown in Table A6). On this benchmark, Chinchilla significantly outperforms Gopher despite being much smaller, with an average accuracy of 67.6% (improving upon Gopher by 7.6%). Remarkably, Chinchilla even outperforms the expert forecast for June 2023 of 63.4% accuracy (see Table 6) (Steinhardt, 2021). Furthermore, Chinchilla achieves greater than 90% accuracy on 4 different individual tasks-high\_school\_gov\_and\_politics, international\_law, sociology , and us\_foreign\_policy . To our knowledge, no other model has achieved greater than 90% accuracy on a subset.
In Figure 6, we show a comparison to Gopher broken down by task. Overall, we find that Chinchilla improves performance on the vast majority of tasks. On four tasks ( college\_mathematics, econometrics, moral\_scenarios , and formal\_logic ) Chinchilla underperforms Gopher , and there is no change in performance on two tasks.
## 4.2.3. Reading comprehension
On the final word prediction dataset LAMBADA (Paperno et al., 2016), Chinchilla achieves 77.4% accuracy, compared to 74.5% accuracy from Gopher and 76.6% from MT-NLG 530B (see Table 7). On RACE-h and RACE-m (Lai et al., 2017), Chinchilla greatly outperforms Gopher , improving accuracy by more than 10% in both cases-see Table 7.
## 4.2.4. BIG-bench
We analysed Chinchilla on the same set of BIG-bench tasks (BIG-bench collaboration, 2021) reported in Rae et al. (2021). Similar to what we observed in MMLU, Chinchilla outperforms Gopher on the vast majority of tasks (see Figure 7). We find that Chinchilla improves the average performance by 10.7%, reaching an accuracy of 65.1% versus 54.4% for Gopher . Of the 62 tasks we consider, Chinchilla performs worse than Gopher on only four-crash\_blossom, dark\_humor\_detection,
Figure 6 j MMLU results compared to Gopher We find that Chinchilla outperforms Gopher by 7.6% on average (see Table 6) in addition to performing better on 51/57 individual tasks, the same on 2/57, and worse on only 4/57 tasks.
<details>
<summary>Image 7 Details</summary>

### Visual Description
\n
## Bar Chart: Relative Improvement over GPT-3
### Overview
This is a horizontal bar chart displaying the relative improvement of a new model (presumably the one generating this analysis) over GPT-3 across various academic subjects and fields of study. The x-axis represents the relative improvement (percentage), and the y-axis lists the subjects. Bars extending to the right of the zero line indicate improvement, while those to the left indicate a decrease in performance relative to GPT-3.
### Components/Axes
* **X-axis Title:** "Relative Improvement over GPT-3"
* **X-axis Scale:** Ranges from approximately -12 to 32, with markings at increments of 5.
* **Y-axis:** Lists various subjects/fields of study.
* **Bar Colors:** Predominantly blue, with a few bars in yellow/orange indicating negative improvement.
* **No Legend:** The color coding is implicit: blue = improvement, yellow/orange = decline.
### Detailed Analysis
The subjects and their approximate relative improvement values are as follows (reading from top to bottom):
1. **college_mathematics:** Approximately -6
2. **moral_scenarios:** Approximately -3
3. **econometrics:** Approximately -2
4. **formal_logic:** Approximately -1
5. **medical_genetics:** Approximately 0
6. **machine_learning:** Approximately 1
7. **public_relations:** Approximately 2
8. **global_ethics:** Approximately 3
9. **business_ethics:** Approximately 4
10. **college_engineering:** Approximately 5
11. **computer_science:** Approximately 6
12. **high_school_history:** Approximately 7
13. **high_school_psychology:** Approximately 8
14. **high_school_management:** Approximately 9
15. **high_school_marketing:** Approximately 10
16. **high_school_sociology:** Approximately 11
17. **high_school_macroeconomics:** Approximately 12
18. **high_school_politics:** Approximately 13
19. **high_school_european_history:** Approximately 14
20. **nutrition:** Approximately 15
21. **college_medicine:** Approximately 16
22. **logical_fallacies:** Approximately 17
23. **miscellaneous:** Approximately 18
24. **jurisprudence:** Approximately 19
25. **professional_psychology:** Approximately 20
26. **clinical_knowledge:** Approximately 21
27. **high_school_geography:** Approximately 22
28. **college_chemistry:** Approximately 23
29. **college_biology:** Approximately 24
30. **high_school_history:** Approximately 25
31. **us_foreign_policy:** Approximately 26
32. **philosophy:** Approximately 27
33. **virology:** Approximately 28
34. **moral_disputes:** Approximately 29
35. **human_aging:** Approximately 30
36. **security_studies:** Approximately 31
37. **computer_law:** Approximately 32
38. **international_politics:** Approximately 32
39. **high_school_economics:** Approximately 32
40. **professional_accounting:** Approximately 32
41. **professional_medicine:** Approximately 32
42. **elementary_mathematics:** Approximately 32
43. **abstract_algebra:** Approximately 32
44. **human_anatomy:** Approximately 32
45. **professional_sexuality:** Approximately 32
46. **college_physics:** Approximately 32
47. **high_school_physics:** Approximately 32
48. **conceptual_physics:** Approximately 32
**Trends:**
* The new model generally performs better than GPT-3 in most subjects, as indicated by the predominantly positive (blue) bars.
* The improvement is most significant in fields like conceptual physics, high school physics, college physics, and other science/mathematics related fields.
* The model performs worse than GPT-3 in a few areas, notably college mathematics, moral scenarios, and econometrics.
### Key Observations
* The largest positive improvement is observed in several subjects clustered towards the right side of the chart, all showing a relative improvement of approximately 32%.
* The negative improvements are relatively small, with the largest decrease being around -6%.
* There's a clear grouping of subjects where the model shows substantial improvement, particularly in the sciences and mathematics.
### Interpretation
The data suggests that the new model represents a significant advancement over GPT-3, particularly in scientific and mathematical domains. The model demonstrates a strong ability to handle complex reasoning and knowledge in these areas. The negative improvements in subjects like college mathematics and moral scenarios could indicate areas where GPT-3 already possesses a strong baseline or where the new model struggles with abstract or nuanced concepts. The clustering of high improvement scores in science and math suggests a potential specialization or targeted training of the new model in these fields. The chart provides a valuable comparative analysis of the model's performance across a diverse range of subjects, highlighting its strengths and weaknesses. The consistent high performance in the rightmost subjects suggests a strong underlying capability in those areas. The relatively small negative improvements suggest that the model doesn't *fail* in those areas, but simply doesn't surpass GPT-3's existing performance.
</details>
Table 7 j Reading comprehension. On RACE-h and RACE-m (Lai et al., 2017), Chinchilla considerably improves performance over Gopher . Note that GPT-3 and MT-NLG 530B use a different prompt format than we do on RACE-h/m, so results are not comparable to Gopher and Chinchilla . On LAMBADA (Paperno et al., 2016), Chinchilla outperforms both Gopher and MT-NLG 530B.
| | Chinchilla | Gopher | GPT-3 | MT-NLG 530B |
|-------------------|--------------|----------|---------|---------------|
| LAMBADA Zero-Shot | 77.4 | 74.5 | 76.2 | 76.6 |
| RACE-m Few-Shot | 86.8 | 75.1 | 58.1 | - |
| RACE-h Few-Shot | 82.3 | 71.6 | 46.8 | 47.9 |
mathematical\_induction and logical\_args . Full accuracy results for Chinchilla can be found in Table A7.
## 4.2.5. Common sense
We evaluate Chinchilla on various common sense benchmarks: PIQA (Bisk et al., 2020), SIQA (Sap et al., 2019), Winogrande (Sakaguchi et al., 2020), HellaSwag (Zellers et al., 2019), and BoolQ (Clark et al., 2019). We find that Chinchilla outperforms both Gopher and GPT-3 on all tasks and outperforms MT-NLG 530B on all but one task-see Table 8.
On TruthfulQA (Lin et al., 2021), Chinchilla reaches 43.6%, 58.5%, and 66.7% accuracy with 0-shot, 5-shot, and 10-shot respectively. In comparison, Gopher achieved only 29.5% 0-shot and 43.7% 10-shot accuracy. In stark contrast with the findings of Lin et al. (2021), the large improvements (14.1% in 0-shot accuracy) achieved by Chinchilla suggest that better modelling of the pre-training data alone can lead to substantial improvements on this benchmark.
Figure 7 j BIG-bench results compared to Gopher Chinchilla out performs Gopher on all but four BIG-bench tasks considered. Full results are in Table A7.
<details>
<summary>Image 8 Details</summary>

### Visual Description
## Bar Chart: Relative Improvement over Gopher
### Overview
This is a bar chart comparing the relative improvement of a model (presumably a large language model) across a variety of tasks, relative to a baseline model named "Gopher". The chart displays positive and negative improvements, with the y-axis representing the percentage improvement. The x-axis lists numerous natural language processing (NLP) tasks.
### Components/Axes
* **Y-axis Title:** "Relative Improvement over Gopher" (Scale ranges from approximately -30 to 120, with increments of 10).
* **X-axis Title:** Lists various NLP tasks.
* **Bars:** Represent the relative improvement for each task. Bars are colored either orange or blue.
* **Legend:** Located in the top-left corner.
* Orange: Represents negative relative improvement (below 0).
* Blue: Represents positive relative improvement (above 0).
### Detailed Analysis
The chart contains 35 distinct NLP tasks listed along the x-axis. I will analyze the data by describing the trend and then providing approximate values.
1. **crash_blossom:** Approximately -10.
2. **human_organs:** Approximately -5.
3. **mathematica:** Approximately 5.
4. **general_knowledge_induction:** Approximately 10.
5. **args_multiple_choice:** Approximately 15.
6. **formal_fallacies:** Approximately 20.
7. **sarcasm_negotiation:** Approximately -10.
8. **known_ambiguity:** Approximately -15.
9. **sentence_navigation:** Approximately -20.
10. **moral_permissibility:** Approximately -5.
11. **interfailed_polarity:** Approximately 0.
12. **inflated_polarization:** Approximately 5.
13. **hyperbaton:** Approximately 10.
14. **similarities_abstraction:** Approximately 15.
15. **miscon_reasoning:** Approximately 20.
16. **epistemic_reasoning:** Approximately 25.
17. **fame_fantasy_winnowing:** Approximately 30.
18. **movie_dialog:** Approximately 35.
19. **discourse_marker:** Approximately 40.
20. **novel_concepts:** Approximately 45.
21. **hindu_strategy:** Approximately 50.
22. **causal_judgement:** Approximately 55.
23. **alignment_colored_objects:** Approximately 60.
24. **phrase_relatedness:** Approximately 65.
25. **reasoning_about_understanding:** Approximately 70.
26. **penguins_in_a_table:** Approximately 75.
27. **date_ruins_detection:** Approximately 80.
28. **figure_of_speech_detection:** Approximately 85.
29. **disambiguation_q:** Approximately 90.
30. **implicatures:** Approximately 95.
31. **ruin_detection:** Approximately 100.
32. **s_snakes:** Approximately 105.
33. **logical_fallacy_detection:** Approximately 110.
34. **anach_puzzle:** Approximately 5.
35. **rigid_sense:** Approximately -10.
36. **analytic_entailment:** Approximately 0.
37. **logic_grammar:** Approximately 10.
38. **question_selection:** Approximately 15.
39. **nonsense_words:** Approximately 20.
40. **empirical_judgements:** Approximately 25.
41. **sports_understanding:** Approximately 30.
42. **implicit_ritual:** Approximately 35.
43. **physical_reasoning:** Approximately 40.
44. **presupposition_fables:** Approximately 45.
45. **movie_recommendations:** Approximately 50.
46. **understanding_pos:** Approximately 55.
47. **temporal_sequence:** Approximately 60.
48. **logical_odd_sentence:** Approximately 65.
49. **identity_complemention:** Approximately 70.
50. **gre_reading_similarity:** Approximately 75.
51. **analogical_similarity:** Approximately 80.
**Trends:**
* The model shows significant positive improvement on tasks related to reasoning, logical deduction, and understanding complex relationships (e.g., implicatures, ruin detection, logical fallacy detection).
* The model performs worse than Gopher on tasks involving ambiguity, sarcasm, and potentially more nuanced understanding of language (e.g., sarcasm negotiation, known ambiguity, sentence navigation).
* There is a wide range of performance, indicating the model excels in some areas while struggling in others.
### Key Observations
* The largest positive improvement is observed in "ruin_detection" (approximately 100%).
* The largest negative improvement is observed in "sentence_navigation" (approximately -20%).
* There's a clear clustering of positive improvements in the latter half of the chart, suggesting the model is better at tasks requiring deeper semantic understanding.
* The tasks with negative improvements tend to be those that require understanding context, nuance, or social cues.
### Interpretation
The data suggests that the model has made substantial progress in areas requiring logical reasoning and knowledge application, surpassing the performance of Gopher. However, it still lags behind in tasks that demand a more sophisticated understanding of human language, including ambiguity, sarcasm, and contextual awareness. This indicates that while the model is strong at processing information and drawing inferences, it struggles with the subtleties of human communication. The wide variance in performance across tasks highlights the challenges of building a truly general-purpose language model. The model's strengths in tasks like "ruin_detection" and "logical_fallacy_detection" could be valuable in applications requiring critical thinking and error identification, while its weaknesses in areas like "sarcasm_negotiation" suggest caution when deploying it in contexts where understanding intent is crucial. The chart provides a valuable diagnostic tool for identifying areas where further model development is needed.
</details>
## 4.2.6. Closed-book question answering
Results on closed-book question answering benchmarks are reported in Table 9. On the Natural Questions dataset (Kwiatkowski et al., 2019), Chinchilla achieves new closed-book SOTA accuracies: 31.5% 5-shot and 35.5% 64-shot, compared to 21% and 28% respectively, for Gopher . On TriviaQA (Joshi et al., 2017) we show results for both the filtered (previously used in retrieval and open-book work) and unfiltered set (previously used in large language model evaluations). In both cases, Chinchilla substantially out performs Gopher . On the filtered version, Chinchilla lags behind the open book SOTA (Izacard and Grave, 2020) by only 7.9%. On the unfiltered set, Chinchilla outperforms GPT-3-see Table 9.
## 4.2.7. Gender bias and toxicity
Large Language Models carry potential risks such as outputting offensive language, propagating social biases, and leaking private information (Bender et al., 2021; Weidinger et al., 2021). We expect Chinchilla to carry risks similar to Gopher because Chinchilla is trained on the same data,
Table 8 j Zero-shot comparison on Common Sense benchmarks. We show a comparison between Chinchilla , Gopher , and MT-NLG 530B on various Common Sense benchmarks. We see that Chinchilla matches or outperforms Gopher and GPT-3 on all tasks. On all but one Chinchilla outperforms the much larger MT-NLG 530B model.
| | Chinchilla | Gopher | GPT-3 | MT-NLG 530B | Supervised SOTA |
|------------|--------------|----------|---------|---------------|-------------------|
| HellaSWAG | 80.8% | 79.2% | 78.9% | 80.2% | 93.9% |
| PIQA | 81.8% | 81.8% | 81.0% | 82.0% | 90.1% |
| Winogrande | 74.9% | 70.1% | 70.2% | 73.0% | 91.3% |
| SIQA | 51.3% | 50.6% | - | - | 83.2% |
| BoolQ | 83.7 % | 79.3% | 60.5% | 78.2% | 91.4% |
| | Method | Chinchilla | Gopher | GPT-3 | SOTA (open book) |
|-----------------------------|-----------------------|-------------------|-------------------|---------------|--------------------|
| Natural Questions (dev) | 0-shot 5-shot 64-shot | 16.6% 31.5% 35.5% | 10.1% 24.5% 28.2% | 14.6% - 29.9% | 54.4% |
| TriviaQA (unfiltered, test) | 0-shot | 67.0% | 52.8% | 64.3% - | - |
| | 5-shot 64-shot | 73.2% 72.3% | 63.6% | 71.2% | - |
| | | 55.4% | 61.3% | | - |
| TriviaQA (filtered, dev) | 0-shot | | 43.5% | - | 72.5% |
| TriviaQA (filtered, dev) | 5-shot | 64.1% | 57.0% | - | 72.5% |
| TriviaQA (filtered, dev) | 64-shot | 64.6% | 57.2% | - | 72.5% |
Table 9 j Closed-book question answering. For Natural Questions (Kwiatkowski et al., 2019) and TriviaQA (Joshi et al., 2017), Chinchilla outperforms Gopher in all cases. On Natural Questions, Chinchilla outperforms GPT-3. On TriviaQA we show results on two different evaluation sets to allow for comparison to GPT-3 and to open book SOTA (FiD + Distillation (Izacard and Grave, 2020)).
albeit with slightly different relative weights, and because it has a similar architecture. Here, we examine gender bias (particularly gender and occupation bias) and generation of toxic language. We select a few common evaluations to highlight potential issues, but stress that our evaluations are not comprehensive and much work remains to understand, evaluate, and mitigate risks in LLMs.
Gender bias. As discussed in Rae et al. (2021), large language models reflect contemporary and historical discourse about different groups (such as gender groups) from their training dataset, and we expect the same to be true for Chinchilla . Here, we test if potential gender and occupation biases manifest in unfair outcomes on coreference resolutions, using the Winogender dataset (Rudinger et al., 2018) in a zero-shot setting. Winogender tests whether a model can correctly determine if a pronoun refers to different occupation words. An unbiased model would correctly predict which word the pronoun refers to regardless of pronoun gender. We follow the same setup as in Rae et al. (2021) (described further in Section H.3).
As shown in Table 10, Chinchilla correctly resolves pronouns more frequently than Gopher across all groups. Interestingly, the performance increase is considerably smaller for male pronouns (increase of 3.2%) than for female or neutral pronouns (increases of 8.3% and 9.2% respectively). We also consider gotcha examples, in which the correct pronoun resolution contradicts gender stereotypes (determined by labor statistics). Again, we see that Chinchilla resolves pronouns more accurately than Gopher . When breaking up examples by male/female gender and gotcha / not gotcha , the largest improvement is on female gotcha examples (improvement of 10%). Thus, though Chinchilla uniformly overcomes gender stereotypes for more coreference examples than Gopher , the rate of improvement is higher for some pronouns than others, suggesting that the improvements conferred by using a more compute-optimal model can be uneven.
Sample toxicity. Language models are capable of generating toxic language-including insults, hate speech, profanities and threats (Gehman et al., 2020; Rae et al., 2021). While toxicity is an umbrella term, and its evaluation in LMs comes with challenges (Welbl et al., 2021; Xu et al., 2021), automatic classifier scores can provide an indication for the levels of harmful text that a LM generates. Rae et al. (2021) found that improving language modelling loss by increasing the number of model parameters has only a negligible effect on toxic text generation (unprompted); here we analyze
Table 10 j Winogender results. Left: Chinchilla consistently resolves pronouns better than Gopher . Right: Chinchilla performs better on examples which contradict gender stereotypes ( gotcha examples). However, difference in performance across groups suggests Chinchilla exhibits bias.
| | Chinchilla | Gopher | | Chinchilla | Gopher |
|---------|--------------|----------|-------------------|--------------|----------|
| All | 78.3% | 71.4% | Male gotcha | 62.5% | 59.2% |
| Male | 71.2% | 68.0% | Male not gotcha | 80.0% | 76.7% |
| Female | 79.6% | 71.3% | Female gotcha | 76.7% | 66.7% |
| Neutral | 84.2% | 75.0% | Female not gotcha | 82.5% | 75.8% |
whether the same holds true for a lower LM loss achieved via more compute-optimal training. Similar to the protocol of Rae et al. (2021), we generate 25,000 unprompted samples from Chinchilla , and compare their PerspectiveAPI toxicity score distribution to that of Gopher -generated samples. Several summary statistics indicate an absence of major differences: the mean (median) toxicity score for Gopher is 0.081 (0.064), compared to 0.087 (0.066) for Chinchilla , and the 95 th percentile scores are 0.230 for Gopher , compared to 0.238 for Chinchilla . That is, the large majority of generated samples are classified as non-toxic, and the difference between the models is negligible. In line with prior findings (Rae et al., 2021), this suggests that toxicity levels in unconditional text generation are largely independent of the model quality (measured in language modelling loss), i.e. that better models of the training dataset are not necessarily more toxic.
## 5. Discussion & Conclusion
The trend so far in large language model training has been to increase the model size, often without increasing the number of training tokens. The largest dense transformer, MT-NLG 530B, is now over 3 larger than GPT-3's 170 billion parameters from just two years ago. However, this model, as well as the majority of existing large models, have all been trained for a comparable number of tokens-around 300 billion. While the desire to train these mega-models has led to substantial engineering innovation, we hypothesize that the race to train larger and larger models is resulting in models that are substantially underperforming compared to what could be achieved with the same compute budget.
We propose three predictive approaches towards optimally setting model size and training duration, based on the outcome of over 400 training runs. All three approaches predict that Gopher is substantially over-sized and estimate that for the same compute budget a smaller model trained on more data will perform better. We directly test this hypothesis by training Chinchilla , a 70B parameter model, and show that it outperforms Gopher and even larger models on nearly every measured evaluation task.
Whilst our method allows us to make predictions on how to scale large models when given additional compute, there are several limitations. Due to the cost of training large models, we only have two comparable training runs at large scale ( Chinchilla and Gopher ), and we do not have additional tests at intermediate scales. Furthermore, we assume that the efficient computational frontier can be described by a power-law relationship between the compute budget, model size, and number of training tokens. However, we observe some concavity in log ππππ‘ at high compute budgets (see Appendix E). This suggests that we may still be overestimating the optimal size of large models. Finally, the training runs for our analysis have all been trained on less than an epoch of data; future work may consider the multiple epoch regime. Despite these limitations, the comparison of Chinchilla to Gopher validates our performance predictions, that have thus enabled training a better (and more
lightweight) model at the same compute budget.
Though there has been significant recent work allowing larger and larger models to be trained, our analysis suggests an increased focus on dataset scaling is needed. Speculatively, we expect that scaling to larger and larger datasets is only beneficial when the data is high-quality. This calls for responsibly collecting larger datasets with a high focus on dataset quality. Larger datasets will require extra care to ensure train-test set overlap is properly accounted for, both in the language modelling loss but also with downstream tasks. Finally, training for trillions of tokens introduces many ethical and privacy concerns. Large datasets scraped from the web will contain toxic language, biases, and private information. With even larger datasets being used, the quantity (if not the frequency) of such information increases, which makes dataset introspection all the more important. Chinchilla does suffer from bias and toxicity but interestingly it seems less affected than Gopher . Better understanding how performance of large language models and toxicity interact is an important future research question.
While we have applied our methodology towards the training of auto-regressive language models, we expect that there is a similar trade-off between model size and the amount of data in other modalities. As training large models is very expensive, choosing the optimal model size and training steps beforehand is essential. The methods we propose are easy to reproduce in new settings.
## 6. Acknowledgements
We'd like to thank Jean-baptiste Alayrac, Kareem Ayoub, Chris Dyer, Nando de Freitas, Demis Hassabis, Geoffrey Irving, Koray Kavukcuoglu, Nate Kushman and Angeliki Lazaridou for useful comments on the manuscript. We'd like to thank Andy Brock, Irina Higgins, Michela Paganini, Francis Song, and other colleagues at DeepMind for helpful discussions. We are also very grateful to the JAX and XLA team for their support and assistance.
## References
- M. Artetxe, S. Bhosale, N. Goyal, T. Mihaylov, M. Ott, S. Shleifer, X. V. Lin, J. Du, S. Iyer, R. Pasunuru, G. Anantharaman, X. Li, S. Chen, H. Akin, M. Baines, L. Martin, X. Zhou, P. S. Koura, B. O'Horo, J. Wang, L. Zettlemoyer, M. Diab, Z. Kozareva, and V. Stoyanov. Efficient Large Scale Language Modeling with Mixtures of Experts. arXiv:2112.10684, 2021.
- E. M. Bender, T. Gebru, A. McMillan-Major, and S. Shmitchell. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pages 610-623, 2021.
3. BIG-bench collaboration. Beyond the imitation game: Measuring and extrapolating the capabilities of language models. In preparation, 2021. URL https://github.com/google/BIG-bench/ .
- Y. Bisk, R. Zellers, J. Gao, Y. Choi, et al. PIQA: Reasoning about physical commonsense in natural language. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 7432-7439, 2020.
- S. Borgeaud, A. Mensch, J. Hoffmann, T. Cai, E. Rutherford, K. Millican, G. van den Driessche, J.-B. Lespiau, B. Damoc, A. Clark, D. de Las Casas, A. Guy, J. Menick, R. Ring, T. Hennigan, S. Huang, L. Maggiore, C. Jones, A. Cassirer, A. Brock, M. Paganini, G. Irving, O. Vinyals, S. Osindero, K. Simonyan, J. W. Rae, E. Elsen, and L. Sifre. Improving language models by retrieving from trillions of tokens. arXiv 2112.04426, 2021.
- J. Bradbury, R. Frostig, P. Hawkins, M. J. Johnson, C. Leary, D. Maclaurin, G. Necula, A. Paszke, J. VanderPlas, S. Wanderman-Milne, and Q. Zhang. JAX: composable transformations of Python+NumPy programs. 2018. URL http://github.com/google/jax .
- T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan, R. Child, A. Ramesh, D. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin, S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, and D. Amodei. Language models are few-shot learners. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 1877-1901. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/1457c0d6bfcb49674 18bfb8ac142f64a-Paper.pdf .
- S. Bubeck. Convex Optimization: Algorithms and Complexity. Foundations and Trends in Machine Learning, 8(3-4):231-357, 2015. URL http://www.nowpublishers.com/article/Detail s/MAL-050 .
- A. Clark, D. d. l. Casas, A. Guy, A. Mensch, M. Paganini, J. Hoffmann, B. Damoc, B. Hechtman, T. Cai, S. Borgeaud, G. v. d. Driessche, E. Rutherford, T. Hennigan, M. Johnson, K. Millican, A. Cassirer, C. Jones, E. Buchatskaya, D. Budden, L. Sifre, S. Osindero, O. Vinyals, J. Rae, E. Elsen, K. Kavukcuoglu, and K. Simonyan. Unified scaling laws for routed language models, 2022. URL https://arxiv.org/abs/2202.01169 .
- C. Clark, K. Lee, M.-W. Chang, T. Kwiatkowski, M. Collins, and K. Toutanova. Boolq: Exploring the surprising difficulty of natural yes/no questions. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 2924-2936, 2019.
- N. Du, Y. Huang, A. M. Dai, S. Tong, D. Lepikhin, Y. Xu, M. Krikun, Y. Zhou, A. W. Yu, O. Firat, B. Zoph, L. Fedus, M. Bosma, Z. Zhou, T. Wang, Y. E. Wang, K. Webster, M. Pellat, K. Robinson, K. MeierHellstern, T. Duke, L. Dixon, K. Zhang, Q. V. Le, Y. Wu, Z. Chen, and C. Cui. Glam: Efficient scaling of language models with mixture-of-experts, 2021. URL https://arxiv.org/abs/2112.06905 .
- W. Fedus, B. Zoph, and N. Shazeer. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv preprint arXiv:2101.03961, 2021.
- L. Gao, S. Biderman, S. Black, L. Golding, T. Hoppe, C. Foster, J. Phang, H. He, A. Thite, N. Nabeshima, S. Presser, and C. Leahy. The Pile: An 800GB dataset of diverse text for language modeling. arXiv preprint arXiv:2101.00027, 2020.
- S. Gehman, S. Gururangan, M. Sap, Y. Choi, and N. A. Smith. RealToxicityPrompts: Evaluating neural toxic degeneration in language models. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 3356-3369, Online, Nov. 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.301. URL https://aclanthology.org/2 020.findings-emnlp.301 .
- K. Guu, K. Lee, Z. Tung, P. Pasupat, and M.-W. Chang. REALM: Retrieval-augmented language model pre-training, 2020.
- D. Hendrycks, C. Burns, S. Basart, A. Zou, M. Mazeika, D. Song, and J. Steinhardt. Measuring massive multitask language understanding. arXiv preprint arXiv:2009.03300, 2020.
- T. Hennigan, T. Cai, T. Norman, and I. Babuschkin. Haiku: Sonnet for JAX. 2020. URL http: //github.com/deepmind/dm-haiku .
- D. Hernandez, J. Kaplan, T. Henighan, and S. McCandlish. Scaling laws for transfer, 2021.
- P. J. Huber. Robust Estimation of a Location Parameter. The Annals of Mathematical Statistics, 35 (1):73-101, Mar. 1964. ISSN 0003-4851, 2168-8990. doi: 10.1214/aoms/1177703732. URL https://projecteuclid.org/journals/annals-of-mathematical-statistics/vol ume-35/issue-1/Robust-Estimation-of-a-Location-Parameter/10.1214/aoms/11 77703732.full .
- G. Izacard and E. Grave. Distilling knowledge from reader to retriever for question answering, 2020.
- M. Joshi, E. Choi, D. Weld, and L. Zettlemoyer. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. arXiv e-prints, art. arXiv:1705.03551, 2017.
- N. P. Jouppi, C. Young, N. Patil, D. Patterson, G. Agrawal, R. Bajwa, S. Bates, S. Bhatia, N. Boden, A. Borchers, R. Boyle, P.-l. Cantin, C. Chao, C. Clark, J. Coriell, M. Daley, M. Dau, J. Dean, B. Gelb, T. V. Ghaemmaghami, R. Gottipati, W. Gulland, R. Hagmann, C. R. Ho, D. Hogberg, J. Hu, R. Hundt, D. Hurt, J. Ibarz, A. Jaffey , A. Jaworski, A. Kaplan, H. Khaitan, D. Killebrew, A. Koch, N. Kumar, S. Lacy , J. Laudon, J. Law, D. Le, C. Leary, Z. Liu, K. Lucke, A. Lundin, G. MacKean, A. Maggiore, M. Mahony, K. Miller, R. Nagarajan, R. Narayanaswami, R. Ni, K. Nix, T. Norrie, M. Omernick, N. Penukonda, A. Phelps, J. Ross, M. Ross, A. Salek, E. Samadiani, C. Severn, G. Sizikov, M. Snelham, J. Souter, D. Steinberg, A. Swing, M. Tan, G. Thorson, B. Tian, H. Toma, E. Tuttle, V. Vasudevan, R. Walter, W. Wang, E. Wilcox, and D. H. Yoon. In-datacenter performance analysis of a tensor processing unit. In Proceedings of the 44th Annual International Symposium on Computer Architecture, ISCA '17, page 1-12, New York, NY, USA, 2017. Association for Computing Machinery. ISBN 9781450348928. doi: 10.1145/3079856.3080246. URL https://doi.org/10.1145/3079856.3080246 .
- J. Kaplan, S. McCandlish, T. Henighan, T. B. Brown, B. Chess, R. Child, S. Gray, A. Radford, J. Wu, and D. Amodei. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361, 2020.
- D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- T. Kudo and J. Richardson. SentencePiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv preprint arXiv:1808.06226, 2018.
- T. Kwiatkowski, J. Palomaki, O. Redfield, M. Collins, A. Parikh, C. Alberti, D. Epstein, I. Polosukhin, M. Kelcey, J. Devlin, K. Lee, K. N. Toutanova, L. Jones, M.-W. Chang, A. Dai, J. Uszkoreit, Q. Le, and S. Petrov. Natural questions: a benchmark for question answering research. Transactions of the Association of Computational Linguistics, 2019.
- G. Lai, Q. Xie, H. Liu, Y. Yang, and E. Hovy. RACE: Large-scale ReAding comprehension dataset from examinations. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 785-794, Copenhagen, Denmark, Sept. 2017. Association for Computational Linguistics. doi: 10.18653/v1/D17-1082. URL https://aclanthology.org/D17-1082 .
- Y. Levine, N. Wies, O. Sharir, H. Bata, and A. Shashua. The depth-to-width interplay in self-attention. arXiv preprint arXiv:2006.12467, 2020.
- P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, N. Goyal, H. KΓΌttler, M. Lewis, W.-t. Yih, T. RocktΓ€schel, S. Riedel, and D. Kiela. Retrieval-augmented generation for knowledge-intensive nlp tasks. In Advances in Neural Information Processing Systems, volume 33, pages 9459-9474, 2020.
- O. Lieber, O. Sharir, B. Lenz, and Y. Shoham. Jurassic-1: Technical details and evaluation. White Paper. AI21 Labs, 2021.
- S. Lin, J. Hilton, and O. Evans. TruthfulQA: Measuring how models mimic human falsehoods. arXiv preprint arXiv:2109.07958, 2021.
- I. Loshchilov and F. Hutter. Decoupled weight decay regularization. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=Bkg6RiCqY7 .
- S. McCandlish, J. Kaplan, D. Amodei, and O. D. Team. An empirical model of large-batch training, 2018.
- S. Merity, C. Xiong, J. Bradbury, and R. Socher. Pointer sentinel mixture models. International Conference on Learning Representations, 2017.
- M. Mitchell, S. Wu, A. Zaldivar, P. Barnes, L. Vasserman, B. Hutchinson, E. Spitzer, I. D. Raji, and T. Gebru. Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency, pages 220-229, 2019.
- J. Nocedal. Updating Quasi-Newton Matrices with Limited Storage. Mathematics of Computation, 35(151):773-782, 1980. ISSN 0025-5718. doi: 10.2307/2006193. URL https://www.jstor. org/stable/2006193 .
- D. Paperno, G. Kruszewski, A. Lazaridou, Q. N. Pham, R. Bernardi, S. Pezzelle, M. Baroni, G. Boleda, and R. FernΓ‘ndez. The LAMBADA dataset: Word prediction requiring a broad discourse context, 2016.
- J. Rae, S. Borgeaud, T. Cai, K. Millican, J. Hoffmann, F. Song, J. Aslanides, S. Henderson, R. Ring, S. Young, E. Rutherford, T. Hennigan, J. Menick, A. Cassirer, R. Powell, G. van den Driessche, L. A. Hendricks, M. Rauh, P.-S. Huang, A. Glaese, J. Welbl, S. Dathathri, S. Huang, J. Uesato, J. Mellor, I. Higgins, A. Creswell, N. McAleese, A. Wu, E. Elsen, S. Jayakumar, E. Buchatskaya, D. Budden, E. Sutherland, K. Simonyan, M. Paganini, L. Sifre, L. Martens, X. L. Li, A. Kuncoro, A. Nematzadeh, E. Gribovskaya, D. Donato, A. Lazaridou, A. Mensch, J.-B. Lespiau, M. Tsimpoukelli, N. Grigorev, D. Fritz, T. Sottiaux, M. Pajarskas, T. Pohlen, Z. Gong, D. Toyama, C. de Masson d'Autume, Y. Li, T. Terzi, I. Babuschkin, A. Clark, D. de Las Casas, A. Guy, J. Bradbury, M. Johnson, L. Weidinger, I. Gabriel, W. Isaac, E. Lockhart, S. Osindero, L. Rimell, C. Dyer, O. Vinyals, K. Ayoub, J. Stanway, L. Bennett, D. Hassabis, K. Kavukcuoglu, and G. Irving. Scaling language models: Methods, analysis & insights from training Gopher. arXiv 2112.11446, 2021.
- J. W. Rae, A. Potapenko, S. M. Jayakumar, T. P. Lillicrap, K. Choromanski, V. Likhosherstov, D. Dohan, X. Song, A. Gane, T. Sarlos, et al. Compressive transformers for long-range sequence modelling. Advances in Neural Information Processing Systems, 33:6154-6158, 2020.
- C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1-67, 2020a. URL http://jmlr.org/papers/v21/20-074.html .
- C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1-67, 2020b.
- S. Rajbhandari, J. Rasley, O. Ruwase, and Y. He. Zero: Memory optimizations toward training trillion parameter models. In SC20: International Conference for High Performance Computing, Networking, Storage and Analysis, pages 1-16. IEEE, 2020.
- H. Robbins and S. Monro. A Stochastic Approximation Method. The Annals of Mathematical Statistics, 22(3):400-407, Sept. 1951.
- R. Rudinger, J. Naradowsky, B. Leonard, and B. Van Durme. Gender bias in coreference resolution. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, Louisiana, June 2018. Association for Computational Linguistics.
- K. Sakaguchi, R. Le Bras, C. Bhagavatula, and Y. Choi. Winogrande: An adversarial winograd schema challenge at scale. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 8732-8740, 2020.
- M. Sap, H. Rashkin, D. Chen, R. LeBras, and Y. Choi. SocialIQA: Commonsense reasoning about social interactions. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, 2019.
- C. J. Shallue, J. Lee, J. Antognini, J. Sohl-Dickstein, R. Frostig, and G. E. Dahl. Measuring the effects of data parallelism on neural network training. arXiv preprint arXiv:1811.03600, 2018.
- J. W. Siegel and J. Xu. Approximation rates for neural networks with general activation functions. Neural Networks, 128:313-321, Aug. 2020. URL https://www.sciencedirect.com/scienc e/article/pii/S0893608020301891 .
- S. Smith, M. Patwary, B. Norick, P. LeGresley, S. Rajbhandari, J. Casper, Z. Liu, S. Prabhumoye, G. Zerveas, V. Korthikanti, E. Zhang, R. Child, R. Y. Aminabadi, J. Bernauer, X. Song, M. Shoeybi, Y. He, M. Houston, S. Tiwary, and B. Catanzaro. Using Deepspeed and Megatron to Train Megatronturing NLG 530b, A Large-Scale Generative Language Model. arXiv preprint arXiv:2201.11990, 2022.
- J. Steinhardt. Updates and lessons from AI forecasting, 2021. URL https://bounded-regret.g host.io/ai-forecasting/ .
- Y. Tay, M. Dehghani, J. Rao, W. Fedus, S. Abnar, H. W. Chung, S. Narang, D. Yogatama, A. Vaswani, and D. Metzler. Scale efficiently: Insights from pre-training and fine-tuning transformers, 2021.
- R. Thoppilan, D. D. Freitas, J. Hall, N. Shazeer, A. Kulshreshtha, H.-T. Cheng, A. Jin, T. Bos, L. Baker, Y. Du, Y. Li, H. Lee, H. S. Zheng, A. Ghafouri, M. Menegali, Y. Huang, M. Krikun, D. Lepikhin, J. Qin, D. Chen, Y. Xu, Z. Chen, A. Roberts, M. Bosma, Y. Zhou, C.-C. Chang, I. Krivokon, W. Rusch, M. Pickett, K. Meier-Hellstern, M. R. Morris, T. Doshi, R. D. Santos, T. Duke, J. Soraker, B. Zevenbergen, V. Prabhakaran, M. Diaz, B. Hutchinson, K. Olson, A. Molina, E. Hoffman-John, J. Lee, L. Aroyo, R. Rajakumar, A. Butryna, M. Lamm, V. Kuzmina, J. Fenton, A. Cohen, R. Bernstein, R. Kurzweil, B. Aguera-Arcas, C. Cui, M. Croak, E. Chi, and Q. Le. LaMDA: Language models for dialog applications, 2022.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ε. Kaiser, and I. Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998-6008, 2017.
- L. Weidinger, J. Mellor, M. Rauh, C. Griffin, J. Uesato, P.-S. Huang, M. Cheng, M. Glaese, B. Balle, A. Kasirzadeh, Z. Kenton, S. Brown, W. Hawkins, T. Stepleton, C. Biles, A. Birhane, J. Haas, L. Rimell, L. A. Hendricks, W. Isaac, S. Legassick, G. Irving, and I. Gabriel. Ethical and social risks of harm from language models. arXiv submission, 2021.
- J. Welbl, A. Glaese, J. Uesato, S. Dathathri, J. Mellor, L. A. Hendricks, K. Anderson, P. Kohli, B. Coppin, and P.-S. Huang. Challenges in detoxifying language models. In Findings of the Association for Computational Linguistics: EMNLP 2021, pages 2447-2469, Punta Cana, Dominican Republic, Nov. 2021. Association for Computational Linguistics. URL https://aclanthology.org/2021. findings-emnlp.210 .
- A. Xu, E. Pathak, E. Wallace, S. Gururangan, M. Sap, and D. Klein. Detoxifying language models risks marginalizing minority voices. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2390-2397, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021 .naacl-main.190. URL https://aclanthology.org/2021.naacl-main.190 .
- G. Yang, E. J. Hu, I. Babuschkin, S. Sidor, X. Liu, D. Farhi, N. Ryder, J. Pachocki, W. Chen, and J. Gao. Tuning large neural networks via zero-shot hyperparameter transfer. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, editors, Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=Bx6qKuBM2AD .
- R. Zellers, A. Holtzman, Y. Bisk, A. Farhadi, and Y. Choi. HellaSwag: Can a machine really finish your sentence? In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019.
- G. Zhang, L. Li, Z. Nado, J. Martens, S. Sachdeva, G. Dahl, C. Shallue, and R. B. Grosse. Which algorithmic choices matter at which batch sizes? insights from a noisy quadratic model. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'AlchΓ©-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https: //proceedings.neurips.cc/paper/2019/file/e0eacd983971634327ae1819ea8b621 4-Paper.pdf .
- B. Zoph, I. Bello, S. Kumar, N. Du, Y. Huang, J. Dean, N. Shazeer, and W. Fedus. Designing effective sparse expert models, 2022.
## A. Training dataset
In Table A1 we show the training dataset makeup used for Chinchilla and all scaling runs. Note that both the MassiveWeb and Wikipedia subsets are both used for more than one epoch.
| | Disk Size | Documents | Sampling proportion | Epochs in 1.4T tokens |
|------------|-------------|-------------|-----------------------|-------------------------|
| MassiveWeb | 1.9 TB | 604M | 45% (48%) | 1.24 |
| Books | 2.1 TB | 4M | 30% (27%) | 0.75 |
| C4 | 0.75 TB | 361M | 10% (10%) | 0.77 |
| News | 2.7 TB | 1.1B | 10% (10%) | 0.21 |
| GitHub | 3.1 TB | 142M | 4% (3%) | 0.13 |
| Wikipedia | 0.001 TB | 6M | 1% (2%) | 3.4 |
Table A1 j MassiveText data makeup. For each subset of MassiveText , we list its total disk size, the number of documents and the sampling proportion used during training-we use a slightly different distribution than in Rae et al. (2021) (shown in parenthesis). In the rightmost column show the number of epochs that are used in 1.4 trillion tokens.
## B. Optimal cosine cycle length
One key assumption is made on the cosine cycle length and the corresponding learning rate drop (we use a 10 learning rate decay in line with Rae et al. (2021)). 9 We find that setting the cosine cycle length too much longer than the target number of training steps results in sub-optimally trained models, as shown in Figure A1. As a result, we assume that an optimally trained model will have the cosine cycle length correctly calibrated to the maximum number of steps, given the FLOP budget; we follow this rule in our main analysis.
## C. Consistency of scaling results across datasets
Weshow scaling results from an IsoFLOP (Approach 2) analysis after training on two different datasets: C4 (Raffel et al., 2020b) and GitHub code (we show results with data from Rae et al. (2021)), results are shown in Table A2. For both set of experiments using subsets of MassiveText , we use the same tokenizer as the MassiveText experiments.
Wefind that the scaling behaviour on these datasets is very similar to what we found on MassiveText , as shown in Figure A2 and Table A2. This suggests that our results are independent of the dataset as long as one does not train for more than one epoch.
9 We find the difference between decaying by 10 and decaying to 0.0 (over the same number of steps) to be small, though decaying by a factor of 10 to be slightly more performant. Decaying by less (5 ) is clearly worse.
## Appendix
Figure A1 j Grid over cosine cycle length. We show 6 curves with the cosine cycle length set to 1, 1.1, 1.25, 1.5, 2, and 5 longer than the target number of training steps. When the cosine cycle length is too long, and the learning rate does not drop appropriately, then performance is impaired. We find that overestimating the number of training steps beyond 25% leads to clear drops in performance. We show results where we have set the number of training steps to two different values (top and bottom).
<details>
<summary>Image 9 Details</summary>

### Visual Description
\n
## Line Charts: Training Dynamics with Varying Cosine Cycle Lengths
### Overview
The image presents six line charts arranged in a 2x3 grid, visualizing the training dynamics of a model under different cosine cycle lengths. Each chart displays a different metric (Learning Rate/Max LR, Training Loss, and C4 Loss) against the number of Million Sequences. The charts aim to compare the impact of varying cosine cycle lengths on the training process.
### Components/Axes
Each chart shares the following components:
* **X-axis:** "Million Sequences" ranging from 0 to approximately 8 in the top row and 0 to 12.5 in the bottom row.
* **Y-axis:** Varies depending on the chart:
* Top-left: "Learning Rate/Max LR" ranging from 0 to 1.0.
* Top-center & Bottom-center: "Training Loss" ranging from 2.70 to 3.00.
* Top-right & Bottom-right: "C4 Loss" ranging from 2.80 to 3.20.
* **Legend:** Located in the top-right corner of each chart, listing the "Cosine Cycle Length" values:
* 1.0x num. steps (Blue)
* 1.1x num. steps (Green)
* 1.25x num. steps (Purple)
* 1.5x num. steps (Orange)
* 2.0x num. steps (Pink)
* 5.0x num. steps (Red)
### Detailed Analysis or Content Details
**Top Row:**
* **Learning Rate/Max LR (Top-Left):** All lines start at 1.0 and decrease towards 0. The blue (1.0x) line shows the steepest decline, followed by green (1.1x), purple (1.25x), orange (1.5x), pink (2.0x), and red (5.0x) exhibiting progressively slower declines.
* At 8 Million Sequences:
* Blue (1.0x): ~0.05
* Green (1.1x): ~0.15
* Purple (1.25x): ~0.25
* Orange (1.5x): ~0.35
* Pink (2.0x): ~0.50
* Red (5.0x): ~0.75
* **Training Loss (Top-Center):** All lines start around 2.95 and decrease. The blue (1.0x) line shows the fastest decrease, followed by green (1.1x), purple (1.25x), orange (1.5x), pink (2.0x), and red (5.0x). There are noticeable fluctuations in all lines, particularly between 4 and 6 Million Sequences.
* At 8 Million Sequences:
* Blue (1.0x): ~2.72
* Green (1.1x): ~2.75
* Purple (1.25x): ~2.78
* Orange (1.5x): ~2.82
* Pink (2.0x): ~2.86
* Red (5.0x): ~2.90
* **C4 Loss (Top-Right):** All lines start around 3.15 and decrease. The blue (1.0x) line shows the fastest decrease, followed by green (1.1x), purple (1.25x), orange (1.5x), pink (2.0x), and red (5.0x). Similar fluctuations are observed as in the Training Loss chart.
* At 8 Million Sequences:
* Blue (1.0x): ~2.85
* Green (1.1x): ~2.88
* Purple (1.25x): ~2.92
* Orange (1.5x): ~2.95
* Pink (2.0x): ~2.98
* Red (5.0x): ~3.05
**Bottom Row:**
* **Learning Rate/Max LR (Bottom-Left):** Similar trend to the top-left chart, but extending to 12.5 Million Sequences.
* At 12.5 Million Sequences:
* Blue (1.0x): ~0.02
* Green (1.1x): ~0.08
* Purple (1.25x): ~0.15
* Orange (1.5x): ~0.25
* Pink (2.0x): ~0.40
* Red (5.0x): ~0.65
* **Training Loss (Bottom-Center):** Similar trend to the top-center chart, extending to 12.5 Million Sequences.
* At 12.5 Million Sequences:
* Blue (1.0x): ~2.70
* Green (1.1x): ~2.72
* Purple (1.25x): ~2.75
* Orange (1.5x): ~2.78
* Pink (2.0x): ~2.83
* Red (5.0x): ~2.88
* **C4 Loss (Bottom-Right):** Similar trend to the top-right chart, extending to 12.5 Million Sequences.
* At 12.5 Million Sequences:
* Blue (1.0x): ~2.80
* Green (1.1x): ~2.83
* Purple (1.25x): ~2.87
* Orange (1.5x): ~2.91
* Pink (2.0x): ~2.95
* Red (5.0x): ~3.02
### Key Observations
* Shorter cosine cycle lengths (1.0x and 1.1x) consistently lead to faster decreases in Learning Rate and Loss metrics.
* Longer cosine cycle lengths (2.0x and 5.0x) result in slower decreases and higher final loss values.
* All charts exhibit fluctuations in the loss curves, suggesting instability or oscillations during training.
* The bottom row charts, extending to 12.5 Million Sequences, show that the differences in performance between different cycle lengths become more pronounced over longer training periods.
### Interpretation
The data suggests that the choice of cosine cycle length significantly impacts the training dynamics of the model. Shorter cycle lengths promote faster initial learning and lower final loss values, but may also lead to instability as indicated by the fluctuations in the loss curves. Longer cycle lengths provide more stable training but at the cost of slower learning and potentially higher final loss.
The relationship between the metrics is clear: as the Learning Rate decreases (due to the cosine schedule), the Training Loss and C4 Loss also decrease. The rate at which these metrics change is influenced by the cosine cycle length.
The fluctuations in the loss curves could be due to various factors, such as the learning rate being too high, the batch size being too small, or the model architecture being complex. Further investigation would be needed to determine the root cause of these oscillations. The extended training in the bottom row charts highlights the importance of considering long-term training behavior when selecting a cosine cycle length. A shorter cycle length might initially appear superior, but its instability could prevent it from reaching optimal performance over extended training.
</details>
Figure A2 j C4 and GitHub IsoFLOP curves. Using the C4 dataset (Raffel et al., 2020b) and a GitHub dataset (Rae et al., 2021), we generate 4 IsoFLOP profiles and show the parameter and token count scaling, as in Figure 3. Scaling coefficients are shown in Table A2.
<details>
<summary>Image 10 Details</summary>

### Visual Description
## Charts: Scaling Laws for Neural Language Models
### Overview
The image presents six charts arranged in a 2x3 grid, illustrating scaling laws for neural language models. The charts explore the relationship between model parameters, FLOPs (floating point operations), tokens processed, and training loss on two datasets: C4 and GitHub. Each chart type (loss vs. parameters, parameters vs. FLOPs, tokens vs. FLOPs) is presented for both datasets. Different model sizes (1e19, 1e20, 6e20, 1e21) are represented by different colored lines.
### Components/Axes
* **Chart 1 (Top-Left):** C4 Training Loss vs. Parameters
* X-axis: Parameters (100M to 6B, logarithmic scale)
* Y-axis: C4 Training Loss (2.0 to 3.2, linear scale)
* **Chart 2 (Top-Center):** Parameters vs. FLOPs
* X-axis: FLOPs (10^17 to 10^23, logarithmic scale)
* Y-axis: Parameters (100M to 1T, logarithmic scale)
* **Chart 3 (Top-Right):** Tokens vs. FLOPs
* X-axis: FLOPs (10^17 to 10^23, logarithmic scale)
* Y-axis: Tokens (1B to 1T, logarithmic scale)
* **Chart 4 (Bottom-Left):** GitHub Training Loss vs. Parameters
* X-axis: Parameters (100M to 6B, logarithmic scale)
* Y-axis: GitHub Training Loss (0.2 to 1.0, linear scale)
* **Chart 5 (Bottom-Center):** Parameters vs. FLOPs
* X-axis: FLOPs (10^17 to 10^23, logarithmic scale)
* Y-axis: Parameters (100M to 1T, logarithmic scale)
* **Chart 6 (Bottom-Right):** Tokens vs. FLOPs
* X-axis: FLOPs (10^17 to 10^23, logarithmic scale)
* Y-axis: Tokens (1B to 1T, logarithmic scale)
**Legend (shared across all charts):**
* Black: 1e19
* Blue: 1e20
* Light Blue: 6e20
* Green: 1e21
### Detailed Analysis or Content Details
**Chart 1: C4 Training Loss vs. Parameters**
* The black line (1e19) starts at approximately 3.0 loss at 100M parameters, decreases sharply to around 2.4 loss at 300M parameters, and then plateaus around 2.3-2.4 loss for larger parameter sizes.
* The blue line (1e20) starts at approximately 2.8 loss at 100M parameters, decreases more rapidly than the black line, reaching around 2.2 loss at 300M parameters, and then plateaus around 2.1-2.2 loss.
* The light blue line (6e20) shows a similar trend, starting at around 2.6 loss, decreasing to approximately 2.0 loss, and then plateauing.
* The green line (1e21) starts at approximately 2.4 loss, decreases to around 1.9 loss, and then plateaus.
**Chart 2: Parameters vs. FLOPs**
* All lines (black, blue, light blue, green) are approximately linear and follow the same trajectory. The lines show a direct proportional relationship between parameters and FLOPs.
* At 10^17 FLOPs, the parameter count is approximately 100M for all models.
* At 10^21 FLOPs, the parameter count is approximately 10B for all models.
* At 10^23 FLOPs, the parameter count is approximately 100B for all models.
* The slope of the lines is roughly consistent, indicating a constant FLOPs-to-parameter ratio.
**Chart 3: Tokens vs. FLOPs**
* Similar to the Parameters vs. FLOPs chart, all lines are approximately linear and follow the same trajectory.
* At 10^17 FLOPs, the token count is approximately 1B for all models.
* At 10^21 FLOPs, the token count is approximately 10B for all models.
* At 10^23 FLOPs, the token count is approximately 100B for all models.
**Chart 4: GitHub Training Loss vs. Parameters**
* The black line (1e19) starts at approximately 0.8 loss at 100M parameters, decreases to around 0.6 loss at 300M parameters, and then plateaus around 0.5-0.6 loss.
* The blue line (1e20) starts at approximately 0.75 loss, decreases more rapidly, reaching around 0.5 loss at 300M parameters, and then plateaus around 0.4-0.5 loss.
* The light blue line (6e20) shows a similar trend, starting at around 0.7 loss, decreasing to approximately 0.4 loss, and then plateauing.
* The green line (1e21) starts at approximately 0.65 loss, decreases to around 0.35 loss, and then plateaus.
**Chart 5: Parameters vs. FLOPs**
* Identical to Chart 2.
**Chart 6: Tokens vs. FLOPs**
* Identical to Chart 3.
### Key Observations
* **Loss Decreases with Parameters:** For both C4 and GitHub datasets, training loss generally decreases as the number of parameters increases, but the improvement diminishes beyond a certain point (plateau).
* **Linear Scaling of FLOPs:** The relationship between parameters and FLOPs, and between tokens and FLOPs, appears to be approximately linear. This suggests that increasing model size or dataset size requires a proportional increase in computational resources.
* **Dataset Dependence:** The absolute values of the training loss are different for the C4 and GitHub datasets, indicating that the difficulty of the training task varies between the two datasets. The GitHub dataset consistently shows lower loss values for a given parameter count.
* **Consistent Scaling Laws:** The scaling laws (relationships between parameters, FLOPs, and tokens) appear to be consistent across different model sizes.
### Interpretation
These charts demonstrate the scaling laws governing the performance of neural language models. They suggest that increasing model size (parameters) and dataset size (tokens) leads to improved performance (lower training loss), but with diminishing returns. The linear relationship between FLOPs and parameters/tokens highlights the computational cost associated with scaling up these models. The differences in loss between the C4 and GitHub datasets suggest that the choice of training data significantly impacts model performance. The consistent scaling laws across different model sizes indicate that these relationships are fundamental and can be used to predict the performance of even larger models. The plateauing of loss curves suggests that simply increasing model size indefinitely may not be the most effective strategy for improving performance; other factors, such as model architecture and training techniques, may become more important at larger scales. The charts provide empirical evidence supporting the idea that computational resources are a key limiting factor in the development of large language models.
</details>
| Approach | Coef. π where π πππ‘ / πΆ π | Coef. π where π· πππ‘ / πΆ π |
|----------------------|-----------------------------|-----------------------------|
| C4 | 0.5 | 0.5 |
| GitHub | 0.53 | 0.47 |
| Kaplan et al. (2020) | 0.73 | 0.27 |
Table A2 j Estimated parameter and data scaling with increased training compute on two alternate datasets. The listed values are the exponents, π and π , on the relationship ππππ‘ / πΆ π and π·πππ‘ / πΆ π . Using IsoFLOP profiles, we estimate the scaling on two different datasets.
## D. Details on the scaling analyses
## D.1. Approach 1: Fixing model sizes and varying training sequences
We use a maximum learning rate of 2 10 4 for the smallest models and 1 25 10 4 for the largest models. In all cases, the learning rate drops by a factor of 10 during training, using a cosine schedule. We make the assumption that the cosine cycle length should be approximately matched to the number of training steps. We find that when the cosine cycle overshoots the number of training steps by more than 25%, performance is noticeably degraded-see Figure A1. 10 We use Gaussian smoothing with a window length of 10 steps to smooth the training curve.
## D.2. Approach 3: Parametric fitting of the loss
In this section, we first show how Equation (2) can be derived. We repeat the equation below for clarity ,
<!-- formula-not-decoded -->
based on a decomposition of the expected risk between a function approximation term and an optimisation suboptimality term. We then give details on the optimisation procedure for fitting the parameters.
Loss decomposition. Formally, we consider the task of predicting the next token π¦ 2 Y based on the previous tokens in a sequence π₯ 2 Y π , with π varying from 0 to π max-the maximum sequence length. We consider a distribution π 2 D'X Y' of tokens in Y and their past in X . A predictor π : X ! D'Y' computes the probability of each token given the past sequence. The Bayes classifier, π β
, minimizes the cross-entropy of π ' π₯ ' with the observed tokens π¦ , with expectation taken on the whole data distribution. We let πΏ be the expected risk
<!-- formula-not-decoded -->
The set of all transformers of size π , that we denote H π , forms a subset of all functions that map sequences to distributions of tokens X ! D'Y' . Fitting a transformer of size π on the expected risk πΏ ' π ' amounts to minimizing such risk on a restricted functional space
<!-- formula-not-decoded -->
When we observe a dataset ' π₯ π π¦ π ' π π 2Β» 1 π· β¦ of size π· , we do not have access to πΌ π , but instead to the empirical expectation Λ πΌ π· over the empirical distribution Λ ππ· . What happens when we are given π·
10 This further emphasises the point of not only determining model size, but also training length before training begins.
datapoints that we can only see once, and when we constrain the size of the hypothesis space to be π -dimensional ? We are making steps toward minimizing the empirical risk within a finite-dimensional functional space H π :
<!-- formula-not-decoded -->
We are never able to obtain Λ π π π· as we typically perform a single epoch over the dataset of size π· . Instead, be obtain Β― π π π· , which is the result of applying a certain number of gradient steps based on the π· datapoints-the number of steps to perform depends on the gradient batch size, for which we use well-tested heuristics.
Using the Bayes-classifier π β
, the expected-risk minimizer π π and the 'single-epoch empirical-risk minimizer' Β― π π π· , we can finally decompose the loss πΏ ' π π· ' into
<!-- formula-not-decoded -->
The loss comprises three terms: the Bayes risk, i.e. the minimal loss achievable for next-token prediction on the full distribution π , a.k.a the 'entropy of natural text.'; a functional approximation term that depends on the size of the hypothesis space; finally, a stochastic approximation term that captures the suboptimality of minimizing Λ πΏ π· instead of πΏ , and of making a single epoch on the provided dataset.
Expected forms of the loss terms. In the decomposition (9), the second term depends entirely on the number of parameters π that defines the size of the functional approximation space. On the set of two-layer neural networks , it is expected to be proportional to 1 π 1 2 (Siegel and Xu, 2020). Finally, given that it corresponds to early stopping in stochastic first order methods, the third term should scale as the convergence rate of these methods, which is lower-bounded by 1 π· 1 2 (Robbins and Monro, 1951) (and may attain the bound). This convergence rate is expected to be dimension free (see e.g. Bubeck, 2015, for a review) and depends only on the loss smoothness; hence we assume that the second term only depends on π· in (2). Empirically, we find after fitting (2) that
<!-- formula-not-decoded -->
with πΈ = 1 69, π΄ = 406 4, π΅ = 410 7. We note that the parameter/data coefficients are both lower than 1 2 ; this is expected for the data-efficiency coefficient (but far from the known lower-bound). Future models and training approaches should endeavor to increase these coefficients.
Fitting the decomposition to data. We effectively minimize the following problem
<!-- formula-not-decoded -->
where πΏππΈ is the log-sum-exp operator. We then set π΄ π΅ πΈ = exp ' π ' exp ' π ' exp ' π ' .
We use the LBFGS algorithm to find local minima of the objective above, started on a grid of initialisation given by: πΌ 2 f 0 0 5 2 g , π½ 2 f 0 0 5 2 g , π 2 f 1 5 1 g , π 2 f 0 5 25 g , and π 2 f 0 5 25 g . We find that the optimal initialisation is not on the boundary of our initialisation sweep.
We use πΏ = 10 3 for the Huber loss. We find that using larger values of πΏ pushes the model to overfit the small compute regime and poorly predict held-out data from larger runs. We find that using a πΏ smaller than 10 3 does not impact the resulting predictions.
## D.3. Predicted compute optimal frontier for all three methods
For Approaches 2 and 3, we show the estimated model size and number of training tokens for a variety of compute budgets in Table A3. We plot the predicted number of tokens and parameters for a variety of FLOP budgets for the three methods in Figure A3.
Table A3 j Estimated optimal training FLOPs and training tokens for various model sizes. Analogous to Table 3, we show the model size/token count projections from Approaches 2 and 3 for various compute budgets.
| | Approach 2 | Approach 2 | Approach 3 | Approach 3 |
|-------------|--------------|----------------|--------------|-----------------|
| Parameters | FLOPs | Tokens | FLOPs | Tokens |
| 400 Million | 1.84e+19 | 7.7 Billion | 2.21e+19 | 9.2 Billion |
| 1 Billion | 1.20e+20 | 20.0 Billion | 1.62e+20 | 27.1 Billion |
| 10 Billion | 1.32e+22 | 219.5 Billion | 2.46e+22 | 410.1 Billion |
| 67 Billion | 6.88e+23 | 1.7 Trillion | 1.71e+24 | 4.1 Trillion |
| 175 Billion | 4.54e+24 | 4.3 Trillion | 1.26e+24 | 12.0 Trillion |
| 280 Billion | 1.18e+25 | 7.1 Trillion | 3.52e+25 | 20.1 Trillion |
| 520 Billion | 4.19e+25 | 13.4 Trillion | 1.36e+26 | 43.5 Trillion |
| 1 Trillion | 1.59e+26 | 26.5 Trillion | 5.65e+26 | 94.1 Trillion |
| 10 Trillion | 1.75e+28 | 292.0 Trillion | 8.55e+28 | 1425.5 Trillion |
.
Figure A3 j Optimal number of tokens and parameters for a training FLOP budget. For a fixed FLOP budget, we show the optimal number of tokens and parameters as predicted by Approaches 1, 2, and 3. For an alternate representation, see Figure 1.
<details>
<summary>Image 11 Details</summary>

### Visual Description
## Scatter Plot: Model Parameters vs. Tokens Trained
### Overview
This image presents a scatter plot visualizing the relationship between the number of tokens trained and the number of parameters in various large language models. The plot uses a logarithmic scale for both axes. Several models are represented as distinct data points, and three approaches are shown as lines representing scaling trends.
### Components/Axes
* **X-axis:** Tokens (logarithmic scale, from 10^10 to 10^13)
* **Y-axis:** Parameters (logarithmic scale, from 10^8 to 10^12)
* **Legend:** Located in the top-left corner, containing the following entries:
* Approach 1 (Blue line)
* Approach 2 (Orange line)
* Approach 3 (Purple line)
* Chinchilla (Teal star)
* Gopher (Yellow star)
* GPT-3 (Red star)
* Megatron-Turing NLG (Purple star)
* **Data Points:** Represent individual models.
* **Grid Lines:** Present to aid in reading values.
* **Value Labels:** Numerical labels are placed along the lines representing the approaches, indicating parameter counts (e.g., 1e+18, 1e+19, 1e+21, 1e+22, 1e+23, 1e+24, 1e+25, 1e+26).
### Detailed Analysis
**Approach 1 (Blue Line):** This line exhibits a strong upward trend, indicating that as the number of tokens increases, the number of parameters also increases significantly.
* Approximately (10^10, 1e+18)
* Approximately (10^11, 1e+19)
* Approximately (10^12, 1e+23)
* Approximately (10^13, 1e+26)
**Approach 2 (Orange Line):** This line also shows an upward trend, but it is less steep than Approach 1.
* Approximately (10^10, 1e+18)
* Approximately (10^11, 1e+20)
* Approximately (10^12, 1e+22)
* Approximately (10^13, 1e+25)
**Approach 3 (Purple Line):** This line is the flattest of the three approaches, indicating a slower increase in parameters relative to tokens.
* Approximately (10^10, 1e+18)
* Approximately (10^11, 1e+21)
* Approximately (10^12, 1e+23)
* Approximately (10^13, 1e+24)
**Individual Models:**
* **Chinchilla (Teal Star):** Approximately (10^11.5, 1e+24)
* **Gopher (Yellow Star):** Approximately (10^11.5, 1e+24)
* **GPT-3 (Red Star):** Approximately (10^11, 1e+22)
* **Megatron-Turing NLG (Purple Star):** Approximately (10^11, 1e+21)
### Key Observations
* The three approaches demonstrate different scaling strategies for model size. Approach 1 scales parameters most aggressively with tokens, while Approach 3 scales them more conservatively.
* Chinchilla and Gopher have similar parameter counts and token counts.
* GPT-3 and Megatron-Turing NLG have lower parameter counts compared to Chinchilla and Gopher, given their token counts.
* The lines representing the approaches generally serve as upper bounds for the individual models, suggesting that these approaches define the feasible scaling limits.
### Interpretation
The plot illustrates the trade-offs between model size (parameters) and training data (tokens) in large language models. The different approaches suggest varying philosophies regarding scaling. Approach 1 prioritizes maximizing model capacity, while Approach 3 focuses on efficiency. The positioning of individual models relative to these approaches reveals their design choices. For example, Chinchilla and Gopher, which are known for their parameter-efficient training, fall close to the scaling trends. The data suggests that there is a strong correlation between the number of tokens trained and the number of parameters, but the optimal scaling strategy depends on specific goals and constraints. The plot highlights the importance of considering both model size and training data when developing large language models. The logarithmic scales emphasize the exponential growth in both parameters and tokens, underscoring the computational challenges associated with scaling these models.
</details>
## D.4. Small-scale comparison to Kaplan et al. (2020)
For 10 21 FLOPs, we perform a head-to-head comparison of a model predicted by Approach 1 and that predicted by Kaplan et al. (2020). For both models, we use a batch size of 0.5M tokens and a
maximum learning rate of 1 5 10 4 that decays by 10 . From Kaplan et al. (2020), we find that the optimal model size should be 4.68 billion parameters. From our approach 1, we estimate a 2.86 billion parameter model should be optimal. We train a 4.74 billion parameter and a 2.80 billion parameter transformer to test this hypothesis, using the same depth-to-width ratio to avoid as many confounding factors as possible. We find that our predicted model outperforms the model predicted by Kaplan et al. (2020) as shown in Figure A4.
Figure A4 j Comparison to Kaplan et al. (2020) at 10 21 FLOPs. We train 2.80 and 4.74 billion parameter transformers predicted as optimal for 10 21 FLOPs by Approach 1 and by Kaplan et al. (2020). We find that our prediction results in a more performant model at the end of training.
<details>
<summary>Image 12 Details</summary>

### Visual Description
\n
## Chart: Training Loss vs. Sequences/FLOPs
### Overview
The image presents two line charts comparing the training loss of a model ("Approach 1") against a baseline ("Kaplan et al (2020)"). The left chart plots training loss against the number of sequences processed, while the right chart plots training loss against the number of FLOPs (Floating Point Operations) performed. Both charts share the same y-axis scale (Training Loss) but differ in their x-axis representation.
### Components/Axes
* **Y-axis (Both Charts):** "Training Loss", ranging from approximately 2.2 to 2.8.
* **Left Chart X-axis:** "Sequences", ranging from 0 to approximately 2.1 x 10<sup>7</sup>.
* **Right Chart X-axis:** "FLOPs x10<sup>21</sup>", ranging from 0.0 to 1.0.
* **Legend (Top-Right):**
* Yellow Line: "Kaplan et al (2020)"
* Blue Line: "Approach 1"
* **Horizontal Dashed Line (Both Charts):** A horizontal dashed line at approximately 2.32 on the Training Loss axis.
### Detailed Analysis or Content Details
**Left Chart (Loss vs. Sequences):**
* **Kaplan et al (2020) - Yellow Line:** The line starts at approximately 2.75 at 0 sequences, rapidly decreases to around 2.35 at approximately 5 x 10<sup>6</sup> sequences, and then plateaus around 2.32.
* **Approach 1 - Blue Line:** The line begins at approximately 2.75 at 0 sequences, decreases more gradually than the yellow line, reaching approximately 2.32 at approximately 1.5 x 10<sup>7</sup> sequences, and then plateaus.
**Right Chart (Loss vs. FLOPs):**
* **Kaplan et al (2020) - Yellow Line:** The line starts at approximately 2.75 at 0 FLOPs, rapidly decreases to around 2.35 at approximately 0.4 x 10<sup>21</sup> FLOPs, and then plateaus around 2.32.
* **Approach 1 - Blue Line:** The line begins at approximately 2.75 at 0 FLOPs, decreases more gradually than the yellow line, reaching approximately 2.32 at approximately 0.8 x 10<sup>21</sup> FLOPs, and then plateaus.
### Key Observations
* Both approaches demonstrate a decreasing training loss as the number of sequences or FLOPs increases.
* "Kaplan et al (2020)" achieves a lower loss for a given amount of computational effort (sequences or FLOPs) initially, but both approaches converge to approximately the same loss value.
* The horizontal dashed line represents a potential lower bound or target loss value.
* The rate of loss reduction slows down for both approaches as they approach the plateau.
### Interpretation
The charts suggest that both "Approach 1" and the method described by "Kaplan et al (2020)" are effective in reducing training loss. However, "Kaplan et al (2020)" appears to be more efficient in the early stages of training, achieving a lower loss with fewer sequences or FLOPs. As training progresses, "Approach 1" catches up, eventually reaching a similar loss level. This could indicate that "Approach 1" requires more computational resources to achieve the same level of performance as "Kaplan et al (2020)", but it is still capable of reaching a comparable result. The plateauing of both curves suggests diminishing returns β further increasing sequences or FLOPs beyond a certain point yields minimal improvement in training loss. The horizontal dashed line could represent a practical limit to the achievable loss, or a point where further optimization becomes less impactful. The comparison highlights a trade-off between computational efficiency and overall performance.
</details>
## E. Curvature of the FLOP-loss frontier
We observe that as models increase there is a curvature in the FLOP-minimal loss frontier. This means that projections from very small models lead to different predictions than those from larger models. In Figure A5 we show linear fits using the first, middle, and final third of frontier-points. In this work, we do not take this in to account and we leave this as interesting future work as it suggests that even smaller models may be optimal for large FLOP budgets.
## F. FLOPs computation
We include all training FLOPs, including those contributed to by the embedding matrices, in our analysis. Note that we also count embeddings matrices in the total parameter count. For large models the FLOP and parameter contribution of embedding matrices is small. We use a factor of 2 to describe the multiply accumulate cost. For the forward pass, we consider contributions from:
- Embeddings
- Attention (Single Layer)
- -2 seq\_len vocab\_size d\_model
- -Key, query and value projections : 2 3 seq\_len d\_model ' key\_size num\_heads '
Figure A5 j Training curve envelopes. We fit to the first third (orange), the middle third (green), and the last third (blue) of all points along the loss frontier. We plot only a subset of the points.
<details>
<summary>Image 13 Details</summary>

### Visual Description
\n
## Chart: Training Loss vs. FLOPS with Parameter Count Color Coding
### Overview
The image presents a chart illustrating the relationship between Training Loss and FLOPS (Floating Point Operations) for a series of machine learning model training runs. The color of each line represents the number of parameters in the model, indicated by a colorbar on the right side of the chart. Several lines are overlaid, showing the training loss decreasing as FLOPS increase. A few lines are highlighted with dashed lines.
### Components/Axes
* **X-axis:** FLOPS, labeled with a logarithmic scale ranging from 10<sup>17</sup> to 10<sup>22</sup>.
* **Y-axis:** Training Loss, labeled with a linear scale ranging from 2.0 to 6.0.
* **Colorbar:** Represents the number of parameters in millions, ranging from 75 to 10000. The color gradient transitions from dark purple (75 million parameters) to bright yellow (10000 million parameters).
* **Lines:** Multiple colored lines representing individual training runs. The color of each line corresponds to the number of parameters in the model.
* **Dashed Lines:** Three dashed lines are overlaid on the chart, each with a distinct color and pattern.
### Detailed Analysis
The chart displays numerous curves, each representing a training run. The lines generally exhibit a downward trend, indicating that training loss decreases as FLOPS increase. The steepness of the decline varies between lines.
* **Lines with Parameter Count:**
* Lines colored dark purple (approximately 75 million parameters) start at a training loss of around 5.5 and decrease relatively slowly.
* Lines colored blue (approximately 250 million parameters) start at a training loss of around 5.0 and decrease at a moderate rate.
* Lines colored green (approximately 500 million parameters) start at a training loss of around 4.5 and decrease at a moderate rate.
* Lines colored orange/red (approximately 2500-5000 million parameters) start at a training loss of around 4.0 and decrease rapidly.
* Lines colored yellow (approximately 10000 million parameters) start at a training loss of around 5.5 and decrease rapidly.
* **Dashed Lines:**
* **Blue Dashed Line:** Starts at approximately FLOPS = 10<sup>17</sup> and Training Loss = 3.8, and decreases linearly to approximately FLOPS = 10<sup>22</sup> and Training Loss = 2.1.
* **Green Dashed Line:** Starts at approximately FLOPS = 10<sup>17</sup> and Training Loss = 3.6, and decreases linearly to approximately FLOPS = 10<sup>22</sup> and Training Loss = 2.0.
* **Red Dashed Line:** Starts at approximately FLOPS = 10<sup>17</sup> and Training Loss = 3.4, and decreases linearly to approximately FLOPS = 10<sup>22</sup> and Training Loss = 1.9.
* **Black Dots:** A series of black dots are plotted along the lower portion of the curves, concentrated between approximately FLOPS = 10<sup>20</sup> and FLOPS = 10<sup>22</sup>, with Training Loss values between 2.0 and 2.5.
### Key Observations
* **Parameter Count and Loss:** Generally, models with more parameters (yellow lines) achieve lower training loss for a given FLOPS value, but also start with a higher initial loss.
* **Scaling:** The logarithmic scale on the x-axis (FLOPS) is crucial for visualizing the data, as the range of FLOPS values is very large.
* **Linear Approximation:** The dashed lines provide a linear approximation of the training loss reduction, potentially representing a theoretical or expected performance.
* **Convergence:** The black dots suggest a region of convergence where further increases in FLOPS yield diminishing returns in terms of training loss reduction.
### Interpretation
The chart demonstrates the trade-off between model size (number of parameters), computational cost (FLOPS), and training loss. Larger models (more parameters) generally require more FLOPS to achieve a given level of training loss, but can potentially reach lower loss values overall. The dashed lines suggest a baseline performance expectation, and the deviation of the individual training runs from these lines could indicate variations in training data, optimization algorithms, or model architecture. The black dots highlight a point of diminishing returns, where increasing FLOPS beyond a certain point does not significantly reduce training loss. This information is valuable for understanding the scaling behavior of machine learning models and optimizing training strategies. The chart suggests that there is a sweet spot in terms of model size and computational resources to achieve optimal performance.
</details>
- -Key @ Query logits : 2 seq\_len seq\_len ' key\_size num\_heads '
- Softmax @ query reductions : 2 seq\_len seq\_len ' key\_size num\_heads '
- Softmax : 3 num\_heads seq\_len seq\_len
- -Final Linear : 2 seq\_len ' key\_size num\_heads ' d\_model
- -2 seq\_len ' d\_model ffw\_size , d\_model ffw\_size '
- Dense Block (Single Layer)
- Final Logits
- Total forward pass FLOPs: embeddings , num\_layers ' total\_attention , dense\_block ' +logits
- -2 seq\_len d\_model vocab\_size
As in Kaplan et al. (2020) we assume that the backward pass has twice the FLOPs of the forward pass. We show a comparison between our calculation and that using the common approximation πΆ = 6 π·π (Kaplan et al., 2020) where πΆ is FLOPs, π· is the number of training tokens, and π is the number of parameters in Table A4. We find the differences in FLOP calculation to be very small and they do not impact our analysis. Compared to the results presented in Rae et al. (2021), we use a slightly more
Table A4 j FLOP comparison. For a variety of different model sizes, we show the ratio of the FLOPs that we compute per sequence to that using the 6 ππ· approximation.
| Parameters | num_layers | d_model | ffw_size | num_heads | k/q size | FLOP Ratio (Ours/6 ππ· ) |
|--------------|--------------|-----------|------------|-------------|------------|---------------------------|
| 73M | 10 | 640 | 2560 | 10 | 64 | 1.03 |
| 305M | 20 | 1024 | 4096 | 16 | 64 | 1.1 |
| 552M | 24 | 1280 | 5120 | 10 | 128 | 1.08 |
| 1.1B | 26 | 1792 | 7168 | 14 | 128 | 1.04 |
| 1.6B | 28 | 2048 | 8192 | 16 | 128 | 1.03 |
| 6.8B | 40 | 3584 | 14336 | 28 | 128 | 0.99 |
accurate calculation giving a slightly different value (6 3 10 23 compared to 5 76 10 23 ).
## G. Other differences between Chinchilla and Gopher
Beyond differences in model size and number of training tokens, there are some additional minor differences between Chinchilla and Gopher . Specifically , Gopher was trained with Adam (Kingma and Ba, 2014) whereas Chinchilla was trained with AdamW (Loshchilov and Hutter, 2019). Furthermore, as discussed in Lessons Learned in Rae et al. (2021), Chinchilla stored a higher-precision copy of the weights in the sharded optimiser state.
We show comparisons of models trained with Adam and AdamW in Figure A6 and Figure A7. We find that, independent of the learning rate schedule, AdamW trained models outperform models trained with Adam. In Figure A6 we show a comparison of an 680 million parameter model trained
Figure A6 j Comparison of other differences. Using an 680 million parameter model, we show a comparison between the setup used to train Gopher and Chinchilla -the change in optimiser and using a higher precision copy of the weights in the optimiser state. The setup used for Chinchilla (orange) clearly outperforms the setup used to train Gopher (green).
<details>
<summary>Image 14 Details</summary>

### Visual Description
\n
## Line Charts: Training Loss, Perplexity, and C4 Loss vs. Million Sequences
### Overview
The image presents three line charts arranged horizontally. Each chart depicts the relationship between "Million Sequences" on the x-axis and a different metric on the y-axis: "Training Loss", "Wikitext103 Perplexity", and "C4 Loss". Each chart contains four data series, differentiated by color, representing different "Training Setup" configurations.
### Components/Axes
* **X-axis (all charts):** "Million Sequences", ranging from 0 to 30.
* **Chart 1 Y-axis:** "Training Loss"
* **Chart 2 Y-axis:** "Wikitext103 Perplexity"
* **Chart 3 Y-axis:** "C4 Loss"
### Data Series (Common to all charts)
The following training setups are represented by different colored lines in each chart:
* **Baseline:** (Color not specified in the description)
* **LoRA:** (Color not specified in the description)
* **QLoRA:** (Color not specified in the description)
* **QLoRA + DDP:** (Color not specified in the description)
### Chart-Specific Observations
* **Training Loss Chart:** The chart shows how the training loss decreases as the number of million sequences increases for each training setup.
* **Wikitext103 Perplexity Chart:** This chart illustrates the perplexity on the Wikitext103 dataset as a function of the number of million sequences. Lower perplexity indicates better performance.
* **C4 Loss Chart:** This chart displays the loss on the C4 dataset in relation to the number of million sequences.
</details>
Figure A7 j Adam vs AdamW. For a 417M (blue) and 1.4B model (green), we find that training with AdamW improves performance over training with Adam.
<details>
<summary>Image 15 Details</summary>

### Visual Description
## Charts: Model Training Performance
### Overview
The image presents three charts displaying the training performance of different language models. The charts show the relationship between model training progress (measured in Million Sequences) and various performance metrics: C4 Loss, WikiText103 Perplexity, and LAMBADA Accuracy. Four different model configurations are compared: 417M with Adam optimizer, 417M with AdamW optimizer, 1.4B with Adam optimizer, and 1.4B with AdamW optimizer.
### Components/Axes
Each chart shares a common x-axis:
* **X-axis:** Million Sequences (ranging from 0 to 150)
The y-axes vary per chart:
* **Left Chart:** C4 Loss (ranging from 2.3 to 2.8)
* **Center Chart:** WikiText103 Perplexity (ranging from 10.0 to 30.0)
* **Right Chart:** LAMBADA Accuracy (ranging from 0.0 to 0.7)
The legend, located in the bottom-right corner, identifies the four model configurations using both size (417M, 1.4B) and optimizer (Adam, AdamW) and corresponding line colors:
* **Blue Solid Line:** 417M, Adam
* **Blue Dashed Line:** 417M, AdamW
* **Green Solid Line:** 1.4B, Adam
* **Green Dashed Line:** 1.4B, AdamW
### Detailed Analysis
**Left Chart: C4 Loss vs. Million Sequences**
* **417M, Adam (Blue Solid):** The line starts at approximately 2.78 at 0 Million Sequences and decreases to approximately 2.55 at 150 Million Sequences. The decrease is most rapid in the first 50 Million Sequences, then plateaus.
* **417M, AdamW (Blue Dashed):** The line starts at approximately 2.75 at 0 Million Sequences and decreases to approximately 2.52 at 150 Million Sequences. The decrease is similar to the Adam variant, but consistently lower.
* **1.4B, Adam (Green Solid):** The line starts at approximately 2.70 at 0 Million Sequences and decreases to approximately 2.40 at 150 Million Sequences. The decrease is more pronounced than the 417M models.
* **1.4B, AdamW (Green Dashed):** The line starts at approximately 2.65 at 0 Million Sequences and decreases to approximately 2.35 at 150 Million Sequences. This line consistently shows the lowest loss values.
**Center Chart: WikiText103 Perplexity vs. Million Sequences**
* **417M, Adam (Blue Solid):** The line starts at approximately 29.0 at 0 Million Sequences and decreases to approximately 22.0 at 150 Million Sequences.
* **417M, AdamW (Blue Dashed):** The line starts at approximately 28.5 at 0 Million Sequences and decreases to approximately 21.5 at 150 Million Sequences.
* **1.4B, Adam (Green Solid):** The line starts at approximately 27.0 at 0 Million Sequences and decreases to approximately 16.0 at 150 Million Sequences.
* **1.4B, AdamW (Green Dashed):** The line starts at approximately 26.0 at 0 Million Sequences and decreases to approximately 15.0 at 150 Million Sequences.
**Right Chart: LAMBADA Accuracy vs. Million Sequences**
* **417M, Adam (Blue Solid):** The line starts at approximately 0.35 at 0 Million Sequences and increases to approximately 0.52 at 150 Million Sequences.
* **417M, AdamW (Blue Dashed):** The line starts at approximately 0.38 at 0 Million Sequences and increases to approximately 0.55 at 150 Million Sequences.
* **1.4B, Adam (Green Solid):** The line starts at approximately 0.45 at 0 Million Sequences and increases to approximately 0.65 at 150 Million Sequences.
* **1.4B, AdamW (Green Dashed):** The line starts at approximately 0.48 at 0 Million Sequences and increases to approximately 0.68 at 150 Million Sequences.
### Key Observations
* **Model Size:** The 1.4B models consistently outperform the 417M models across all metrics.
* **Optimizer:** AdamW generally leads to better performance than Adam, as indicated by lower loss and perplexity, and higher accuracy.
* **Convergence:** All models show diminishing returns in performance as training progresses beyond 100 Million Sequences.
* **C4 Loss & Perplexity Correlation:** The C4 Loss and WikiText103 Perplexity charts exhibit an inverse relationship; as loss decreases, perplexity also decreases.
* **LAMBADA Accuracy & Training:** LAMBADA Accuracy increases with training, indicating the model's ability to learn long-range dependencies.
### Interpretation
These charts demonstrate the impact of model size and optimizer choice on language model training. The larger 1.4B models achieve significantly better performance than the 417M models, suggesting that increasing model capacity is beneficial. The use of the AdamW optimizer consistently improves performance over Adam, likely due to its weight decay regularization, which helps prevent overfitting.
The diminishing returns observed after 100 Million Sequences suggest that further training may not yield substantial improvements. The inverse relationship between C4 Loss and WikiText103 Perplexity indicates that reducing loss on the C4 dataset leads to better generalization performance on the WikiText103 dataset. The increasing LAMBADA Accuracy demonstrates the model's ability to capture long-range dependencies, which is crucial for tasks requiring contextual understanding.
The data suggests that a 1.4B model trained with the AdamW optimizer is the most effective configuration among those tested. The charts provide valuable insights into the training dynamics of language models and can inform decisions about model architecture and training strategies.
</details>
with and without the higher precision copy of the weights and with Adam/AdamW for comparison.
## H. Results
## H.1. The Pile
In Table A5 we show the bits-per-byte (bpb) on The Pile (Gao et al., 2020) of Chinchilla , Gopher , and Jurassic-1. Chinchilla outperforms Gopher on all subsets. Jurassic-1 outperforms Chinchilla on 2 subsets-dm\_mathematics and ubuntu\_irc .
| Subset | Chinchilla (70B) | Gopher (280B) | Jurassic-1 (170B) |
|-------------------|--------------------|-----------------|---------------------|
| pile_cc | 0.667 | 0.691 | 0.669 |
| pubmed_abstracts | 0.559 | 0.578 | 0.587 |
| stackexchange | 0.614 | 0.641 | 0.655 |
| github | 0.337 | 0.377 | 0.358 |
| openwebtext2 | 0.647 | 0.677 | - |
| arxiv | 0.627 | 0.662 | 0.680 |
| uspto_backgrounds | 0.526 | 0.546 | 0.537 |
| freelaw | 0.476 | 0.513 | 0.514 |
| pubmed_central | 0.504 | 0.525 | 0.579 |
| dm_mathematics | 1.111 | 1.142 | 1.037 |
| hackernews | 0.859 | 0.89 | 0.869 |
| nih_exporter | 0.572 | 0.59 | 0.590 |
| opensubtitles | 0.871 | 0.9 | 0.879 |
| europarl | 0.833 | 0.938 | - |
| books3 | 0.675 | 0.712 | 0.835 |
| philpapers | 0.656 | 0.695 | 0.742 |
| gutenberg_pg_19 | 0.548 | 0.656 | 0.890 |
| bookcorpus2 | 0.714 | 0.741 | - |
| ubuntu_irc | 1.026 | 1.09 | 0.857 |
Table A5 j Bits-per-Byte on The Pile. We show the bpb on The Pile for Chinchilla compared to Gopher and Jurassic-1.
## H.2. MMLU
In Table A6 we show the performance of Chinchilla and Gopher on each subset of MMLU.
## H.3. Winogender Setup
We follow the same setup as in Rae et al. (2021). To test coreference resolution in Chinchilla , we input a sentence which includes a pronoun reference (e.g., 'The librarian helped the child pick out a book because {pronoun} liked to encourage reading.'), then measure the probability of the model completing the sentence ''{Pronoun}' refers to the' with different sentence roles ('librarian' and 'child' in this example). Each example is annotated with the correct pronoun resolution (the pronoun corresponds to the librarian in this example). Each sentence is tested with a female, male, and gender-neutral pronoun. An unbiased model would correctly predict which word the pronoun refers to regardless of pronoun gender.
## H.4. BIG-bench
In Table A7 we show Chinchilla and Gopher performance on each subset of BIG-bench that we consider.
## I. Model Card
We present the Chinchilla model card in Table A8, following the framework presented by Mitchell et al. (2019).
| Task | Chinchilla | Gopher | Task | Chinchilla | Gopher |
|------------------------------|--------------|----------|------------------------------|--------------|----------|
| abstract_algebra | 31 | 25 | anatomy | 70.4 | 56.3 |
| astronomy | 73 | 65.8 | business_ethics | 72.0 | 70.0 |
| clinical_knowledge | 75.1 | 67.2 | college_biology | 79.9 | 70.8 |
| college_chemistry | 51 | 45 | college_computer_science | 51.0 | 49.0 |
| college_mathematics | 32 | 37 | college_medicine | 66.5 | 60.1 |
| college_physics | 46.1 | 34.3 | computer_security | 76.0 | 65.0 |
| conceptual_physics | 67.2 | 49.4 | econometrics | 38.6 | 43.0 |
| electrical_engineering | 62.1 | 60 | elementary_mathematics | 41.5 | 33.6 |
| formal_logic | 33.3 | 35.7 | global_facts | 39.0 | 38.0 |
| high_school_biology | 80.3 | 71.3 | high_school_chemistry | 58.1 | 47.8 |
| high_school_computer_science | 58 | 54 | high_school_european_history | 78.8 | 72.1 |
| high_school_geography | 86.4 | 76.8 | high_school_gov_and_politics | 91.2 | 83.9 |
| high_school_macroeconomics | 70.5 | 65.1 | high_school_mathematics | 31.9 | 23.7 |
| high_school_microeconomics | 77.7 | 66.4 | high_school_physics | 36.4 | 33.8 |
| high_school_psychology | 86.6 | 81.8 | high_school_statistics | 58.8 | 50.0 |
| high_school_us_history | 83.3 | 78.9 | high_school_world_history | 85.2 | 75.1 |
| human_aging | 77.6 | 66.4 | human_sexuality | 86.3 | 67.2 |
| international_law | 90.9 | 77.7 | jurisprudence | 79.6 | 71.3 |
| logical_fallacies | 80.4 | 72.4 | machine_learning | 41.1 | 41.1 |
| management | 82.5 | 77.7 | marketing | 89.7 | 83.3 |
| medical_genetics | 69 | 69 | miscellaneous | 84.5 | 75.7 |
| moral_disputes | 77.5 | 66.8 | moral_scenarios | 36.5 | 40.2 |
| nutrition | 77.1 | 69.9 | philosophy | 79.4 | 68.8 |
| prehistory | 81.2 | 67.6 | professional_accounting | 52.1 | 44.3 |
| professional_law | 56.5 | 44.5 | professional_medicine | 75.4 | 64.0 |
| professional_psychology | 75.7 | 68.1 | public_relations | 73.6 | 71.8 |
| security_studies | 75.9 | 64.9 | sociology | 91.0 | 84.1 |
| us_foreign_policy | 92 | 81 | virology | 53.6 | 47.0 |
| world_religions | 87.7 | 84.2 | | | |
Table A6 j Chinchilla MMLU results. For each subset of MMLU (Hendrycks et al., 2020), we show Chinchilla 's accuracy compared to Gopher .
| Model Details | Model Details |
|-----------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Organization Developing the Model | DeepMind |
| Model Date | March 2022 |
| Model Type | Autoregressive Transformer Language Model (Section 4.1 for details) |
| Feedback on the Model | {jordanhoffmann, sborgeaud, amensch,sifre}@deepmind.com |
| Intended Uses | Intended Uses |
| Primary Intended Uses | The primary use is research on language models, including: research on the scaling behaviour of language models along with those listed in Rae et al. (2021). |
## Primary Intended Users
DeepMind researchers. We will not make this model available
| | publicly. |
|---------------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Out-of-Scope Uses | Uses of the language model for language generation in harm- ful or deceitful settings. More generally, the model should not be used for downstream applications without further safety and fairness mitigations. |
| Factors | Factors |
| Card Prompts - Relevant Factor | Relevant factors include which language is used. Our model is trained on English data. Furthermore, in the analysis of mod- els trained on the same corpus in Rae et al. (2021), we found it has unequal performance when modelling some dialects (e.g., African American English). Our model is designed for research. The model should not be used for downstream ap- plications without further analysis on factors in the proposed downstream application. |
| Card Prompts - Evaluation Factors | See the results in Rae et al. (2021) which analyzes models trained on the same text corpus. |
| Metrics | Metrics |
| Model Performance Measures | β’ Perplexity and bits per byte on language modelling datasets β’ Accuracy on completion tasks, reading comprehension, MMLU, BIG-bench and fact checking. β’ Exact match accuracy for question answering. β’ Generation toxicity from Real Toxicity Prompts (RTP) alongside toxicity classification accuracy. β’ Gender and occupation bias. Test include comparing the probability of generating different gender terms and the Winogender coreference resolution task. We principally focus on Chinchilla 's performance compared to Gopher on text likelihood prediction. |
| Decision thresholds | N/A |
| Approaches to Uncertainty and Vari- ability | Due to the costs of training large language models, we did not train Chinchilla multiple times. However, the breadth of our evaluation on a range of different task types gives a reasonable estimate of the overall performance of the model. Furthermore, the existence of another large model trained on the same dataset ( Gopher ) provides a clear point of com- parison. |
| Evaluation Data | Evaluation Data |
## Datasets
| | β’ Language modelling on LAMBADA, Wikitext103 (Mer- ity et al., 2017), C4 (Raffel et al., 2020a), PG-19 (Rae et al., 2020) and the Pile (Gao et al., 2020). β’ Language understanding, real world knowledge, mathematical and logical reasoning on the Massive Multitask Language Understanding (MMLU) bench- mark (Hendrycks et al., 2020) and on the 'Beyond the Imitation Game Benchmark' (BIG-bench) (BIG-bench collaboration, 2021). β’ Question answering (closed book) on Natural Ques- tions (Kwiatkowski et al., 2019) and TriviaQA (Joshi et al., 2017). β’ Reading comprehension on RACE (Lai et al., 2017) β’ Common sense understanding on HellaSwag (Zellers et al., 2019), PIQA (Bisk et al., 2020), Wino- grande (Sakaguchi et al., 2020), SIQA (Sap et al., 2019), BoolQ (Clark et al., 2019), and TruthfulQA (Lin et al., 2021). |
|---------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Motivation | We chose evaluations from Rae et al. (2021) to allow us to most directly compare to Gopher . |
| Preprocessing | Input text is tokenized using a SentencePiece tokenizer with a vocabulary of size 32,000. Unlike the tokenizer used for Gopher , the tokenizer used for Chinchilla does not perform NFKC normalization. |
## Training Data
The same dataset is used as in Rae et al. (2021). Differences in sampling are shown in Table A1.
## Quantitative Analyses
| Unitary Results | Section 4.2 gives a detailed description of our analysis. Main take-aways include: β’ Our model is capable of outputting toxic language as measured by the PerspectiveAPI. This is particularly true when the model is prompted with toxic prompts. β’ Gender: Our model emulates stereotypes found in our dataset, with occupations such as 'dietician' and 're- ceptionist' being more associated with women and 'car- penter' and 'sheriff' being more associated with men. β’ Race/religion/country sentiment: Prompting our model to discuss some groups leads to sentences with lower or higher sentiment, likely reflecting text in our dataset. |
|-------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
Intersectional Results
We did not investigate intersectional biases.
| | Ethical Considerations |
|-----------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Data | The data is the same as described in Rae et al. (2021). |
| Human Life | The model is not intended to inform decisions about matters central to human life or flourishing. |
| Mitigations | We considered filtering the dataset to remove toxic content but decided against it due to the observation that this can introduce new biases as studied by Welbl et al. (2021). More work is needed on mitigation approaches to toxic content and other types of risks associated with language models, such as those discussed in Weidinger et al. (2021). |
| Risks and Harms | The data is collected from the internet, and thus undoubtedly there is toxic/biased content in our training dataset. Fur- thermore, it is likely that personal information is also in the dataset that has been used to train our models. We defer to the more detailed discussion in Weidinger et al. (2021). |
| Use Cases | Especially fraught use cases include the generation of fac- tually incorrect information with the intent of distributing it or using the model to generate racist, sexist or otherwise toxic text with harmful intent. Many more use cases that could cause harm exist. Such applications to malicious use are discussed in detail in Weidinger et al. (2021). |
Table A8 j Chinchilla model card. We follow the framework presented in Mitchell et al. (2019).
## J. List of trained models
In Table A9 we list the model size and configuration of all models used in this study. Many models have been trained multiple times, for a different number of training steps.
Table A7 j Chinchilla BIG-bench results. For each subset of BIG-bench (BIG-bench collaboration, 2021), we show Chinchilla and Gopher 's accuracy.
| Task | Chinchilla | Gopher | Task | Chinchilla | Gopher |
|---------------------------------|--------------|----------|------------------------------|--------------|----------|
| hyperbaton | 54.2 | 51.7 | movie_dialog_same_or_diff | 54.5 | 50.7 |
| causal_judgment | 57.4 | 50.8 | winowhy | 62.5 | 56.7 |
| formal_fallacies_syllogisms_neg | 52.1 | 50.7 | movie_recommendation | 75.6 | 50.5 |
| crash_blossom | 47.6 | 63.6 | moral_permissibility | 57.3 | 55.1 |
| discourse_marker_prediction | 13.1 | 11.7 | strategyqa | 68.3 | 61 |
| general_knowledge_json | 94.3 | 93.9 | nonsense_words_grammar | 78 | 61.4 |
| sports_understanding | 71 | 54.9 | metaphor_boolean | 93.1 | 59.3 |
| implicit_relations | 49.4 | 36.4 | navigate | 52.6 | 51.1 |
| penguins_in_a_table | 48.7 | 40.6 | presuppositions_as_nli | 49.9 | 34 |
| intent_recognition | 92.8 | 88.7 | temporal_sequences | 32 | 19 |
| reasoning_about_colored_objects | 59.7 | 49.2 | question_selection | 52.6 | 41.4 |
| logic_grid_puzzle | 44 | 35.1 | logical_fallacy_detection | 72.1 | 58.9 |
| timedial | 68.8 | 50.9 | physical_intuition | 79 | 59.7 |
| epistemic_reasoning | 60.6 | 56.4 | physics_mc | 65.5 | 50.9 |
| ruin_names | 47.1 | 38.6 | identify_odd_metaphor | 68.8 | 38.6 |
| hindu_knowledge | 91.4 | 80 | understanding_fables | 60.3 | 39.6 |
| misconceptions | 65.3 | 61.7 | logical_sequence | 64.1 | 36.4 |
| implicatures | 75 | 62 | mathematical_induction | 47.3 | 57.6 |
| disambiguation_q | 54.7 | 45.5 | fantasy_reasoning | 69 | 64.1 |
| known_unknowns | 65.2 | 63.6 | SNARKS | 58.6 | 48.3 |
| dark_humor_detection | 66.2 | 83.1 | crass_ai | 75 | 56.8 |
| analogical_similarity | 38.1 | 17.2 | entailed_polarity | 94 | 89.5 |
| sentence_ambiguity | 71.7 | 69.1 | irony_identification | 73 | 69.7 |
| riddle_sense | 85.7 | 68.2 | evaluating_info_essentiality | 17.6 | 16.7 |
| date_understanding | 52.3 | 44.1 | phrase_relatedness | 94 | 81.8 |
| analytic_entailment | 67.1 | 53 | novel_concepts | 65.6 | 59.1 |
| odd_one_out | 70.9 | 32.5 | empirical_judgments | 67.7 | 52.5 |
| logical_args | 56.2 | 59.1 | figure_of_speech_detection | 63.3 | 52.7 |
| alignment_questionnaire | 91.3 | 79.2 | english_proverbs | 82.4 | 57.6 |
| similarities_abstraction | 87 | 81.8 | Human_organs_senses_mcc | 85.7 | 84.8 |
| anachronisms | 69.1 | 56.4 | gre_reading_comprehension | 53.1 | 27.3 |
Table A9 j All models. We list the hyperparameters and size of all models trained as part of this work. Many shown models have been trained with multiple learning rate schedules/number of training tokens.
| Parameters (million) | d_model | ffw_size | kv_size | n_heads | n_layers |
|------------------------|-----------|------------|-----------|-----------|------------|
| 44 | 512 | 2048 | 64 | 8 | 8 |
| 57 | 576 | 2304 | 64 | 9 | 9 |
| 74 | 640 | 2560 | 64 | 10 | 10 |
| 90 | 640 | 2560 | 64 | 10 | 13 |
| 106 | 640 | 2560 | 64 | 10 | 16 |
| 117 | 768 | 3072 | 64 | 12 | 12 |
| 140 | 768 | 3072 | 64 | 12 | 15 |
| 163 | 768 | 3072 | 64 | 12 | 18 |
| 175 | 896 | 3584 | 64 | 14 | 14 |
| 196 | 896 | 3584 | 64 | 14 | 16 |
| 217 | 896 | 3584 | 64 | 14 | 18 |
| 251 | 1024 | 4096 | 64 | 16 | 16 |
| 278 | 1024 | 4096 | 64 | 16 | 18 |
| 306 | 1024 | 4096 | 64 | 16 | 20 |
| 425 | 1280 | 5120 | 128 | 10 | 18 |
| 489 | 1280 | 5120 | 128 | 10 | 21 |
| 509 | 1408 | 5632 | 128 | 11 | 18 |
| 552 | 1280 | 5120 | 128 | 10 | 24 |
| 587 | 1408 | 5632 | 128 | 11 | 21 |
| 632 | 1536 | 6144 | 128 | 12 | 19 |
| 664 | 1408 | 5632 | 128 | 11 | 24 |
| 724 | 1536 | 6144 | 128 | 12 | 22 |
| 816 | 1536 | 6144 | 128 | 12 | 25 |
| 893 | 1792 | 7168 | 128 | 14 | 20 |
| 1,018 | 1792 | 7168 | 128 | 14 | 23 |
| 1,143 | 1792 | 7168 | 128 | 14 | 26 |
| 1,266 | 2048 | 8192 | 128 | 16 | 22 |
| 1,424 | 2176 | 8704 | 128 | 17 | 22 |
| 1,429 | 2048 | 8192 | 128 | 16 | 25 |
| 1,593 | 2048 | 8192 | 128 | 16 | 28 |
| 1,609 | 2176 | 8704 | 128 | 17 | 25 |
| 1,731 | 2304 | 9216 | 128 | 18 | 24 |
| 1,794 | 2176 | 8704 | 128 | 17 | 28 |
| 2,007 | 2304 | 9216 | 128 | 18 | 28 |
| 2,283 | 2304 | 9216 | 128 | 18 | 32 |
| 2,298 | 2560 | 10240 | 128 | 20 | 26 |
| 2,639 | 2560 | 10240 | 128 | 20 | 30 |
| 2,980 | 2560 | 10240 | 128 | 20 | 34 |
| 3,530 | 2688 | 10752 | 128 | 22 | 36 |
| 3,802 | 2816 | 11264 | 128 | 22 | 36 |
| 4,084 | 2944 | 11776 | 128 | 22 | 36 |
| 4,516 | 3072 | 12288 | 128 | 24 | 36 |
| 6,796 | 3584 | 14336 | 128 | 28 | 40 |
| 9,293 | 4096 | 16384 | 128 | 32 | 42 |
| 11,452 | 4352 | 17408 | 128 | 32 | 47 |
| 12,295 | 4608 | 18432 | 128 | 36 | 44 |
| 12,569 | 4608 | 18432 | 128 | 32 | 47 |
| 13,735 | 4864 | 19456 | 128 | 32 | 47 |
| 14,940 | 4992 | 19968 | 128 | 32 | 49 |
| 16,183 | 5120 | 20480 | 128 | 40 | 47 |