Reading time: 38 minute
...

๐Ÿ“ Original Info

  • Title:
  • ArXiv ID: 2512.12167
  • Date:
  • Authors: Unknown

๐Ÿ“ Abstract

So far, expensive finetuning beyond the pretraining sequence length has been a requirement for effectively extending the context of language models (LM). In this work, we break this key bottleneck by Dropping the Positional Embeddings of LMs after training (DroPE). Our simple method is motivated by three key theoretical and empirical observations. First, positional embeddings (PEs) serve a crucial role during pretraining, providing an important inductive bias that significantly facilitates convergence. Second, over-reliance on this explicit positional information is also precisely what prevents test-time generalization to sequences of unseen length, even when using popular PE-scaling methods. Third, positional embeddings are not an inherent requirement of effective language modeling and can be safely removed after pretraining following a short recalibration phase. Empirically, DroPE yields seamless zero-shot context extension without any long-context finetuning, quickly adapting pretrained LMs without compromising their capabilities in the original training context. Our findings hold across different models and dataset sizes, far outperforming previous specialized architectures and established rotary positional embedding scaling methods.

๐Ÿ“„ Full Content

Transformers established themselves as the predominant architecture for training foundation models at unprecedented scale in language and beyond (Brown et al., 2020;Dosovitskiy et al., 2020;Jumper et al., 2021;Team et al., 2023). The defining feature of transformers is abandoning explicit architectural biases such as convolutions and recurrences in favor of highly general self-attention layers (Vaswani et al., 2017), while injecting positional information about the sequence through positional embeddings (PEs) and causal masking. However, despite significant efforts to scale attention to long sequences on modern hardware (Dao et al., 2022;Liu and Abbeel, 2023;Liu et al., 2023a), this powerful layer is inherently bottlenecked by quadratic token-totoken operations, which makes pretraining at long sequence lengths computationally intractable at scale. As a result, enabling models to use contexts beyond their pretraining length without additional long-context fine-tuning (i.e., "zero-shot context extension") has emerged as a central challenge for the next generation of foundation models (Chi et al., 2023;Press et al., 2021).

When inference sequence lengths exceed the pretraining context, the performance of modern transformer-based LMs degrades sharply. This is directly caused by their use of explicit PEs such as the ubiquitous rotary positional embeddings (RoPE) (Su et al., 2024), which become out-of-distribution at unseen sequence lengths. To address this issue, careful scaling techniques that adapt RoPE frequencies on longer sequences were introduced (bloc97, 2023;Chen et al., 2023;Ding et al., 2024;Peng et al., 2023). However, despite their popularity, these methods still rely on an expensive, long-context finetuning phase to meaningfully use tokens beyond the original sequence length, failing to generalize out of the box (Lu et al., 2024a). Beyond RoPE transformers, alternative architectures and positional embedding schemes have shown early promise in reducing costs by attenuating the underlying quadratic computational burden Choromanski et al. (2020); Wang et al. (2020); Xiong et al. (2021); Zaheer et al. (2020) or maintaining better out-of-context generalization (Kazemnejad et al., 2023;Puvvada et al., 2025;Yang et al., 2025b). Yet, these parallel efforts are still far from challenging established pipelines, introducing notable performance and stability trade-offs that prevent wide adoption.

In this work, we challenge the conventional role of RoPE in language modeling, and propose to overcome this inherent trade-off by Dropping the Positional Embeddings (DroPE) of LMs after pretraining. Our method is based on three key theoretical and empirical observations. First, explicit positional embeddings significantly facilitate pretraining convergence by baking in an important inductive bias that is difficult to recover from data alone. Second, over-reliance on positional embeddings is precisely what prevents test-time generalization to sequences of unseen length, with RoPE-scaling context extension methods focusing on recent tokens instead of ones deeper in the context to retain perplexity. Third, explicit PE is not an inherent requirement for effective language modeling and can be removed after pretraining, following a short recalibration phase which is performed at the original context length.

Empirically, DroPE models generalize zero-shot to sequences far beyond their training context, marking a sharp contrast to traditional positional scaling techniques. Moreover, we show that adapting RoPE models with DroPE does not compromise their original in-context capabilities, preserving both perplexity and downstream task performance. Our findings hold across LMs of different architectures and sizes up to 7B parameters pretrained on trillions of tokens, establishing a new standard for developing robust and scalable long-context transformers.

Contributions. In summary, our main contributions are as follows:

(1) In Section 3, we provide empirical and theoretical analysis of the role of positional embeddings in LM training, showing their importance in significantly accelerating convergence. (2) In Section 4, we discuss why RoPE-scaling methods fail to reliably attend across far-away tokens when evaluated zero-shot on long sequences, showing that these approaches inevitably shift attention weights, hindering the model’s test-time behavior.

(3) In Section 5, we introduce DroPE, a new method that challenges the conventional role of positional embeddings in transformers, motivated by our empirical and theoretical analyses of its role as a transient but critical training inductive bias. (4) We demonstrate that DroPE enables zero-shot generalization of pretrained RoPE transformers far beyond their original sequence length, without any long-context finetuning. DroPE can be incorporated at no extra cost into established training pipelines, and can be used to inexpensively empower arbitrary pretrained LLMs in the wild.

We share our code to facilitate future work and extensions toward developing foundation models capable of handling orders-of-magnitude longer contexts.

Self-attention.

where ๐‘‘ ๐‘˜ is the head dimension. A multi-head attention block computes multiple attention outputs

, concatenates them, and projects to the model dimension:

].

Language and positional embeddings. State-of-the-art autoregressive transformer LMs use information about sequence positions provided both implicitly via causal masking of the attention scores1 , and explicitly with positional embeddings. In particular, the modern literature has settled on the Rotary PE (RoPE) scheme (Su et al., 2024), providing relative positional information to each attention head by rotating ๐‘ž ๐‘– and ๐‘˜ ๐‘— in 2D chunks before the inner product in Equation 1:

Here, each ๐‘…(๐œ” ๐‘š ) โˆˆ โ„ 2ร—2 is a planar rotation of angle ๐œ” ๐‘š = ๐‘ -2(๐‘š-1)/๐‘‘ ๐‘˜ acting on the (2๐‘š, 2๐‘š + 1) subspace of ๐‘ž ๐‘– and ๐‘˜ ๐‘— . The base ๐‘ is commonly taken to be 10,000.

where ๐œ… ๐‘š โˆˆ [0, 1] interpolates between 0 and 1 as the base frequency ๐œ” ๐‘š grows (see Appendix A). These methods, referred to as RoPE-scaling, still require additional finetuning on long sequences, and don’t generalize to long-context downstream tasks out of the box (Lu et al., 2024b).

NoPE transformers. In a parallel line of work, there have been efforts to train transformers without PEs, commonly referred to as NoPE architectures (Haviv et al., 2022;Kazemnejad et al., 2023), to avoid the need for rescaling RoPE frequencies. While NoPE was shown to be a viable LM architecture, it has failed to gain traction due to degraded performance (Haviv et al., 2022;Yang et al., 2025b) compared to RoPE architectures. For an in-depth introduction to the above concepts, see Appendix A. While NoPE transformers were shown to be expressive enough for effective sequence modeling (Haviv et al., 2022;Kazemnejad et al., 2023), we find that they consistently underperform RoPE architectures throughout our experiments. As illustrated in Figure 3, NoPE transformers maintain visibly worse perplexity throughout training. These empirical results are consistent with past literature (Haviv et al., 2022;Yang et al., 2025b), yet the reasons why positional embeddings are key for effective language model training have never been fully understood.

From a purely mechanistic perspective, even without explicit positional embeddings, NoPE transformers can exploit the causal mask to encode positional information, maintaining the same expressivity as their RoPE counterparts (Haviv et al., 2022;Kazemnejad et al., 2023). Specifically, Kazemnejad et al. (2023) prove that the first attention layer in a NoPE transformer can perfectly reconstruct sequence positions, and subsequent layers can emulate the effects of relative or absolute positional embeddings.

As detailed in Section 3.1, rather than looking at theoretical expressivity, we investigate this empirical performance discrepancy from an optimization perspective, providing theoretical analysis of the positional bias of NoPE transformers during training. The theoretical and empirical analysis in this section can be summarized in the following observation. At a high level, our analysis focuses on the rate at which NoPE and RoPE transformers can develop positional bias in their self-attention heads, which captures their non-uniformity. We quantify attention positional bias as a linear functional on the attention map:

Definition 3.1 (Attention positional bias). Given centered positional weights ๐‘ ๐‘– ๐‘— โˆˆ โ„ with ๐‘—โ‰ค๐‘– ๐‘ ๐‘– ๐‘— = 0, the positional bias of the attention weights ๐›ผ ๐‘– ๐‘— is

Attention heads with a strong positional bias would maximize the average value of A ๐‘ across input sequences for some weights ๐‘. For example, a “diagonal” attention head, focusing mass on the current token, is exactly the maximizer of A ๐‘ , with ๐‘ ๐‘– ๐‘— having 1s on the diagonal and -1 ๐‘–-1 otherwise. Figure 4 | RoPE transformers have higher positional bias gradients at initialization. We compare the average norm of A ๐‘ across layers, for RoPE and NoPE transformers. In 4a we plot the gradient norms of positional bias towards a diagonal head, and in 4b, we take bias towards previous token attention, off-diagonal head. In both cases, the gradient norm is consistently higher for RoPE across layers, meaning that RoPE heads can learn these patterns faster.

To validate the theory behind Observation 1, we empirically compare the gradients of the attention positional bias functional in attention heads of RoPE and NoPE transformers. Specifically, we measure the average gradient norm at initialization in the direction of two common language modeling patterns: diagonal attention heads, placing mass on the current token, and off-diagonal heads, capturing immediate previous token context. As illustrated in Figure 4, the gradient magnitudes of NoPE transformers are far lower than those of RoPE transformers, with the gap between the two growing in deeper layers. This means that diagonal and off-diagonal heads are slower to develop under NoPE, reflecting its difficulty in recovering positional information. In the next section, we theoretically analyze the causes of this gradient norm gap.

We detail our findings, summarized in Observation 1, with a series of formal results, bounding the rate at which positional bias can develop early in training. We provide full proofs and an extended analysis of these results in Appendix B. 2) query and key gradients vanish: ๐œ•L/๐œ•๐‘Š ๐‘„ = ๐œ•L/๐œ•๐‘Š ๐พ = 0, (3) for all heads and any positional weights A ๐‘ = 0, โˆ‡ ๐œƒ A ๐‘ = 0 , and (4) the output is constant:

The explicit positional information injected into attention heads in RoPE transformers circumvents this issue. Enabling non-zero A ๐‘ gradients even on constant sequences. Proposition 3.3. For a non-trivial RoPE attention head, even if the input sequence is constant, there are positional weights ๐‘, for which A ๐‘ > 0, and โˆฅโˆ‡ ๐œƒ A ๐‘ โˆฅ > 0.

NoPE transformers propagate embedding uniformity. At initialization, the entries of the embedding matrix are drawn i.i.d. from a distribution with a fixed small variance (commonly, ๐œŽ 2 = 0.02). Therefore, the token embeddings are close to uniform at the beginning of training. The next theorem shows that for NoPE transformers, this uniformity persists throughout the network, and bounds the attention positional bias A ๐‘ and its gradients.

Theorem 3.4. Define the he prefix-spread of the hidden states at layer ๐‘™ as

, where h(๐‘™)

๐‘— .

For NoPE transformers, there exists ๐œ€ > 0 and constants ๐ถ 1 , ๐ถ 2 , and ๐ถ 3 such that if the initial embeddings ฮ” (1) โ„Ž โ‰ค ๐œ€, then for all layers ๐‘™ โ‰ค ๐ฟ:

with high probability over the initialization distribution. The constants only depend on the number of layers and heads, and not on the sequence length.

The main idea in the proof of Theorem 3.4 is that uniformity in the embeddings causes uniformity in the attention maps, so ๐›ผ ๐‘– ๐‘— โ‰ˆ 1/๐‘–. Uniform mixing of tokens cannot increase the prefix spread; thus, uniformity persists throughout the network. This result explains the discrepancy between RoPE and NoPE transformers illustrated in Figure 4.

In summary, we demonstrate that while NoPE attention can learn positional bias, attention nonuniformity develops slowly early in training due to bounded A ๐‘ gradients at initialization. State-of-the-art RoPE scaling methods fail to effectively generalize to sequences longer than those seen in training without additional long-context finetuning. While YaRN and other popular frequency scaling techniques do avoid perplexity degradation on longcontext sequence (bloc97, 2023;Peng et al., 2023), they exhibit sharp performance drops on downstream tasks whenever important information is present deep in the sequence, beyond the training context (Liu et al., 2023b;Lu et al., 2024b). We empirically demonstrate this phenomenon, comparing the perplexity and needle-in-a-haystack (NIAH) (Hsieh et al., 2024;Kamradt, 2023) performance of a RoPE transformer scaled with YaRN and to a cropped context baseline. As illustrated in Figure 5, YaRN’s zeroshot behavior closely matches that of simply cropping the sequence length to the pretraining context, maintaining constant perplexity but ignoring information present outside the cropped window.

The cause of this limitation lies in the way context extension methods scale different RoPE frequencies.

As detailed in Section 2, elaborated on in Appendix A, and illustrated in Figure 6, the scaling factors of PI (Chen et al., 2023), RoPE-NTK (bloc97, 2023), and YaRN (Peng et al., 2023) have a strong effect on low frequencies. In Section 4.1, we discuss why this aggressive scaling of low frequencies leads to the observed failures, yielding our second observation.

Observation 2. RoPE-scaling methods must compress low frequencies to keep positional phases indistribution. This, in turn, shifts semantic attention heads at large relative distances, causing the observed failures on downstream tasks, preventing zero-shot context extension. Effect of RoPE scaling. RoPE scaling methods modify the frequencies at inference time to evaluate sequences that are longer than those seen during pretraining. In each (2๐‘š, 2๐‘š+1) subspace, the RoPE phase at relative distance ฮ” is ๐œ™ ๐‘š (ฮ”) = ๐œ” ๐‘š ฮ”, so scaling the frequency to ๐œ” โ€ฒ ๐‘š = ๐›พ ๐‘š ๐œ” ๐‘š is equivalent to using a phase

As illustrated in Figure 6, most scaling methods leave high frequencies nearly unchanged (๐›พ ๐‘š โ‰ˆ 1) but all of them compress the low frequencies (๐›พ ๐‘š โ‰ˆ 1/๐‘ ). As demonstrated both theoretically and empirically in Barbero et al. (2024), high RoPE frequencies are primarily used by positional heads, with attention patterns based on relative token positions (e.g., diagonal or previous-token heads). In contrast, low frequencies are predominantly used by semantic heads that attend based on query/key content. Consequently, positional heads are largely unaffected by scaling, but semantic attention is shifted. Moreover, the effect on low-frequency dominated semantic heads is exacerbated for distant tokens, since the relative phase ๐œ™ ๐‘š (ฮ”) is larger, and thus the 1/๐‘  scaling factor has a greater effect. In other words, scaling Figure 8 | RoPE scaling shifts semantic attention mass. Attention weights of the last token (query) with tokens from a retrieval target (keys) in a semantic head evaluated on a NIAH probe. Since the head uses low frequencies and the relative distance is non-trivial, the impact of YaRN is substantial, shifting attention mass between tokens.

warps low-frequency phases, shifting long-range attention in precisely the subspaces most used for semantic matching.

In Figure 7 and Figure 8, we illustrate this behavior in practice. We start by selecting a positional attention head in a pretrained Qwen2.5-0.5B model by examining its average attention positional bias (Definition 3.1) across layers. In Figure 7, we show the average attention weights in this positional head under YaRN scaling with ๐‘  = 2. Because high frequencies, which are least affected by YaRN, dominate positional heads, the average attention profiles are similar. In Figure 8, we then contrast this behavior with that of a semantic head for a long needle-in-a-haystack sequence, plotting the average attention of the last token (query) with tokens around the needle (keys). YaRN’s aggressive scaling of low frequencies substantially shifts attention mass across tokens, reflecting the impact of frequency compression at longer ranges.

Why this is inevitable. In a standard RoPE setup, low-frequency phases never make a full cycle over the original context length: ๐œ™ ๐‘š (๐ถ train ) = ๐œ” ๐‘š ๐ถ train < 2๐œ‹ for small ๐œ” ๐‘š . E.g. for a standard RoPE base ๐‘ = 10 4 , a transformer with head dimension ๐‘‘ ๐‘˜ = 64, will have at least five low frequencies for which ๐œ™ ๐‘š (๐ถ train ) < 2๐œ‹, even at a training context of ๐ถ train = 32,000. If we leave ๐œ” ๐‘š unchanged at an extended length ๐ถ test > ๐ถ train , the new maximal relative phase ๐œ™ ๐‘š (๐ถ test ) is pushed outside the training regime and becomes out of distribution for the head. Therefore, to constrain phases to remain in range, any scaling method must choose ๐›พ ๐‘š โ‰ค ๐ถ train ๐ถ test = 1 ๐‘  , which becomes increasingly small as the extension factor ๐‘  grows. In other words, when applying a RoPE transformer to sequences longer than those seen in training, any post-hoc scaling method must compress the low frequencies. But this compression, in turn, shifts attention weights at long relative distances.

Taken together, Observations 1 and 2 imply that providing explicit positional information with PE is a key component for effective LM training, but is also a fundamental barrier to long-context generalization. This raises a natural question: is it possible to harness the inductive bias from positional embeddings exclusively during training? We answer in the affirmative. In this section, we demonstrate that it is possible to drop all positional embeddings from a pretrained transformer and quickly recover the model’s in-context capabilities with a brief recalibration phase. Most notably, this simple new procedure (termed DroPE) unlocks strong zero-shot long context generalization to unseen sequence lengths, far beyond highly-tuned RoPE extensions and prior alternative architectures.

Observation 3. Positional embeddings can be removed after pretraining, allowing LMs to generalize zero-shot to unseen sequence lengths without compromising their in-context performance after short recalibration on a fraction of the training tokens at the original context size.

We extensively validate DroPE across different LM and dataset scales, showing it outperforms prior approaches both as a zero cost integration into pretraining recipes and as an inexpensive way to adapt any LM in the wild already pretrained on trillions of tokens. For all experiments in this paper, we provide full implementation details of each evaluated architecture and optimization phase, including comprehensive hyperparameter lists in Appendix C.

Integrating DroPE at no extra cost. For our first set of experiments, we train from scratch different LMs with half a billion parameters on 16B fineweb tokens (Penedo et al., 2024), over twice the chinchilla-optimal rate (Hoffmann et al., 2022). We repeat this recipe for RoPE and NoPE transformers, as well as an ALiBi model (Press et al., 2021) and an RNoPE-SWA model Yang et al. (2025b), which are alternative architectures specifically aimed at long-context capabilities. We implement DroPE by taking the 14B tokens RoPE transformer checkpoint, removing positional embeddings from every layer, and resuming training for the final 2B tokens. Despite only recalibrating at the very end of training, at no extra cost, DroPE matches the final in-context validation perplexity of RoPE trained on the full 16B tokens, showing a clear edge over the NoPE baseline trained without positional embedding all the way (Figure 2). We provide further analysis and ablations on the recalibration starting point in Appendix D.1, validating the importance of harnessing the inductive bias of RoPE for a substantial amount of training, in line with the core motivation of our new method. To evaluate the long-context generalization of each method, we select three tasks from the RULER benchmark (Hsieh et al., 2024): (1) multi-query: retrieve needles for several listed keys, (2) multi-key: retrieve the needle for one specified key, and

(3) multi-value: retrieve all needles for one key with a single query. For the base RoPE transformer, we consider three context extension strategies: PI (Chen et al., 2023), NTK-RoPE (bloc97, 2023), and the popular YaRN (Peng et al., 2023) Figure 10 | SmolLM-DroPE recalibration.

We compare three recipes, using 30B, 60B, and 120B training tokens.

We then evaluate our SmolLM-DroPE models’ zeroshot length generalization on four different tasks from LongBench (Bai et al., 2023), a challenging benchmark even for closed-source LMs, including knowledge-extraction problems longer than 80 times SmolLM’s pretraining context (2048 tokens). We compare our method with the base SmolLM and three RoPE extensions: PI, RoPE-NTK, and YaRN. As shown in Table 2, despite a significant difficulty spike compared to our prior evaluations, DroPE still displays a clear edge over prior approaches, improving the base SmolLM’s average score by over 10 times. These gains are far beyond all prior zero-shot RoPE extensions currently used across modern LMs. We refer to Appendix D.2 for a fine-grained analysis of task performance as a function of extension factor.

Scaling to billion-parameter models. Given the remarkable efficiency of recalibration, we test DroPE’s ability to scale to larger LMs in the wild, such as SmolLM-1.7B (Allal et al., 2024) and Llama2-7B (Touvron et al., 2023), which were trained on 1 trillion and 4 trillion tokens, respectively. For both of these models, we perform recalibration on 20B tokens, which only represents 2% of the pretraining budget for SmolLM-1.7B, and only 0.5% for Llama2-7B. As demonstrated in Table 3, consistently with all our prior results on a smaller scale, SmolLM-1.7B-DroPE and Llama2-7B-DroPE once again outperform state-of-the-art RoPE-scaling methods on long-context questionanswering and summarization, providing strong evidence towards the scalability and immediate potential of DroPE. Overall, our in-context and out-of-context results demonstrate DroPE is an efficient and effective long-context extension method, which we believe can have meaningful implications for reducing the cost of training pipelines and for tackling the canonical context scalability challenges of transformers.

We complement this section with additional experimental results, including the entire LongBench benchmark, and a performance by query length breakdown in Appendix D.

Recent improvements to RoPE include variants based on Fourier and wavelet transforms (Hua et al., 2025;Oka et al., 2025) and methods such as ๐‘-RoPE (Barbero et al., 2025), NRoPE-SWA (Yang et al., 2025b), and SWAN-GPT (Puvvada et al., 2025), which occupy a middle ground between RoPE and NoPE. Our approach represents a fundamentally different paradigm, replacing RoPE with NoPE at different stages of training. These directions are complementary to ours and can be used in place of RoPE within the DroPE framework. Another orthogonal direction seeks length generalization while retaining a dedicated positional vector yet modifying its indexing or adaptivity (Wu et al., 2024;Zheng et al., 2024;zican Dong et al., 2024).

Our findings support a reinterpretation of positional embeddings in transformer LMs as a useful inductive bias that is essential for efficient training (Observation 1), but inherently constrains zero-shot context extension (Observation 2). Based on these findings, we propose DroPE, a new method rethinking the conventional role of PEs as a temporary scaffold that can and should be removed after serving their training-time purpose (Observation 3).

where [โ€ข, . . . , โ€ข] represents concatenation along the feature dimension. When clear from context, we omit layer and head indices.

Positional embeddings in transformers. The attention mechanism does not directly encode relative distances between queries and keys. Therefore, attention is invariant to prefix permutations: for any permutation ๐œŽ โˆˆ ๐‘† ๐‘ of the first ๐‘ input tokens, attn(๐‘ฅ ๐œŽ -1 (1) , . . . , ๐‘ฅ ๐œŽ -1 ( ๐‘) , ๐‘ฅ ๐‘ , . . . , ๐‘ฅ ๐‘‡ ) ๐‘– = attn(๐‘ฅ 1 , . . . , ๐‘ฅ ๐‘‡ ) ๐‘– for every ๐‘– > ๐‘. In other words, pure attention is blind to token positions. To address this, Vaswani et al. (2017) introduced absolute positional embeddings, adding position information to the token embeddings before the first transformer block. More recently, many architectures replace absolute embeddings with relative schemes that inject pairwise positional information directly into the attention mechanism. The most widely used approach is Rotary Position Embedding (RoPE) (Su et al., 2024). RoPE modifies the attention scores in Equation 5 by rotating queries and keys before taking their inner product:

where, ๐‘… โˆˆ ๐‘‚(๐‘‘ ๐‘˜ ) is a block-diagonal orthogonal matrix composed out of 2 ร— 2 rotation blocks:

In the standard RoPE parameterization, ๐œ” ๐‘š = ๐‘ -2 ๐‘š-1 ๐‘‘ ๐‘˜ with ๐‘ = 10,000.

Language model context extension. Generalizing to contexts longer than those seen during training is a key challenge for transformer-based language models. The key issue is that when applying a transformer on a longer context, the attention mechanism must operate over more tokens than it was trained to handle. This issue is exacerbated with

NTK-RoPE (bloc97, 2023) uses

so that low frequencies (๐‘š โ‰ˆ ๐‘‘ ๐‘˜ /2) are scaled similarly to PI and for high frequencies ๐›พ ๐‘š โ‰ˆ 1. YaRN (Peng et al., 2023) uses

with tunable ๐‘ and ๐‘ž parameters, originally chosen as ๐‘ = 1, ๐‘ž = 32. See Figure 11 for a comparison between these different RoPE scaling methods with ๐‘  = 2, 3, and 4.

In this section, we analyze the behavior of positional bias, or attention non-uniformity, in NoPE transformers and RoPE transformers early in training. We provide formal statements and proofs for all the results from Section 3, starting with Propositions 3.2 and 3.3, followed by Theorem 3.4. The notation of this section follows that of Appendix A. 2) query and key gradients vanish: ๐œ•L/๐œ•๐‘Š ๐‘„ = ๐œ•L/๐œ•๐‘Š ๐พ = 0, (3) for all heads and any positional weights A ๐‘ = 0, โˆ‡ ๐œƒ A ๐‘ = 0 , and (4) the output is constant:

Proof. Let ๐‘ฅ 1 , . . . , ๐‘ฅ ๐‘‡ be a constant input sequence, ๐‘ฅ 1 = โ€ข โ€ข โ€ข = ๐‘ฅ ๐‘‡ , and let M be a NoPE transformer, i.e. a transformer with no positional encodings and causal self attention. The order of the proof is (4) โ‡’ (1) โ‡’ (2 + 3).

(4) Layer outputs, and thus model outputs, are constant. At the first layer, inputs are identical

= โ„Ž. This means that for every attention head and every 1 โ‰ค ๐‘— โ‰ค ๐‘‡ ๐‘ฃ ๐‘— โ‰ก ๐‘ฃ = ๐‘Š ๐‘‰ โ„Ž.

Therefore, the output of the attention head is

independent of ๐‘–. Concatenating heads and applying ๐‘Š ๐‘‚ preserves equality across positions. Residual connections, LayerNorm, and the MLP are positionwise (the same function is applied independently at each position), so identical inputs produce identical outputs at every position. Thus the layer output remains constant. By repeating this argument layer-by-layer, every subsequent layer receives identical inputs and outputs identical states, so in the end

(1) Uniform causal attention. Using (4), we know that for every layer 1 โ‰ค ๐‘™ โ‰ค ๐ฟ โ„Ž (๐‘™)

Therefore, for every attention head and every 1 โ‰ค ๐‘— โ‰ค ๐‘‡ ๐‘ž ๐‘— โ‰ก ๐‘ž := ๐‘Š ๐‘„ โ„Ž, ๐‘˜ ๐‘— โ‰ก ๐‘˜ := ๐‘Š ๐พ โ„Ž, ๐‘ฃ ๐‘— โ‰ก ๐‘ฃ := ๐‘Š ๐‘‰ โ„Ž.

Thus, for each 1 โ‰ค ๐‘— โ‰ค ๐‘– โ‰ค ๐‘‡, the attention scores ๐‘  ๐‘– ๐‘— = ๐‘ž โŠค ๐‘˜/ โˆš ๐‘‘ ๐‘˜ โ‰ก ๐‘ are constant (independent of ๐‘– or ๐‘—). Hence ๐›ผ ๐‘– ๐‘— = softmax(๐‘, . . . , ๐‘ ๐‘– entries

(2 + 3) Vanishing ๐‘Š ๐‘„ , ๐‘Š ๐พ gradients. Since, the inputs for every layer are constant, we know from (1) that every attention head has ๐›ผ ๐‘– ๐‘— โ‰ก 1/๐‘–, independant of ๐‘Š ๐‘„ and ๐‘Š ๐พ . Therefore ๐œ•๐›ผ ๐‘– ๐‘— /๐œ•๐‘Š ๐‘„ = ๐œ•๐›ผ ๐‘– ๐‘— /๐œ•๐‘Š ๐พ = 0. Since the attention bias A ๐‘ depends on the parameters ๐œƒ only through ๐›ผ ๐‘– ๐‘— and the loss L depends on ๐‘Š ๐‘„ and ๐‘Š ๐พ only through ๐›ผ ๐‘– ๐‘— , all these gradients vanish. More formally, using the chain rule,

Additionally, since the heads are uniform the attention bias is zero to begin with

  1. Note that part (4) of the proposition holds for RoPE transformers as well. Parts (1), ( 2) and ( 3) do not. The relative rotations break attention uniformity and thus changing the magnitude of ๐‘Š ๐‘„ and โˆฅ๐‘Š ๐พ โˆฅ can affect the attention weights. This is formally demonstrated in the next section.

Proposition 3.3. For a non-trivial RoPE attention head, even if the input sequence is constant, there are positional weights ๐‘, for which A ๐‘ > 0, and โˆฅโˆ‡ ๐œƒ A ๐‘ โˆฅ > 0.

Proof. Let ๐‘ฅ 1 = โ€ข โ€ข โ€ข = ๐‘ฅ ๐‘‡ = ๐‘ฅ โˆˆ โ„ ๐‘‘ be the inputs to a RoPE attention head, and let ๐‘Š ๐‘„ , ๐‘Š ๐พ โˆˆ โ„ ๐‘‘ ๐‘˜ ร—๐‘‘ be the query and key projection parameters. Since the projection maps are shared across tokens, the queries and keys are constant as well:

Set the positional bias weights to be

Since ๐‘—โ‰ค๐‘– ๐›ผ ๐‘– ๐‘— = 1, we have ๐‘—โ‰ค๐‘– ๐‘ ๐‘– ๐‘— = 0 as required. The positional bias A ๐‘ is

By Cauchy-Schwarz,

with equality only when

with equality iff ๐›ผ ๐‘– ๐‘— = 1/๐‘– is uniform. Therefore, A ๐‘ > 0 unless ๐›ผ ๐‘– ๐‘— is uniform for all ๐‘–. The following lemma asserts that this is not the case Lemma B.2. For any non-degenerate RoPE head and input embeddings ๐‘ฅ 1 = โ€ข โ€ข โ€ข = ๐‘ฅ ๐‘ก = ๐‘ฅ, there exists ๐‘– โ‰ฅ 1 such that ๐‘  ๐‘–1 , . . . , ๐‘  ๐‘–๐‘– and ๐›ผ ๐‘–1 , . . . , ๐›ผ ๐‘–๐‘– are not uniform.

The proof of Lemma B.2 is at the end of this subsection. As for โˆ‡ ๐œƒ A ๐‘ , rewrite A ๐‘ as

so the dependence in the parameters ๐œƒ is entirely through

From the definition of RoPE, we have

Consider scaling ๐‘ž by a scalar ๐œ† > 0: ๐‘ž โ†ฆ โ†’ ๐œ†๐‘ž. For fixed prefix ๐‘–, define

Then

, where ๐ด ๐‘– (๐œ†) := log ๐‘ ๐‘– (๐œ†) is the log-partition function. The second derivative of the log partition function is the logit variance

2 ๐‘  ๐‘– ๐‘— are not all equal and ๐›ผ ๐‘– ๐‘— (๐œ†) > 0. Thus, ๐ด โ€ฒ ๐‘– (๐œ†) is strictly increasing in ๐œ†. Hence, for any ๐‘– with non-constant logits,

and in particular at ๐œ† = 1,

By the chain rule for ๐‘ž โ†ฆ โ†’ ๐œ†๐‘ž, ๐‘‘ ๐‘‘๐œ† ๐น ๐‘– (๐œ†)

Thus โˆ‡ ๐‘ž ๐น ๐‘– (๐‘ž) โ‰  0 (otherwise the dot product with ๐‘ž couldn’t be strictly positive). Finally, since ๐‘ž = ๐‘Š ๐‘„ ๐‘ฅ,

and with ๐‘ฅ โ‰  0 we get โˆฅโˆ‡ ๐œƒ ๐น ๐‘– โˆฅ โ‰ฅ โˆ‡ ๐‘Š ๐‘„ ๐น ๐‘– > 0. Therefore

has strictly positive norm (a sum of nonzero matrices sharing the same nonzero right factor ๐‘ฅ โŠค cannot be the zero matrix unless all left factors vanish, which they don’t for ๐‘– โ‰ฅ 2). โ–ก

To conclude this section, we now prove Lemma B.2. Lemma B.2. For any non-degenerate RoPE head and input embeddings ๐‘ฅ 1 = โ€ข โ€ข โ€ข = ๐‘ฅ ๐‘ก = ๐‘ฅ, there exists ๐‘– โ‰ฅ 1 such that ๐‘  ๐‘–1 , . . . , ๐‘  ๐‘–๐‘– and ๐›ผ ๐‘–1 , . . . , ๐›ผ ๐‘–๐‘– are not uniform.

Proof. RoPE acts as independent 2 ร— 2 rotations on disjoint coordinate pairs. Thus

with pairwise distinct frequencies ๐œ” ๐‘š โˆˆ (0, 2๐œ‹). Decompose ๐‘ž = (๐‘ž 1 , . . . , ๐‘ž ๐‘€ ), ๐‘˜ = (๐‘˜ 1 , . . . , ๐‘˜ ๐‘€ ), ๐‘Ž ๐‘š , ๐‘ ๐‘š โˆˆ โ„ 2 , so ๐‘  ๐‘– ๐‘— = ๐‘“ ( ๐‘— -๐‘–) where

Let

Define ๐ด ๐‘š := ๐‘ž โŠค ๐‘š ๐‘˜ ๐‘š and ๐ต ๐‘š := ๐‘ž โŠค ๐‘š ๐ฝ๐‘ ๐‘š . Then

where

Assume ๐‘“ (ฮ”) is constant in ฮ” for ฮ” = 0, . . . , 2๐‘€ = ๐‘‘ ๐‘˜ , and denote the constant value by -1 2 ๐ถ 0 . Then we have ๐‘€ โˆ‘๏ธ ๐‘š=-๐‘€ ๐ถ ๐‘š ๐‘’ ๐‘–ฮ”๐œ” ๐‘š โ‰ก 0 were ๐ถ -๐‘š := C๐‘š , and ๐œ” -๐‘š = -๐œ” ๐‘š . Since {๐‘’ -๐‘–๐œ” ๐‘€ , . . . , ๐‘’ -๐‘–๐œ” 1 , 1, ๐‘’ ๐‘–๐œ” 1 , . . . , ๐‘’ ๐‘–๐œ” ๐‘€ } are all distinct, by Vandermonde’s identity this means ๐ถ ๐‘š = C๐‘š = 0 for ๐‘š = 1, . . . , ๐‘€, โ‡’ ๐ด ๐‘š = ๐ต ๐‘š = 0 for ๐‘š = 1, . . . , ๐‘€. Now

If ๐‘˜ ๐‘š โ‰  0, then {๐‘˜ ๐‘š , ๐ฝ๐‘˜ ๐‘š } spans โ„ 2 , forcing ๐‘ž ๐‘š = 0. Thus for every block ๐‘š, either ๐‘ž ๐‘š = 0 or ๐‘˜ ๐‘š = 0, which results in a degenerate RoPE head, contradicting the assumption. Therefore, for ๐‘– โ‰ฅ ๐‘‘ ๐‘˜ + 1 the attention logits ๐‘  ๐‘– ๐‘— are not constant, and thus the attention weight ๐›ผ ๐‘– ๐‘— are not constant. โ–ก

In this section, we prove Theorem 3.4. To do so, we first need to prove a sequence of Propositions and Lemmas. First, we restate the theorem here.

Theorem 3.4. Define the he prefix-spread of the hidden states at layer ๐‘™ as

, where h(๐‘™)

For NoPE transformers, there exists ๐œ€ > 0 and constants ๐ถ 1 , ๐ถ 2 , and ๐ถ 3 such that if the initial embeddings ฮ”

(1) โ„Ž โ‰ค ๐œ€, then for all layers ๐‘™ โ‰ค ๐ฟ:

with high probability over the initialization distribution. The constants only depend on the number of layers and heads, and not on the sequence length.

Since all weight matrices are drawn from a Gaussian distribution with a fixed variance, there exists a constant ๐ต, depending only on the architecture, such that with high probability the operator norms of Proof. Notice that

Therefore, by Cauchy-Swartz

By the linearity of ๐‘Š ๐พ we get ๐‘˜ ๐‘— -k๐‘– = ๐‘Š ๐พ (โ„Ž ๐‘— -h๐‘– ) โ‰ค โˆฅ๐‘Š ๐พ โˆฅ โ„Ž ๐‘— -h๐‘– โ‰ค โˆฅ๐‘Š ๐พ โˆฅ ฮ” (๐‘™) โ„Ž โ‰ค ๐ตฮ” (๐‘™) โ„Ž . As for โˆฅ๐‘ž ๐‘– โˆฅ = ๐‘Š ๐‘„ โ„Ž ๐‘– , recall that โ„Ž ๐‘– are the output of a normalization layer, and therefore (at initialization)

Putting it all together gives

To finish the proof, take a maximum over ๐‘— โ‰ค ๐‘–. โ–ก

To bound the effect on the attention probabilities, we need the following Lemma. Lemma B.4. For any ๐‘ โˆˆ โ„ ๐‘› ,

โˆž for all ๐‘ฅ, ๐‘‘ โˆˆ โ„ ๐‘› (see Theorem 2.1.6 in Nesterov (2013)). Take ๐‘“ (๐‘ฅ) = log ๐‘› ๐‘–=1 ๐‘’ ๐‘ฅ ๐‘– . ๐‘“ is ๐ถ 2 , convex and โˆ‡ ๐‘“ (๐‘ฅ) = softmax(๐‘ฅ). Therefore, all we need to show is that for all ๐‘ฅ, ๐‘‘ โˆˆ โ„ ๐‘› ๐‘‘ โŠค โˆ‡softmax(๐‘ฅ)๐‘‘ = ๐‘‘โˆ‡ 2 ๐‘“ (๐‘ฅ)๐‘‘ โ‰ค โˆฅ๐‘‘โˆฅ 2 . and indeed,

as required. โ–ก

Using Lemma B.4, we can bound the uniformity of ๐›ผ ๐‘– ๐‘— and the prefix spread of the head outputs.

Proposition B.5. Let ๐‘ข ๐‘– = 1 ๐‘– 1 โˆˆ โ„ ๐‘– . In any layer ๐‘™,

and,

Proof. To get Equation 12, let ๐‘Ž be the constant vector (s ๐‘– , . . . , s๐‘– ) โˆˆ โ„ ๐‘– and let ๐‘ = ๐‘  ๐‘– -๐‘Ž. By Lemma B.4

Now, notice that โˆฅ๐‘โˆฅ โˆž = max ๐‘—โ‰ค๐‘– ๐‘  ๐‘– ๐‘— -s๐‘– , therefore Proposition B.3 gives us the desired inequality. For Equation 13 notice that,

.

We now bound the next layer’s spread in terms of the current one. Denote by ฮ” (๐‘™) ๐‘ง := max ๐‘– max ๐‘—โ‰ค๐‘– ๐‘ง ๐‘— -z๐‘– the prefix spread of an attention head’s output. First, we’ll give a bound for ฮ” (๐‘™) ๐‘ง , and then use this bound to prove the entire propagation result. Before, we need a short lemma. Lemma B.6. For any sequence (๐‘ฅ ๐‘— ) and ๐‘— โ‰ค ๐‘–,

To finish the proof, take the maximum over ๐‘– and ๐‘— โ‰ค ๐‘–. โ–ก

There exist constants ๐ด 1 , ๐ด 2 depending only on ๐ต, and ๐ป, such that

Proof. From Proposition B.7, the single-head spread is bounded by a linear term 2๐ตฮ” โ„Ž plus a quadratic term 2๐ต 3 โˆš ๐ป. Concatenation and ๐‘Š ๐‘‚ multiply by at most โˆฅ๐‘Š ๐‘‚ โˆฅ (up to a fixed constant depending on number of heads). Adding the residual preserves a linear contribution in ฮ” (โ„“) โ„Ž . The positionwise LayerNorm/MLP, being ๐ต-Lipschitz, scales the spread by at most ๐ต. Collecting the constants into ๐ด 1 and, ๐ด 2 gives the desired result. โ–ก

We can now proof the full propagation result. Theorem B.9. For any finite depth ๐ฟ, there exists ๐œ€ > 0 (depending on ๐ต, ๐ฟ, and ๐ป) such that if ฮ” (1) โ„Ž โ‰ค ๐œ€, then for all ๐‘™ โ‰ค ๐ฟ,

This conclude the first part of the proof, regarding uniformity propagation across depth. Note that the bounds in the proof do not depend on the number of tokens in the input sequence.

A ๐‘ bound. Recall that,

where ๐‘ ๐‘– ๐‘— are centered positional weights, i.e. ๐‘—โ‰ค๐‘– ๐‘ ๐‘– ๐‘— = 0. For any such ๐‘ ๐‘– ๐‘— we have

Now, from direct computation and an application of Lemma B.10, we have

Let’s analyse the norm:

This concludes the proof of Theorem 3.4. et al., 2024) for over 16B tokens with a sequence length of 1024. We note this is well over 2 times the chinchilla optimal number of tokens from Hoffmann et al. (2022). We use a Qwen2 (Yang et al., 2024) tokenizer and follow the specifications (number of layers/hidden dimensions) from the 0.5B model from the same family. We implemented all our baselines on top of this architecture, pretraining them for the same large number of tokens. We use the AdamW optimizer Loshchilov and Hutter (2017) with a small warmup phase of 520 steps, a batch size of 1024, a peak learning rate of 3.0 ร— 10 -4 , and a cosine decay thereafter. For DroPE we followed a similar optimization setup, but only training for 2B total tokens using a shorter warmup of 70 steps and a slightly larger learning rate of 1.0 ร— 10 -3 to compensate for the shorter training budget. We provide a full list of hyperparameters and training specifications for this setting in the left column of Table 4.

DroPE from a pretrained SmolLM . For the second part of our experimental evaluation, we use a SmolLM (Allal et al., 2024) with around 362 million parameters already extensively pretrained on the SmolLM corpus (Ben Allal et al., 2024) for over 600B tokens with a sequence length of 2048 -almost 100 times the chinchilla optimal number. This model used a GPT2 (Radford et al., 2019) tokenizer and its architecture was designed to be similar to models of the Llama2 family (Touvron et al., 2023). While not all training details have been disclosed, Allal et al. ( 2024) explicitly mentions using the AdamW optimizer Loshchilov and Hutter (2017), a batch size of 512, a peak learning rate of 3.0 ร— 10 -3 , and a cosine decay thereafter. For DroPE we again tried to follow a similar optimization setup, across our different 30B/60B/120B training regimes, introducing a short warmup of 490 steps and a slightly lower learning rate of 1.0 ร— 10 -3 as we found their reported 3.0 ร— 10 -3 led to instabilities from the small batch size. Given the more extended training period, we used a simple QKNorm (Henry et al., 2020) after dropping the positional embeddings, which we found beneficial to mitigate sporadic instabilities from large gradients. We note that preliminary experiments showed that normalizing only the queries led to even faster learning and also successfully stabilized long training. We believe further exploration of this new Q-norm method could be an exciting direction for future work to train transformers without positional embeddings at even larger scales. We provide a full list of hyperparameters and training specifications for this setting in the right column of Table 4.

Needle-in-a-haystack. We evaluate long-context retrieval using the needle-in-a-haystack (NIAH) setup, which places a short “needle” inside a long distractor “haystack.” Following prior work (Kamradt, 2023), our haystack is a random excerpt from Paul Graham’s essays, and each needle is a seven-digit “magic number” paired with a short key/descriptor. We study three variants:

โ€ข (Standard NIAH) We insert a single needle and prompt the model to retrieve it.

โ€ข Multi-Query NIAH: We insert multiple (key, value) pairs and prompt the model to return as many values as possible for a given list of keys. For example: The special magic numbers for whispering-workhorse and elite-butterfly mentioned in the provided text are:. โ€ข (Multi-Key NIAH) We insert multiple (key, value) pairs but query for a single key, e.g., The special magic number for elite-butterfly mentioned in the provided text is:

โ€ข (Multi-Value NIAH) We associate multiple values with one key and ask for all of them without pointing to specific positions, e.g., What are all the special magic numbers for cloistered-colonization mentioned in the provided text?

Inserted needles and example targets are formatted in natural language, for instance, two examples include One of the special magic numbers for whispering-workhorse is: 1019173 and One of the special magic numbers for elite-butterfly is: 4132801. For the standard NIAH variant, we report the average success rate over all possible needle depths. For the multiple needles NIAH variants, we always insert four (key, value) needle pairs, placed at random sequence locations. Unless otherwise noted, we use greedy decoding (logit temperature = 0) for reproducibility.

Long-context evaluations. We use standard implementations of PI, RoPE-NTK, and YaRN. For tasks that require a fixed maximum context length (e.g., NIAH at 2ร— the training context), we set the extension factor ๐‘  manually. For settings that require reasoning across multiple context lengths and extended generations, we employ a dynamic scaling schedule that adjusts ๐›พ as a function of the generation length as detailed in Peng et al. (2023).

For DroPE, we follow Wang et al. (2024) and apply softmax temperature scaling when evaluating on longer sequences. In practice, we tune a single scalar logit scale (equivalently, the inverse temperature) on a held-out set at the target length. Analogous to (Peng et al., 2023), we fit this coefficient by minimizing perplexity to obtain the optimal scaling. For the DroPE model trained from scratch, the best-performing scale is ๐›ฝ โ˜… = 1 + 0.412 ln(๐‘ ), and for SmolLM-DroPE the optimal scale is ๐›ฝ โ˜… = 1 + 0.103 ln(๐‘ ), Where ๐‘  = ๐ถ test /๐ถ train is the context extension factor. Unless otherwise specified, all other decoding settings are held fixed across lengths.

Language modeling benchmarks. We evaluate SmolLM and SmolLM-DroPE on six standard multiple-choice benchmarks using the LightEval harness (Habib et al., 2023): ARC-E/C: gradeschool science QA split into Easy and Challenge sets, the latter defined by questions that defeat simple IR and co-occurrence baselines (Clark et al., 2018); HellaSwag: adversarially filtered commonsense sentence completion that is easy for humans but challenging for LMs (Zellers et al., 2019); Open-BookQA: combining a small “open book” of science facts with broad commonsense to answer 6K

โ€ข Lower learning rates (3 ร— 10 -5 , 3 ร— 10 -4 ). DroPE works effectively without QKNorm. At the lowest learning rate (3 ร— 10 -5 ), the model without QK Norm achieves a slightly better final loss (2.713 vs. 3.102). Together with the 3 ร— 10 -4 setting (2.530 vs.2.555), this indicates that QK Norm does not consistently improve performance in low-volatility regimes and is not the source of our gains. โ€ข High learning rate (10 -3 ). At the highest learning rate, the model without QKNorm becomes unstable (loss spikes, gradient explosions), leading to poor convergence (final loss 6.334). In contrast, adding QKNorm stabilizes training and allows us to leverage the higher learning rate to achieve the best overall performance (final loss 2.496).

Figure 12 shows the corresponding training curves with and without QK Norm, highlighting the presence of loss spikes at higher learning rates, in line with observations reported in OLMo et al. (2024a). These results empirically demonstrate that the primary role of QK Norm is to act as a stabilizer that enables the use of a more aggressive, compute-efficient learning rate. Importantly, DroPE can still be applied without QK Norm by using a moderate learning rate (e.g., (3 ร— 10 -4 ), which is our default setting for all experiments except the longer SmolLM-360M recalibration phases.

Observation 1. Positional information and attention non-uniformity, which are crucial for sequence modeling, develop at a bounded rate in NoPE transformers. In contrast, explicit PE methods, such as RoPE, provide a strong bias from the

rate schedule accordingly. Given the extended training periods, only for these experiments, we also add QKNorm(Henry et al., 2020) after dropping the positional embeddings, which we find beneficial for mitigating training instabilities, as noted by OLMo et al. (2024b) (See Appendix D.3).

rate schedule accordingly. Given the extended training periods, only for these experiments, we also add QKNorm(Henry et al., 2020)

rate schedule accordingly. Given the extended training periods, only for these experiments, we also add QKNorm

We start by analyzing how quickly our SmolLM-DroPE models can recover SmolLM’s in-context performance across six different LM reasoning benchmarks(Bisk et al., 2020;Clark et al., 2018;Mihaylov et al., 2018;Sakaguchi et al., 2021;Zellers et al., 2019). As shown in Figures 9 and 10 as well as Table5, even with our shortest training schedule, SmolLM-DroPE almost entirely matches SmolLM on every task, while with our longest schedule our new model manages to exceed its original performance. Furthermore, inspecting our model at every checkpoint throughout training, we find that DroPE recovers over 95% of SmolLM’s performance after less than 5B tokens, representing a minuscule 0.8% of SmolLM’s original budget.

We start by analyzing how quickly our SmolLM-DroPE models can recover SmolLM’s in-context performance across six different LM reasoning benchmarks(Bisk et al., 2020;Clark et al., 2018;Mihaylov et al., 2018;Sakaguchi et al., 2021;Zellers et al., 2019). As shown in Figures 9 and 10 as well as Table5

We start by analyzing how quickly our SmolLM-DroPE models can recover SmolLM’s in-context performance across six different LM reasoning benchmarks(Bisk et al., 2020;Clark et al., 2018;Mihaylov et al., 2018;Sakaguchi et al., 2021;Zellers et al., 2019). As shown in Figures 9 and 10 as well as Table

We start by analyzing how quickly our SmolLM-DroPE models can recover SmolLM’s in-context performance across six different LM reasoning benchmarks(Bisk et al., 2020;Clark et al., 2018;Mihaylov et al., 2018;Sakaguchi et al., 2021;Zellers et al., 2019)

We start by analyzing how quickly our SmolLM-DroPE models can recover SmolLM’s in-context performance across six different LM reasoning benchmarks

We empirically validate DroPE across different models and data scales, showing its effectiveness and potential to be integrated as a new core component of future state-of-the-art training pipelines. More broadly, our work demonstrates that canonical trade-offs in LM Attention. ๐‘‡ โˆˆ โ„

๐‘—โ‰ค๐‘–๐‘” ๐‘– ๐‘— (๐‘Ž ๐‘— -ฤ๐‘– ).๐‘—โ‰ค๐‘– ๐›ผ ๐‘– ๐‘— (๐‘ ๐‘– ๐‘— -๐‘ ๐›ผ ๐‘– ๐‘— ) = 1 ๐‘‡ ๐”ผ ๐‘—โˆผ๐›ผ ๐‘– [๐‘ ๐‘– ๐‘— -๐”ผ ๐‘โˆผ๐›ผ ๐‘– [๐‘ ๐‘–๐‘ ]] = 0.For the second part, observe that โˆ‘๏ธ ๐‘—โ‰ค๐‘– ๐‘” ๐‘– ๐‘— (๐‘Ž ๐‘— -ฤ๐‘– ) = โˆ‘๏ธ ๐‘—โ‰ค๐‘– ๐‘” ๐‘– ๐‘— ๐‘Ž ๐‘— -ฤ๐‘– โˆ‘๏ธ ๐‘—โ‰ค๐‘– ๐‘” ๐‘– ๐‘— = โˆ‘๏ธ ๐‘—โ‰ค๐‘– ๐‘” ๐‘– ๐‘— ๐‘Ž ๐‘— .

๐‘—โ‰ค๐‘–๐‘” ๐‘– ๐‘— (๐‘Ž ๐‘— -ฤ๐‘– ).๐‘—โ‰ค๐‘– ๐›ผ ๐‘– ๐‘— (๐‘ ๐‘– ๐‘— -๐‘ ๐›ผ ๐‘– ๐‘— ) = 1 ๐‘‡ ๐”ผ ๐‘—โˆผ๐›ผ ๐‘– [๐‘ ๐‘– ๐‘— -๐”ผ ๐‘โˆผ๐›ผ ๐‘– [๐‘ ๐‘–๐‘ ]] = 0.

๐‘—โ‰ค๐‘–๐‘” ๐‘– ๐‘— (๐‘Ž ๐‘— -ฤ๐‘– ).

๐‘—โ‰ค๐‘–

C.1

Note the softmax in Equation 1 is taken on the first ๐‘– tokens, implementing a causal mask.

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

โ†‘โ†“
โ†ต
ESC
โŒ˜K Shortcut