More Than Bits Multi-Envelope Double Binary Factorization for Extreme Quantization

Reading time: 32 minute
...

πŸ“ Original Paper Info

- Title: More Than Bits Multi-Envelope Double Binary Factorization for Extreme Quantization
- ArXiv ID: 2512.24545
- Date: 2025-12-31
- Authors: Yuma Ichikawa, Yoshihiko Fujisawa, Yudai Fujimoto, Akira Sakai, Katsuki Fujisawa

πŸ“ Abstract

For extreme low-bit quantization of large language models (LLMs), Double Binary Factorization (DBF) is attractive as it enables efficient inference without sacrificing accuracy. However, the scaling parameters of DBF are too restrictive; after factoring out signs, all rank components share the same magnitude profile, resulting in performance saturation. We propose Multi-envelope DBF (MDBF), which retains a shared pair of 1-bit sign bases but replaces the single envelope with a rank-$l$ envelope. By sharing sign matrices among envelope components, MDBF effectively maintains a binary carrier and utilizes the limited memory budget for magnitude expressiveness. We also introduce a closed-form initialization and an alternating refinement method to optimize MDBF. Across the LLaMA and Qwen families, MDBF enhances perplexity and zero-shot accuracy over previous binary formats at matched bits per weight while preserving the same deployment-friendly inference primitive.

πŸ’‘ Summary & Analysis

1. **Identifying DBF Bottleneck:** The bottleneck in DBF is its reliance on a single envelope constraint. This limits the expressivity of magnitude variations, leading to diminishing accuracy improvements with increased sign diversity. It’s like trying to paint an intricate picture with only one color. 2. **Multi-Envelope DBF Proposal:** MDBF overcomes this by introducing multiple envelope modes while maintaining the same deployment-friendly binary inference path. Imagine using multiple colors but still relying on simple tools for painting. 3. **ADB Refinement:** The initialization and ADMM refinement process in MDBF ensures efficient optimization, akin to refining a painting with multiple colors to achieve greater detail and accuracy.

πŸ“„ Full Paper Content (ArXiv Source)

# Introduction

Large language models (LLMs) support many NLP systems; however, their size renders deployment expensive, since storing FP16 or FP32 parameters and moving them through the memory hierarchy often dominates both the memory footprint and inference latency. Quantization is therefore a central tool for efficient deployment. Post-training quantization (PTQ) is particularly appealing because it can be applied to a pretrained model with minimal overhead, avoiding the need for full retraining. Although recent PTQ methods maintain strong accuracy at around 4-bit precision, performance typically degrades as precision approaches the 2–1-bit regime, where the per-layer information budget is extremely limited. To push below 2 bits lower, many approaches move beyond elementwise quantization and adopt structured parameterizations . Binary and near-binary schemes are especially appealing because they provide a clear hardware fast path: most computation can be performed by specialized kernels operating on bit-packed sign matrices, with only lightweight higher-precision scaling.

A prominent family of methods factorizes each weight matrix into low-rank components and then binarizes the factors. OneBit shows that appropriate scaling can stabilize 1-bit factors, while Double Binary Factorization (DBF) makes the binary path explicit by composing two binary matrix multiplications with interleaved diagonal scalings. LittleBit further enhances extreme-bit accuracy through multi-scale scaling and residual compensation, utilizing quantization-aware training (QAT) across multiple GPUs. Despite these advances, existing formats share a key structural limitation: after demodulation, factor magnitudes are confined to a single rank-one envelope. Increasing the inner rank primarily enhances sign diversity rather than magnitude expressiveness. As a result, under a fixed bits-per-weight budget, accuracy can saturate because gains come more from signs than magnitudes.

/>
Layer-wise reconstruction error vs.Β envelope rank and decomposition depth (LLaMA2 7B). We conduct experiments under a 1.5 bit quantization setting and report the relative Frobenius error $\nicefrac{\|W-\widehat{W}\|_{F}}{\|W\|_{F}}$ of MDBF as a function of the envelope rank l and the decomposition depth P, evaluated on three representative Transformer blocks of layers 0, 15, and 30 and central attention/MLP projections. Across layers and modules, larger l, which increases magnitude expressivity while sharing sign bases, consistently lowers reconstruction error. In contrast, larger P, which adds extra decompositions and sign bases, often worsens reconstruction at matched bits per weight. The best configuration in this sweep is (l*, P*) = (16, 1).

This paper addresses the DBF bottleneck by explicitly allocating limited expressivity to the critical components at extremely low precision. We propose Multi-Envelope Double Binary Factorization (MDBF), which retains the shared 1-bit sign bases and a deployment-friendly binary fast path while replacing the rank-one magnitude envelope with multiple demodulated envelope modes. FigureΒ 1 shows that increasing the envelope rank $`l`$ systematically reduces reconstruction error; however, while increasing the residual path $`P`$, as in LittleBit , is often less effective within a fixed bits-per-weight budget. MDBF adds a small number of real-valued degrees of freedom for magnitude modeling, better aligning with the empirically observed low-rank structure of Transformer weights, which are rarely rank-one. To make MDBF applicable for layer-wise PTQ, we introduce a layer-wise optimization pipeline with closed-form initialization followed by ADMM refinement. Across the LLaMA and Qwen families, MDBF improves perplexity and zero-shot accuracy compared to previous binary formats at matched bits per weight, particularly in the challenging 2–1 bit range, while maintaining the same deployment-friendly binary inference primitive.

Contributions.

  • Identifying DBF Bottleneck: We identify the bottleneck as the single envelope constraint. Under a fixed bits-per-weight budget, modeling magnitude variation yields greater accuracy gains than increasing sign diversity.

  • Multi-Envelope Generalization of DBF: We propose Multi-Envelope DBF (MDBF), which retains the shared 1-bit sign bases and maintains the same deployment-friendly binary fast path while replacing the rank-one magnitude envelope with a rank-$`l`$ envelope.

  • Initialization and ADMM Refinement for MDBF: We generalize the initialization of LittleBit and the ADMM-based refinement of DBF to a multi-envelope setting. This results in a closed-form initializer and an efficient alternating ADMM refinement procedure.

  • Empirical Validation: Across the LLaMA and Qwen model families, MDBF consistently reduces reconstruction error and improves perplexity and zero-shot accuracy compared to prior binary formats at matched BPW.

Notation

Vectors are denoted by bold lowercase letters, e.g., $`{\bm x}`$, and matrices by uppercase letters, e.g., $`W`$. Throughout, $`W\in{\mathbb R}^{N\times M}`$ denotes a real-valued weight matrix. We denote by $`\odot`$ the Hadamard elementwise product, and for a vector $`{\bm a}`$, let $`D_{{\bm a}}`$ represent the diagonal matrix with $`(D_{{\bm a}})_{ii}=a_i`$. We use $`\|\cdot\|_F`$ for the Frobenius norm and $`\langle A,B\rangle_{F}\coloneqq \mathrm{Tr}(A^\top B)`$ for the corresponding Frobenius inner product. For any matrix $`A`$, its singular values are denoted by $`\sigma_1(A)\ge \cdots \ge \sigma_{\min(N,M)}(A)\ge 0`$. For a target rank $`R\le \min(N,M)`$, we denote the rank-$`R`$ truncated SVD of $`W`$ as $`W_R = U_R\Sigma_R V_R^\top`$, where $`U_R\in{\mathbb R}^{N\times R}`$ and $`V_R\in{\mathbb R}^{M\times R}`$ have orthonormal columns and $`\Sigma_R=\mathrm{diag}(\sigma_1,\dots,\sigma_R)\in{\mathbb R}^{R\times R}`$. Finally, the entrywise sign function $`\mathrm{sign}(\cdot)`$ maps to $`\{\pm 1\}`$, with $`\mathrm{sign}(0)=+1`$.

Preliminaries

Low-Rank Approximation

A common approach to model compression exploits the empirically observed approximate low-rank structure of weight matrices. Given a weight matrix $`W \in {\mathbb R}^{N\times M}`$, we approximate it with a rank-$`R`$ factorization:

MATH
\begin{equation}
    \label{eq:low-rank-basis}
    W \approx U V^{\top},~~
    U \in {\mathbb R}^{N \times R},~V \in {\mathbb R}^{M \times R}.
\end{equation}
Click to expand and view more

Equivalently, $`UV^{\top}`$ expresses $`W`$ as a sum of $`R`$ rank-one components, $`UV^{\top}=\sum_{j=1}^{R}{\bm u}_{j}{\bm v}_{j}^{\top}`$, where $`{\bm u}_{j}`$ and $`{\bm v}_{j}`$ are the $`j`$-th columns of $`U`$ and $`V`$.

However, low rank alone does not guarantee meaningful memory savings. If $`U`$ and $`V`$ are stored in standard high-precision formats, such as FP16 or FP32, the parameter count becomes $`(N+M)R`$, which may be comparable to the original $`NM`$ parameters unless $`R \ll \min(N, M)`$. Moreover, maintaining accuracy often requires a moderate $`R`$, further limiting the compression benefit. Therefore, to achieve significant savings at very low bit widths while maintaining a sufficiently large effective rank, it is crucial to quantize the factors or introduce additional structures that facilitate both compact storage and efficient computation.

Double Binary Factorization

Double Binary Factorization (DBF)Β  represents a weight matrix $`W\in{\mathbb R}^{N\times M}`$ by utilizing two binary sign bases and diagonal rescalings:

MATH
\begin{equation}
    \label{eq:dbf-basis}
    \widehat{W}_{\mathrm{DBF}} \coloneqq D_{{\bm a}} S_{a} D_{{\bm m}} S_{b}^{\top} D_{{\bm b}},
\end{equation}
Click to expand and view more

where $`{\bm a}\in{\mathbb R}^{N}`$ and $`{\bm b}\in{\mathbb R}^{M}`$ are the row and column scaling vectors, $`{\bm m}\in{\mathbb R}^{R}`$ is an inner scaling vector, and $`S_{a}\in\{\pm 1\}^{N\times R}`$ and $`S_{b}\in\{\pm 1\}^{M\times R}`$ are binary sign matrices. EquationΒ [eq:dbf-basis] can be regarded as a structured rank-$`R`$ factorization:

MATH
\begin{multline}
    \label{eq:dbf-factors}
    \widehat{W}_{\mathrm{DBF}}=\widehat{U}\widehat{V}^{\top},\\
    \widehat{U} \coloneqq D_{{\bm a}} S_{a} D_{{\bm m}^{\nicefrac{1}{2}}},~
    \widehat{V} \coloneqq D_{{\bm b}} S_{b} D_{{\bm m}^{\nicefrac{1}{2}}}.
\end{multline}
Click to expand and view more

Single-Envelope Constraint.

For a sign mask $`S \in \{\pm1\}^{p\times q}`$ and a matrix $`Z\in{\mathbb R}^{p\times q}`$, we define the demodulated envelope of $`Z`$ concerning $`S`$ as $`E_S(Z) \coloneqq S\odot Z`$. In DBF, the demodulated envelopes of factors $`\widehat{U}`$ and $`\widehat{V}`$ satisfy the following exact identities:

MATH
\begin{multline}
    \label{eq:dbf-demod-envelope}
    E_{S_a}(\widehat{U}) = S_a\odot\widehat{U} = {\bm a}({\bm m}^{\nicefrac{1}{2}})^{\top}, \\
    E_{S_b}(\widehat{V}) = S_b\odot\widehat{V} = {\bm b}({\bm m}^{\nicefrac{1}{2}})^{\top}.
\end{multline}
Click to expand and view more

Therefore, the following holds:

MATH
\begin{equation}
    \mathrm{rank}(E_{S_a}(\widehat{U}))\le 1,~~
    \mathrm{rank}(E_{S_b}(\widehat{V}))\le 1,
\end{equation}
Click to expand and view more

after demodulation by the shared sign bases, both factor envelopes constitute rank-one outer products. Equivalently, each column of $`\widehat{U}`$ shares the same row-wise envelope profile $`{\bm a}`$ up to a scalar multiplier $`m_j^{1/2}`$, and similarly $`{\bm b}`$ applies to $`\widehat{V}`$. We refer to this rank-one demodulated-envelope constraint as the single envelope constraint.

Inference.

Let $`X\in{\mathbb R}^{T\times N}`$ denote the input activations and $`Y=X\widehat{W}_{\mathrm{DBF}}\in{\mathbb R}^{T\times M}`$ the output. DBF admits an efficient evaluation order in which diagonal scalings are interleaved with two binary matrix multiplications:

MATH
\begin{equation}
    \label{eq:DBF-inference-form}
    Y = X\widehat{W}_{\mathrm{DBF}}
    = \Bigl(\Bigl(\Bigl(\bigl(X D_{{\bm a}}\bigr) S_{a}\Bigr) D_{{\bm m}}\Bigr) S_{b}^{\top}\Bigr) D_{{\bm b}}.
\end{equation}
Click to expand and view more

Multiplication by a diagonal matrix is equivalent to elementwise rescaling and is typically bandwidth-light compared to matrix multiplication. Therefore, DBF substitutes one high-precision GEMM with two binary GEMMs and inexpensive diagonal operations. The effective bits-per-weight is determined by the storage of bitpacked $`\{S_a,S_b\}`$ and real-valued vectors $`\{{\bm a},{\bm b},{\bm m}\}`$, relative to the $`NM`$ entries of the original matrix.

Residual Compensation.

LittleBit improves fidelity by augmenting a single DBF approximation with an additional DBF term trained on the residual. LittleBit specifically employs a two-term decomposition as follows:

MATH
\begin{multline}
\label{eq:dbf-residual-comp}
    \widehat{W}_{\mathrm{LittleBit}}
    =D_{{\bm a}^{(1)}} S_{a}^{(1)} D_{{\bm m}^{(1)}} S_{b}^{(1) \top} D_{{\bm b}^{(1)}}
    \\
    +D_{{\bm a}^{(2)}} S_{a}^{(2)} D_{{\bm m}^{(2)}} S_{b}^{(2) \top} D_{{\bm b}^{(2)}}.
\end{multline}
Click to expand and view more

A natural optimization strategy is stagewise residual fitting: first, fit the initial term, then compute the residual term:

MATH
\begin{equation}
    \widehat{R} \coloneqq W-D_{{\bm a}^{(1)}} S_{a}^{(1)} D_{{\bm m}^{(1)}} S_{b}^{(1) \top} D_{{\bm b}^{(1)}}.
\end{equation}
Click to expand and view more

Fit the second DBF term to approximate $`\widehat{R}`$. This residual-compensation scheme captures structures that are difficult to represent with a single DBF component.

Method

/>
Entropy-based effective rank across Transformer layers for LLaMA2 13B. We report the average effective rank of the full weight matrices W, their binarized signs sign(W), and the demodulated envelopes |U| and |V|. The envelopes remain consistently above rank one, indicating multiple magnitude modes and motivating the relaxation of the single-envelope constraint.

Bottleneck: Single-Envelope Constraint

In practice, DBF exhibits a performance ceiling; increasing the inner dimension $`R`$ results in diminishing returns, and accuracy saturates below that of a real-valued low-rank approximation with a comparable parameter count. This gap is not merely empirical but structural. As shown in Section 3, demodulation indicates that each DBF factor is constrained by a rank-one amplitude envelope shared across all $`R`$ columns. Enlarging $`R`$ primarily increases sign-pattern diversity rather than enhancing the capacity to represent magnitudes, fundamentally limiting achievable accuracy.

The assumption of a rank-one envelope is inconsistent with the empirical structure of LLM weights. To quantify this point, we analyze LLaMA2 13B and compute the entropy-based effective ranks of the demodulated envelopes $`|U|`$ and $`|V|`$; see AppendixΒ 10.1 for a detailed definition. FigureΒ 2 shows the average effective rank across Transformer blocks and their linear layers. Although the envelopes are low-rank, they are consistently not rank-one; their effective ranks remain moderately above 1, suggesting multiple magnitude modes that DBF cannot represent under the single-envelope constraint.

These observations clarify why merely increasing sign variation, such as in residual-style multi-term designs like LittleBit, is not the most direct solution. The goal is to preserve a deployment-friendly inference path, where shared binary carriers facilitate efficient 1-bit computation while alleviating the bottleneck. This motivation leads to Multi-Envelope DBF, which enhances expressivity by allowing higher-rank amplitude envelopes to be used under the same shared binary sign bases, without altering the inference-time execution model.

More Than Bits: DBF is fundamentally limited by its rank-one amplitude envelope, rather than by the available bit budget.

Multi-Envelope Double Binary Factorization

We propose Multi-Envelope DBF (MDBF), which preserves the shared 1-bit sign bases while allowing the demodulated envelopes of the factor matrices to have a rank of at most $`l`$ as follows:

MATH
\begin{equation}
    \label{eq:me-dbf-basis}
    \widehat{W}_{l}\coloneqq \widehat{U}_{l}\widehat{V}_{l}^{\top},
    \widehat{U}_{l} = S_{a}\odot A Q^{\top},
    \widehat{V}_{l} = S_{b}\odot B G^{\top},
\end{equation}
Click to expand and view more

where $`S_{a}\in\{\pm 1\}^{N\times R}`$ and $`S_{b}\in\{\pm 1\}^{M\times R}`$ are binary sign matrices, while $`A\in{\mathbb R}^{N\times l}`$, $`Q\in{\mathbb R}^{R\times l}`$, $`B\in{\mathbb R}^{M\times l}`$, and $`G\in{\mathbb R}^{R\times l}`$ are real-valued factors. By construction, the demodulated envelopes are low-rank:

MATH
\begin{multline}
    \label{eq:me-dbf-envelope-rank}
    E_{S_a}(\widehat{U}_l)=S_a\odot \widehat{U}_l = AQ^{\top}, \\
    E_{S_b}(\widehat{V}_l)=S_b\odot \widehat{V}_l = BG^{\top},
\end{multline}
Click to expand and view more

leads to $`\mathrm{rank}(E_{S_a}(\widehat{U}_l))\le l`$ and $`\mathrm{rank}(E_{S_b}(\widehat{V}_l))\le l`$. Let $`A=[{\bm a}^{(1)},\dots,{\bm a}^{(l)}]`$ and $`Q=[{\bm q}^{(1)},\dots,{\bm q}^{(l)}]`$, along with $`B=[{\bm b}^{(1)},\dots,{\bm b}^{(l)}]`$ and $`G=[{\bm g}^{(1)},\dots,{\bm g}^{(l)}]`$, yield the expansions

MATH
\begin{equation}
\label{eq:me-dbf-expand}
    \widehat{U}_{l}=\sum_{t=1}^{l} D_{{\bm a}^{(t)}} S_{a} D_{{\bm q}^{(t)}},
    \widehat{V}_{l}=\sum_{s=1}^{l} D_{{\bm b}^{(s)}} S_{b} D_{{\bm g}^{(s)}}.
\end{equation}
Click to expand and view more

Thus, $`\widehat{W}_{l}`$ admits an equivalent sum-of-DBF form:

MATH
\begin{equation}
\label{eq:me-dbf-sum}
    \widehat{W}_{l}
    =
    \sum_{t=1}^{l}\sum_{s=1}^{l}
    D_{{\bm a}^{(t)}} S_{a} D_{{\bm q}^{(t)} \odot {\bm g}^{(s)}} S_{b}^{\top} D_{{\bm b}^{(s)}}.
\end{equation}
Click to expand and view more

MDBF generalizes DBF: when $`l=1`$, EquationΒ [eq:me-dbf-basis] reduces to a single DBF term involving $`{\bm m}={\bm q}^{(1)}\odot {\bm g}^{(1)}`$; if one enforces $`{\bm m}\ge 0`$, $`{\bm q}^{(1)}={\bm g}^{(1)}={\bm m}^{\nicefrac{1}{2}}`$ may be set to column-sign absorption in $`S_{a},S_{b}`$.

Multi-Envelope Structure

We next formalize the role of envelope rank $`l`$ and derive the corresponding closed-form optimizer using a fixed sign mask. The key idea is to decompose each factor into (i) a binary sign mask and (ii) a demodulated envelope, while imposing a rank constraint on the envelope.

Demodulation under Fixed Sign Mask.

Fix a sign mask $`S\in\{\pm1\}^{N\times R}`$ and consider an arbitrary factor matrix $`U\in{\mathbb R}^{N\times R}`$. We define the demodulated envelope as follows:

MATH
\begin{equation}
    E_{S}(U) \coloneqq S \odot U.
\end{equation}
Click to expand and view more

Since $`S_{ij}^{2}=1`$ holds for all $`(i,j)`$, we obtain $`S\odot(S\odot Z)=Z`$ for any matrix $`Z`$. Thus, once $`S`$ is fixed, $`U`$ is determined by its envelope $`E_S(U)`$.

Consider an approximation $`\widehat{U}`$ that shares the same mask $`S`$. It can be expressed as $`\widehat{U}=S\odot \widehat{E}`$ for some envelope matrix $`\widehat{E}\in{\mathbb R}^{N\times R}`$. Moreover, by LemmaΒ 5, the mapping $`Z\mapsto S\odot Z`$ preserves Frobenius norms:

MATH
\begin{equation}
    \label{eq:envelope-isometry}
    \|U-\widehat{U}\|_F =
    \|E_{S}(U)-\widehat{E}\|_F.
\end{equation}
Click to expand and view more

Therefore, under a fixed mask $`S`$, approximating $`U`$ in Frobenius norm is equivalent to approximating its demodulated envelope.

Rank-$`l`$ Envelope Class.

MDBF constrains the envelope to a low rank. For $`l\ge 1`$, we define the feasible family as follows:

Definition 1. *Let $`S\in\{\pm1\}^{N\times R}`$ and $`l\ge 1`$. Define

MATH
\begin{equation}
    {\mathcal F}_l(S)\coloneqq
    \ab\{S\odot E  \mid E\in{\mathbb R}^{N\times R}, \mathrm{rank}(E) \le l \}.
\end{equation}
```*

</div>

Equivalently, $`\widehat{U}\in{\mathcal F}_l(S)`$ if and only if its
envelope $`E_S(\widehat{U})=S\odot \widehat{U}`$ has a rank of at most
$`l`$. This implies the MDBF factor structure; for example,
$`\widehat{U}_l=S_a\odot(AQ^\top)`$ satisfies
$`\mathrm{rank}(E_{S_a}(\widehat{U}_l))=\mathrm{rank}(AQ^\top)\le l`$.

#### Closed-form Minimizer.

By EquationΒ <a href="#eq:envelope-isometry" data-reference-type="eqref"
data-reference="eq:envelope-isometry">[eq:envelope-isometry]</a>, the
best approximation of $`U`$ within $`{\mathcal F}_l(S)`$ becomes the
classical best rank-$`l`$ approximation problem for the envelope matrix
$`E_S(U)`$. The following theorem demonstrates this equivalence and
provides both the optimal error and a minimizer.

<div id="thm:rankl-error-main" class="theorem">

**Theorem 2**. *Let $`U\in{\mathbb R}^{N \times R}`$ fix a sign mask
$`S\in\{\pm 1\}^{N \times R}`$. Let
$`\sigma_1(E_S(U))\ge \cdots \ge \sigma_{\min(N,R)}(E_S(U))\ge 0`$ be
the singular values of $`E_S(U)=S\odot U`$. Fix an integer $`l`$ such
that $`1\le l\le \min(N, R)`$. Then
``` math
\begin{equation}
    \label{eq:rankl-error}
    \min_{\widehat{U}\in{\mathcal F}_l(S)} \|U-\widehat{U}\|_{F}
    =
    \ab(\sum_{i=l+1}^{\min(N, R)}\sigma_i\bigl(E_S(U)\bigr)^2)^{\nicefrac{1}{2}}.
\end{equation}
Click to expand and view more

Moreover, one minimizer is

MATH
\begin{equation}
        \label{eq:Ustar}
        \widehat{U}^{\ast}=S\odot \mathrm{TSVD}_{l}(E_{S}(U)),
\end{equation}
Click to expand and view more

where $`\mathrm{TSVD}_l(\cdot)`$ denotes the rank-$`l`$ truncated SVD, which provides the best approximation in the Frobenius norm.*

The minimizer EquationΒ [eq:Ustar] directly follows from EquationΒ [eq:envelope-isometry]. Writing $`\widehat{U}=S\odot \widehat{E}`$ with $`\mathrm{rank}(\widehat{E})\le l`$ yields $`\|U-\widehat{U}\|_F=\|E_S(U)-\widehat{E}\|_F`$, thereby identifying the optimal choice as the best rank-$`l`$ approximation $`\widehat{E}=\mathrm{TSVD}_l(E_S(U))`$ according to the Eckart–Young–Mirsky theorem; remodulating with $`S`$ results in EquationΒ [eq:Ustar].

TheoremΒ 2 clarifies the role of $`l`$: for a fixed binary basis $`S`$, MDBF approximates the demodulated envelope $`E_S(U)`$ using a rank-$`l`$ matrix and then reattaches the same mask. Thus, MDBF does not allocate additional degrees of freedom to sign patterns; rather, it enhances capacity by permitting multiple envelope modes in the demodulated domain. DBF represents the special case $`l=1`$, which constrains the envelope to rank-one and requires all $`R`$ columns to share a single envelope direction up to scalar weights. For $`l>1`$, the feasible set strictly expands, $`{\mathcal F}_1(S)\subseteq {\mathcal F}_l(S)`$, and the optimal error in Equation [eq:rankl-error] can only decrease.

Layer-Wise Optimization

We adopt a layer-wise PTQ setting, in which we quantize each weight matrix $`W\in{\mathbb R}^{N\times M}`$ independently under strict memory and computational constraints. Even with the binary sign bases fixed, optimizing MDBF remains non-convex; the objective is bilinear in the factor matrices, and we impose low-rank constraints on their demodulated envelopes. We propose a two-stage pipeline: (i) a closed-form initialization based on Multi-Envelope SVID (MSVID) and (ii) local refinement using alternating updates of ADMM. The resulting factors also provide a robust warmup start for QAT.

Initialization via Multi-Envelope SVID

We begin by defining the operator used in both initialization and refinement. For any matrix $`Z \in {\mathbb R}^{N \times M}`$ and any $`l`$ with $`1\le l\le \min(N,M)`$, define

MATH
\begin{equation}
\label{eq:msvid-def}
\mathrm{MSVID}_{l}[Z] \coloneqq \mathrm{sign}(Z) \odot \mathrm{TSVD}_{l}(|Z|).
\end{equation}
Click to expand and view more

Equivalently, let $`S_{Z} \coloneqq \mathrm{sign}(Z)`$, and noting that the demodulated envelope satisfies $`E_{S_{Z}}(Z)=S_{Z}\odot Z =|Z|`$, the operator $`\mathrm{MSVID}_l[Z]`$ preserves the sign mask $`S_Z`$ while replacing the envelope with its optimal rank-$`l`$ approximation. The special case $`l=1`$ corresponds to a rank-one envelope.

Given $`W\in{\mathbb R}^{N\times M}`$ and an inner dimension $`R`$ with $`1\le R\le \min(N,M)`$, we first compute the rank-$`R`$ truncated SVD, $`W_R = U_R\Sigma_R V_R^{\top}`$, to form balanced factors

MATH
\begin{equation}
    \label{eq:mesvid-balanced}
    U_{0} \coloneqq U_R\Sigma_{R}^{\nicefrac{1}{2}} \in{\mathbb R}^{N\times R},~
    V_{0} \coloneqq V_R\Sigma_R^{\nicefrac{1}{2}}\in{\mathbb R}^{M\times R},
\end{equation}
Click to expand and view more

so that $`W_R=U_0 V_0^{\top}`$. We then initialize MDBF by applying Equation [eq:msvid-def] to each continuous factor:

MATH
\begin{equation}
    \label{eq:mesvid-init-compact}
    \widehat{U}_l^{(0)} \coloneqq \mathrm{MSVID}_l[U_0],~
    \widehat{V}_l^{(0)} \coloneqq \mathrm{MSVID}_l[V_0].
\end{equation}
Click to expand and view more

By construction, $`\widehat{U}_{l}^{(0)}`$ and $`\widehat{V}_{l}^{(0)}`$ possess rank-$`l`$ demodulated envelopes under their induced sign masks, conforming to the MDBF factor structure.

ADMM Refinement

Starting from $`\widehat{W}^{(0)}=\widehat{U}_{l}^{(0)}\widehat{V}_{l}^{(0) \top}`$, we refine the factors through alternating updates that minimize the reconstruction objective $`\|W-UV^{\top}\|_{F}^{2}`$ while repeatedly imposing the multi-envelope structure via $`\mathrm{MSVID}_l[\cdot]`$. Specifically, with $`V`$ fixed, we conduct a limited number of ADMM inner iterations using an auxiliary variable $`\widetilde{U}`$ and a scaled dual variable $`\Lambda`$:

MATH
\begin{align}
    \widetilde{U}^{(k+1)}
    &= \ab(WV^{(k)}+\rho (U^{(k)}-\Lambda^{(k)})) \\
    &\times \ab(V^{(k) \top} V^{(k)}+\rho I)^{-1},
    \label{eq:admm-U-tilde}\\
    U^{(k+1)}
    &=
    \mathrm{MSVID}_{l} \ab[\widetilde{U}^{(k+1)}+\Lambda^{(k)}],
    \label{eq:admm-U-prox}\\
    \Lambda^{(k+1)}
    &=
    \Lambda^{(k)}+\widetilde{U}^{(k+1)}-U^{(k+1)},
    \label{eq:admm-U-dual}
\end{align}
Click to expand and view more

where $`\rho>0`$ and $`I`$ are the $`R \times R`$ identity matrix. We update $`V`$ analogously by applying the same steps to the transposed problem while keeping $`U`$ fixed.

Since $`\mathrm{MSVID}_l[\cdot]`$ updates the sign mask via $`\mathrm{sign}(\cdot)`$, it is not an exact Frobenius projection onto any fixed-mask constraint set, and standard ADMM convergence guarantees do not apply directly. Therefore, we consider the method as an ADMM-inspired scheme that alternates between closed-form least-squares updates and a mask-adaptive rank-$`l`$ envelope step. Crucially, Theorem 2 provides the theoretical backbone: with strong initialization, subsequent updates function as a refinement mechanism that improves the solution in practice.

Experiments

Model Bits (l = 1, P = 1) (l = 1, P = 2) (l = 2, P = 1) (l = 8, P = 1) (l = 16, P = 1)
Tiny LLaMA (1.1B) 1.00 78.14 114.84 70.17 70.04 60.40
1.25 21.96 37.71 22.50 21.67 19.98
1.50 15.85 19.85 16.41 14.94 14.27
Qwen3 (0.6B) 1.00 343.80 773.01 415.91 262.56 277.12
1.25 120.89 221.01 121.23 128.35 111.70
1.50 100.84 121.72 78.77 85.15 70.97
LLaMA3.2 (1B) 1.00 118.69 131.20 118.17 119.82 104.28
1.25 47.99 84.12 42.35 39.88 41.46
1.50 47.78 44.90 39.50 41.06 38.54
LLaMA2 (7B) 1.00 27.96 37.92 28.01 29.72 28.31
1.25 13.40 21.67 12.82 15.83 18.26
1.50 9.81 12.14 9.83 9.94 10.50
LLaMA3 (8B) 1.00 1222.32 9118.75 1024.80 2094.86 1639.64
1.25 58.61 165.43 50.13 92.02 164.24
1.50 35.62 41.40 36.11 30.14 31.70
Qwen3 (8B) 1.00 189.02 309.30 108.90 134.00 115.41
1.25 41.10 101.54 53.24 63.73 69.18
1.50 31.72 39.02 40.22 30.09 32.28

Setting

Baselines and Quantization Methods.

We use representative low-rank binary-factor formats as baselines: DBF, corresponding to $`(l,P)=(1,1)`$, and LittleBit, corresponding to $`(l,P)=(1,2)`$. For each method, we optimize the binary-factor parameters using ADMM for 1,000 outer iterations, with 3 inner ADMM updates per outer step. For formats utilizing $`P`$ decomposed terms, we apply this ADMM procedure independently for each term, resulting in a total of $`1{,}000 \times P`$ outer iterations. Following ADMM, we perform an additional gradient-based refinement using Adam for 1,500 steps, with a learning rate of $`0.01`$. During this refinement stage, we update only the real-valued parameters by minimizing the squared reconstruction loss $`\|W-\widehat{W}\|_F^2`$. For each target bit budget of $`\{1.00, 1.25, 1.50\}`$ bits per weight, we choose the largest inner rank that meets the budget, following .

To enhance accuracy in the low-bit regime, we incorporate an error-compensation technique. Specifically, we adopt quantization error propagation (QEP) with $`\alpha=0.5`$ and use 512 samples from WikiText2 to estimate the required statistics. Finally, in line with standard practice, we retain the first four and last four layers in full precision to ensure a fair and stable comparison across methods.

Models and Datasets.

We evaluate the proposed low-rank binary-factor format and baseline variants on widely used open-weight LLM families, including LLaMA2 , LLaMA3 , and Qwen. Specifically, we analyze the standard benchmarks of Tiny LLaMA 1.1B, LLaMA2 7B, LLaMA3 8B, and LLaMA3.2 1B in prior quantization work . To assess robustness across architectural variations, we also evaluate Qwen3 0.6B and Qwen3 8B, which differ from LLaMA models by incorporating RMSNorm variants in their attention components.

Evaluation.

We follow established evaluation practices for LLM quantization . We report perplexity (PPL) on WikiText2 and evaluate downstream zero-shot performance on ARC-Easy, and PIQA. In the main text, we present the average accuracy across tasks and include per-task results in AppendixΒ 11. All experiments were conducted on a single NVIDIA B200 GPU.

Result

Model Bits (l = 1, P = 1) (l = 1, P = 2) (l = 2, P = 1) (l = 8, P = 1) (l = 16, P = 1)
Tiny LLaMA (1.1B) 1.00 0.4398 0.4355 0.4423 0.4557 0.4449
1.25 0.4945 0.4608 0.5049 0.5045 0.5179
1.50 0.5252 0.5053 0.5273 0.5311 0.5259
Qwen3 (0.6B) 1.00 0.4237 0.4110 0.4244 0.4120 0.4228
1.25 0.4390 0.4213 0.4449 0.4449 0.4457
1.50 0.4519 0.4317 0.4604 0.4544 0.4609
LLaMA3.2 (1B) 1.00 0.4639 0.4716 0.4691 0.4680 0.4691
1.25 0.4977 0.4842 0.5039 0.5196 0.5099
1.50 0.5082 0.5202 0.5222 0.5386 0.5240
LLaMA2 (7B) 1.00 0.5263 0.4919 0.5217 0.5229 0.5229
1.25 0.5981 0.5522 0.6096 0.5767 0.5785
1.50 0.6398 0.6110 0.6480 0.6243 0.6341
LLaMA3 (8B) 1.00 0.4078 0.3997 0.4104 0.4185 0.4095
1.25 0.4796 0.4370 0.5004 0.4716 0.4443
1.50 0.5208 0.5056 0.5059 0.5069 0.5000
Qwen3 (8B) 1.00 0.4635 0.4508 0.4588 0.4486 0.4589
1.25 0.4971 0.4646 0.4993 0.4819 0.4854
1.50 0.5186 0.4880 0.5159 0.5263 0.5119

Reconstruction Error.

A central motivation for MDBF is that, under an extremely tight bit budget, additional sign diversity may inefficiently utilize capacity compared to enhancing magnitude expressivity. We investigate this by analyzing how layer-wise reconstruction error depends on (i) the envelope rank $`l`$, which increases magnitude capacity while keeping the shared binary sign bases, and (ii) the residual path $`P`$, which introduces extra factor terms and additional sign bases, as seen in residual binary decompositions . We evaluate LLaMA2-7B on three representative Transformer blocks (layers $`0`$, $`15`$, and $`30`$) and compute the reconstruction error for the attention and MLP projection matrices. For each weight matrix $`W`$ and its quantized approximation $`\widehat{W}`$, we report the relative Frobenius error $`\nicefrac{\|W-\widehat{W}\|_F}{\|W\|_F}`$ while sweeping $`l \in \{1,2,3,8,16\}`$ and $`P \in \{1,2,3\}`$ at matched bits per weight. Although larger envelope ranks, e.g., $`l \ge 16`$, may further improve performance, we defer these settings to future work due to their higher computational cost; notably, $`l=16`$ is sufficient to demonstrate their effectiveness.

FigureΒ 1 shows a consistent trend across layers and modules. For any fixed $`P`$, increasing $`l`$ reduces reconstruction error, whereas increasing $`P`$ is generally ineffective within the same bit budget. This behavior supports the design choice of MDBF: to reuse a shared binary carrier and allocate the limited capacity to a richer magnitude structure, rather than duplicating sign bases. In contrast, a larger $`P`$ necessitates the storage of additional sign matrices, diverting bits from magnitude modeling and potentially increasing approximation error. The effect is strongest in the earliest block, where the gap across $`(l,P)`$ is most significant among several projections. Since quantization noise introduced early can propagate through the network, minimizing errors in the early layers is particularly important.

Perplexity and Zero-shot Accuracy.

TableΒ [tab:ppl_wikitext2_bits_by_lP] demonstrates that MDBF is the strongest across models when comparing matched bits-per-weight, resulting in consistent reductions in perplexity. Notably, the improvements are most significant in the extremely low bit regime. TableΒ [tab:avg_acc_bits_by_lP] also presents downstream zero-shot performance and displays trends closely aligned with those in TableΒ [tab:ppl_wikitext2_bits_by_lP].

Across a fixed bit budget, we find that increasing the envelope rank $`l`$ contributes more to performance than increasing $`P`$. Specifically, larger $`l`$ yields improvements in both perplexity and zero-shot accuracy, while larger $`P`$ provides more modest enhancements. In this regime, our method significantly outperforms the baselines, often by a substantial margin. By contrast, for larger models around the 7B scale, the gains begin to saturate. We observe that the quantization error continues to decline, suggesting that the method may be overfitting the calibration data instead of translating these improvements into downstream gains. Exploring whether these trends persist with significantly larger calibration sets is an important direction for future work.

Related Work

Post-Training Quantization at $`\sim`$4 Bits.

Most practical low-bit deployments of LLMs rely on PTQ in the roughly 4-bit range. Layer-wise calibration methods, such as GPTQ and AWQ , achieve strong accuracy without full retraining by minimizing layer-wise reconstruction errors or by identifying and preserving salient weight channels during calibration. However, when the bit budget increases to $`\leq 2`$ bits per weight, these techniques often result in significant accuracy degradation, underscoring the need for more structured parameterizations.

Structured Representations for 2–3 Bits.

To achieve the 2–3 bit regime, recent studies often go beyond purely elementwise quantization to adopt structured weight representations. A representative line combines incoherence processing with lattice- and trellis-coded quantizers, as seen in QuIP# , QTIP , and AQLM . By encoding weights using structured codebooks instead of independent rounding, these methods enhance effective expressivity within tight bit budgets. In this regime, heuristic STE-based adjustments tend to be unstable ; therefore, refinement is typically framed as a discrete optimization problem .

Methods Targeting 1Β Bit and Below.

At or below 1 bit precision, maintaining model quality generally requires more flexibility than naive rounding permits. Representative approaches include SVID-based parameterizations , token-adaptive mixtures of scaling factors , and structured sparsity . In parallel, binary-factor formats maintain efficient inference primitives by decomposing each weight matrix into two bit-packed sign matrices with low-cost diagonal scaling, resulting in inference primarily governed by 1-bit kernels. A common theme is the explicit addition of degrees of freedom, such as adaptive scales, learnable sparsity patterns, or residual corrections, to mitigate the information bottleneck caused by extreme quantization.

Conclusion

We study extreme low-bit quantization using binary-factor representations and identify a key limitation of DBF: after sign demodulation, all weight magnitudes are constrained to a single shared envelope. This collapse significantly constrains expressivity and leads to diminishing returns as the inner rank increases, explaining the observed performance saturation. To address this bottleneck, we introduce MDBF, which enables deployment-friendly binary computation by preserving shared 1-bit sign bases while replacing the single-envelope with a rank-$`l`$ demodulated envelope that can represent multiple magnitude modes. We present a practical layer-wise PTQ pipeline featuring closed-form initialization followed by an alternating refinement procedure. Experiments across the LLaMA and Qwen model families demonstrate that MDBF improves perplexity and zero-shot accuracy compared to prior binary formats with the same bits per weight, highlighting that limited capacity is more effectively allocated to modeling magnitudes rather than replicating sign patterns.

Limitations

We focus on layer-wise PTQ. Accordingly, we do not consider submodule-levelΒ  or block-wiseΒ  quantization, end-to-end QAT, or distillation-based methods such as LittleBitΒ . While these alternatives may yield additional accuracy gains, they typically require significantly greater computational resources and introduce nontrivial engineering overhead. Our approach minimizes a non-convex objective through heuristic alternating updates. As a result, we do not provide guarantees of global optimality; the best-performing hyperparameter settings, e.g., envelope rank $`l`$ and decomposition depth $`P`$, may depend on the optimizer and initialization. Developing more principled optimization strategies or employing robust discrete general optimization methods and model-selection criteria for extremely low-bit binary-factor formats remains an important open problem.

Acknowledgments

This work was supported by JST BOOST, Japan (Grant No. JPMJBY24D0), and RIKEN Center for AIP. Additionally, this work received support from the Strategic Innovation Promotion Program (SIP) of the Cabinet Office, Government of Japan.

Derivation

Proofs of Main Results

This section presents detailed proofs of the results stated in the main text, along with the necessary auxiliary lemmas for completeness.

Demodulated envelopes and Envelope Class

Demodulated Envelope.

Let $`S\in\{\pm1\}^{p\times q}`$ represent a sign mask and $`Z\in{\mathbb R}^{p\times q}`$ an arbitrary matrix. We define the demodulated envelope of $`Z`$ concerning $`S`$ as follows:

MATH
\begin{equation}
    \label{eq:demod-envelope-app}
    E_S(Z)\coloneqq S\odot Z,
\end{equation}
Click to expand and view more

Since $`S_{ij}^2=1`$ holds for all $`(i,j)`$, we have $`S\odot S=\mathbf{1}`$ entrywise; therefore, $`Z = S\odot E_S(Z)`$.

Envelope Class.

We recall the rank-$`l`$ envelope class associated with a fixed sign mask.

Definition 3. *Let $`S\in\{\pm1\}^{p\times q}`$ and $`l\ge 1`$. Define

MATH
\begin{equation}
    \label{eq:Fl-rank-form}
    {\mathcal F}_l(S)\coloneqq
    \left\{
    S\odot E  \mid E\in{\mathbb R}^{p\times q}, \mathrm{rank}(E)\le l
    \right\}
    \subset {\mathbb R}^{p\times q}.
\end{equation}
Click to expand and view more

Equivalently, $`\widehat{Z}\in{\mathcal F}_l(S)`$ if and only if its demodulated envelope $`E_S(\widehat{Z})`$ has a rank of at most $`l`$.*

The next lemma demonstrates that $`{\mathcal F}_l(S)`$ can be equivalently characterized as a sum of diagonal scalings of the fixed mask $`S`$, thereby relating envelopes to DBF-like parameterizations.

Lemma 4. *Let $`S\in\{\pm1\}^{p\times q}`$ and $`l\ge 1`$. Then

MATH
\begin{equation}
    \label{eq:Fl-diag-form}
    {\mathcal F}_l(S)
    =
    \left\{
    \sum_{t=1}^{l} D_{{\bm x}^{(t)}} S D_{{\bm y}^{(t)}} \mid
    {\bm x}^{(t)}\in{\mathbb R}^{p}, {\bm y}^{(t)}\in{\mathbb R}^{q}
    \right\}.
\end{equation}
```*

</div>

<div class="proof">

*Proof.* We prove both inclusions. Let
$`\widehat{Z}\in{\mathcal F}_l(S)`$. By Definition
<a href="#def:Fl-app" data-reference-type="ref"
data-reference="def:Fl-app">3</a>, there exists
$`E\in{\mathbb R}^{p\times q}`$ such that $`\mathrm{rank}(E)\le l`$ with
$`\widehat{Z}=S\odot E`$. Since $`\mathrm{rank}(E)\le l`$, there exist
vectors $`\{{\bm x}^{(t)}\}_{t=1}^l\subset{\mathbb R}^p`$ and
$`\{{\bm y}^{(t)}\}_{t=1}^l\subset{\mathbb R}^q`$ such that
$`E=\sum_{t=1}^l {\bm x}^{(t)}({\bm y}^{(t)})^\top.`$ For any $`(i,j)`$,
we compute
``` math
\begin{align}
        \widehat{Z}_{ij}
        &=(S\odot E)_{ij}
        =S_{ij}E_{ij}
        =S_{ij}\sum_{t=1}^l x_i^{(t)}y_j^{(t)}
        =\sum_{t=1}^l x_i^{(t)}S_{ij}y_j^{(t)} =\sum_{t=1}^l \bigl(D_{{\bm x}^{(t)}} S D_{{\bm y}^{(t)}}\bigr)_{ij}.
\end{align}
Click to expand and view more

Therefore, $`\widehat{Z}=\sum_{t=1}^l D_{{\bm x}^{(t)}} S D_{{\bm y}^{(t)}}`$, which shows that $`\widehat{Z}`$ belongs to the right-hand side of Equation [eq:Fl-diag-form]. Suppose that $`\widehat{Z}=\sum_{t=1}^l D_{{\bm x}^{(t)}} S D_{{\bm y}^{(t)}}`$. Define $`E\coloneqq \sum_{t=1}^l {\bm x}^{(t)}\bigl({\bm y}^{(t)}\bigr)^\top`$. Then $`\mathrm{rank}(E)\le l`$. Moreover, for any $`(i,j)`$,

MATH
\begin{align}
        (S\odot E)_{ij}
        &=S_{ij}E_{ij}
        =S_{ij}\sum_{t=1}^l x_i^{(t)}y_j^{(t)}
        =\sum_{t=1}^l x_i^{(t)}S_{ij}y_j^{(t)}
        =\sum_{t=1}^l \bigl(D_{{\bm x}^{(t)}} S D_{{\bm y}^{(t)}}\bigr)_{ij}
        =\widehat{Z}_{ij}.
\end{align}
Click to expand and view more

Hence, $`\widehat{Z}=S\odot E\in{\mathcal F}_l(S)`$ completes the proof.Β β—»

Isometry Induced by Sign Mask

Lemma 5. *Let $`S\in\{\pm1\}^{p\times q}`$ and $`A,B\in{\mathbb R}^{p\times q}`$.

MATH
\begin{equation}
        \|S\odot A\|_{F}=\|A\|_{F},~~
        \|S\odot(A-B)\|_{F}=\|A-B\|_{F}.
\end{equation}
Click to expand and view more

The linear map $`T_S:{\mathbb R}^{p\times q}\to {\mathbb R}^{p\times q}`$ defined by $`T_S(Z)\coloneqq S\odot Z`$ is an isometry with respect to $`\|\cdot\|_{F}`$.*

Proof. Using the definition of the Frobenius norm and the fact that $`S_{ij}^2=1`$ for all $`(i,j)`$, we have

MATH
\begin{align}
        \|S\odot A\|_{F}^2
        &=\sum_{i=1}^{p}\sum_{j=1}^{q}(S_{ij}A_{ij})^2
        =\sum_{i=1}^{p}\sum_{j=1}^{q}S_{ij}^2\,A_{ij}^2
        =\sum_{i=1}^{p}\sum_{j=1}^{q}A_{ij}^2
        =\|A\|_{F}^{2}.
\end{align}
Click to expand and view more

Taking the square roots yields $`\|S\odot A\|_F=\|A\|_F`$. Applying the same argument to $`A`$-$`B`$ yields

MATH
\begin{equation}
        \|S\odot(A-B)\|_F=\|A-B\|_F,
\end{equation}
Click to expand and view more

which establishes the distance-preservation property and, consequently, the isometry claim.Β β—»

Optimal Demodulated-Envelope Approximation

We restate and prove TheoremΒ 2 from the perspective of the demodulated-envelope.

Theorem 6. *Let $`U\in {\mathbb R}^{N\times R}`$ fix a sign mask $`S\in\{\pm1\}^{N\times R}`$. Let $`\sigma_1(E_S(U))\ge\cdots\ge\sigma_{\min(N,R)}(E_S(U))\ge 0`$ denote the singular values of the demodulated envelope $`E_S(U)\coloneqq S\odot U`$. Fix an integer $`l`$ with $`1\le l\le \min(N,R)`$. Then

MATH
\begin{equation}
    \label{eq:rankl-error-app}
    \min_{\widehat{U}\in{\mathcal F}_l(S)} \|U-\widehat{U}\|_{F}^2
    =
    \sum_{i=l+1}^{\min(N, R)}\sigma_i\bigl(E_S(U)\bigr)^2.
\end{equation}
Click to expand and view more

Moreover, a minimizer is provided by

MATH
\begin{equation}
    \label{eq:rankl-opt-app}
    \widehat{U}^{\star}=S\odot \mathrm{TSVD}_l\bigl(E_S(U)\bigr),
\end{equation}
Click to expand and view more

where $`\mathrm{TSVD}_l(\cdot)`$ denotes the best rank-$`l`$ approximation in the Frobenius norm.*

Proof. By DefinitionΒ 1, $`\widehat{U}\in{\mathcal F}_l(S)`$ if and only if there exists a matrix $`E\in{\mathbb R}^{N\times R}`$ such that $`\mathrm{rank}(E)\le l`$ as follows:

MATH
\begin{equation}
        \label{eq:proof-param}
        \widehat{U}=S\odot E.
\end{equation}
Click to expand and view more

Hence, the constrained minimization over $`\widehat{U}\in{\mathcal F}_l(S)`$ is equivalent to minimizing over $`E`$ with $`\mathrm{rank}(E)\le l`$ under the parameterization in EquationΒ [eq:proof-param]. Define $`E_U\coloneqq E_S(U)=S\odot U`$. Since $`S_{ij}^2=1`$ for all $`(i,j)`$, we have the involution property $`S\odot(S\odot Z)=Z`$ for any $`Z`$, particularly $`U=S\odot E_U`$. For any feasible $`\widehat{U}=S\odot E`$, we obtain $`U-\widehat{U}=S\odot(E_U-E)`$. Applying LemmaΒ 5 yields

MATH
\begin{equation}
        \|U-\widehat{U}\|_F=\|E_U-E\|_F,~~
        \|U-\widehat{U}\|_F^2=\|E_U-E\|_F^2.
        \label{eq:reduce-rank-prob}
\end{equation}
Click to expand and view more

Substituting Equations [eq:proof-param] and [eq:reduce-rank-prob] gives

MATH
\begin{equation}
        \min_{\widehat{U}\in{\mathcal F}_l(S)} \|U-\widehat{U}\|_{F}^2
        =
        \min_{\substack{E\in{\mathbb R}^{N\times R} \\ \mathrm{rank}(E)\le l}}
        \|E_U-E\|_{F}^2,
\end{equation}
Click to expand and view more

which is exactly the classical best rank-$`l`$ approximation problem for $`E_U`$ in Frobenius norm. By the Eckart–Young–Mirsky theorem applied to $`E_U`$, we have

MATH
\begin{equation}
        \min_{\mathrm{rank}(E)\le l}\|E_U-E\|_{F}^2 =
        \sum_{i=l+1}^{\min(N,R)}\sigma_i(E_U)^2 =
        \sum_{i=l+1}^{\min(N,R)}\sigma_i\bigl(E_S(U)\bigr)^2,
\end{equation}
Click to expand and view more

which proves Equation [eq:rankl-error-app]. Moreover, the same theorem implies that an optimizer is $`E^\star=\mathrm{TSVD}_l(E_U)`$; substituting into Equation [eq:proof-param] yields $`\widehat{U}^\star=S\odot \mathrm{TSVD}_l(E_U)=S\odot \mathrm{TSVD}_l(E_S(U))`$, which corresponds to Equation [eq:rankl-opt-app].Β β—»

Remark 7. If we choose the mask $`S=\mathrm{sign}(U)`$, the demodulated envelope satisfies $`E_S(U)=S\odot U=|U|`$ entrywise. In this case, Equation [eq:rankl-error-app] specializes to the standard singular-value tail characterization of the best rank-$`l`$ approximation error for $`|U|`$ in the Frobenius norm.

Corollary 8. *Under the assumptions of TheoremΒ 6, setting $`l=1`$ yields

MATH
\begin{equation}
        \min_{\widehat{U}\in{\mathcal F}_1(S)} \|U-\widehat{U}\|_{F}^2
        =
        \sum_{i=2}^{\min(N,M)}\sigma_i\bigl(E_S(U)\bigr)^2,.
\end{equation}
Click to expand and view more

In particular, the single envelope constraint is lossless (in Frobenius norm) if and only if $`\mathrm{rank}\bigl(E_S(U)\bigr)\le 1`$.*

Proof. This is the specialization of TheoremΒ 6 to $`l=1`$.Β β—»

Exact Projection

Proposition 9. *Fix $`S\in\{\pm1\}^{N\times R}`$ and $`l\ge 1`$. For any $`Z\in{\mathbb R}^{N\times R}`$, the Frobenius projection onto $`{\mathcal F}_l(S)`$,

MATH
\begin{equation}
        \Pi_{{\mathcal F}_l(S)}(Z)\in \arg\min_{Y\in{\mathcal F}_l(S)}\|Z-Y\|_F,
\end{equation}
Click to expand and view more

selected as

MATH
\begin{equation}
        \label{eq:proj-Fl-app}
        \Pi_{{\mathcal F}_l(S)}(Z) = S\odot \mathrm{TSVD}_l\bigl(E_S(Z)\bigr) = S\odot \mathrm{TSVD}_l\bigl(S\odot Z\bigr).
\end{equation}
```*

</div>

<div class="proof">

*Proof.* By DefinitionΒ <a href="#def:Fl-app" data-reference-type="ref"
data-reference="def:Fl-app">3</a>, any $`Y\in{\mathcal F}_l(S)`$ can be
expressed as $`Y=S\odot E`$ for some $`E\in{\mathbb R}^{N\times R}`$
with $`\mathrm{rank}(E)\le l`$. Hence,
``` math
\begin{align}
        \min_{Y\in{\mathcal F}_l(S)}\|Z-Y\|_F^2
        &= \min_{\substack{E\in{\mathbb R}^{N\times R}\\ \mathrm{rank}(E)\le l}} \|Z-S\odot E\|_F^2.
\end{align}
Click to expand and view more

Using the sign-mask isometry (LemmaΒ 5) and the identity $`S\odot(S\odot E)=E`$, we obtain

MATH
\begin{align}
        \|Z-S\odot E\|_F
        &=\|S\odot Z - S\odot(S\odot E)\|_F
        =\|S\odot Z - E\|_F
        =\|E_S(Z)-E\|_F.
\end{align}
Click to expand and view more

Therefore,

MATH
\begin{equation}
        \min_{Y\in{\mathcal F}_l(S)}\|Z-Y\|_F^2
        =
        \min_{\mathrm{rank}(E)\le l}\|E_S(Z)-E\|_F^2.
\end{equation}
Click to expand and view more

The right-hand side represents the classical best rank-$`l`$ approximation problem for $`E_S(Z)`$ in the Frobenius norm, with minimizers provided by the truncated SVD: $`E^\star=\mathrm{TSVD}_l(E_S(Z))`$ (Eckart–Young–Mirsky theorem). Substituting back yields $`Y^\star=S\odot E^\star=S\odot \mathrm{TSVD}_l (E_S(Z))`$, which is Equation [eq:proj-Fl-app].Β β—»

Additional Implementation Details

Entropy-based effective rank

We quantify the number of envelope modes present using the entropy-based effective rank (entropy rank) . For a matrix $`E\in{\mathbb R}^{N \times M}`$ (in our case, a demodulated envelope such as $`E_{\mathrm{sign}(U)}(U)=|U|`$ or $`E_{\mathrm{sign}(V)}(V)=|V|`$), let singular values and $`\sigma_1(E)\ge \cdots \ge \sigma_{\min(N,M)}(E)\ge 0`$ represent its normalized spectrum. Define the normalized spectrum

MATH
\begin{equation}
    \pi_i(E)\coloneqq \frac{\sigma_i(E)}{\sum_{j=1}^{\min(N,M)}\sigma_j(E)},~~~ i=1,\dots,\min(N,M),
\end{equation}
Click to expand and view more

with the convention $`\pi_i(E)=1/\min(N,M)`$, if $`E=0`$ and the spectral entropy are considered:

MATH
\begin{equation}
    H(E)\coloneqq -\sum_{i=1}^{\min(N,M)}\pi_i(E)\log \pi_i(E).
\end{equation}
Click to expand and view more

The effective rank of entropy is $`\mathrm{erank}(E)\coloneqq \exp\bigl(H(E)\bigr)`$. This quantity is scale-invariant and satisfies $`1\le \mathrm{erank}(E)\le \min(N,M)`$. Moreover, $`\mathrm{erank}(E)=1`$ when $`E`$ is a nonzero rank-one matrix with all spectral mass on a single singular value, and $`\mathrm{erank}(E)\approx k`$ when roughly $`k`$ singular values contribute comparably.

Additional Experiments

In the main text, we present results aggregated across models (and, where applicable, across tasks) to emphasize overall trends. For completeness and transparency, this appendix includes the individual results for each model and target bitwidth. All experiments follow the same evaluation protocol as the main paper; the reported metric remains the same normalized accuracy (higher is better).

TablesΒ [tab:arc_easy_acc_norm_bits_by_lP] andΒ [tab:piqa_acc_norm_bits_by_lP] summarize normalized accuracy on ARC-Easy and PIQA, respectively. Rows indicate the target bit width, while columns correspond to the bit-parameter choices $`(l,P)`$. Overall, the per-model numbers align with the aggregate results in the main text: changing $`(l,P)`$ typically results in modest variations, with $`(l{=}2,P{=}1)`$ providing a strong and stable default across models, while increasing $`l`$ often leads to diminishing returns.

Model Bits (l = 1, P = 1) (l = 1, P = 2) (l = 2, P = 1) (l = 8, P = 1) (l = 16, P = 1)
Tiny LLaMA (1.1B) 1.00 0.3279 0.3258 0.3215 0.3304 0.3283
1.25 0.3737 0.3384 0.3960 0.3822 0.3927
1.50 0.4057 0.3838 0.4045 0.4066 0.4082
Qwen3 (0.6B) 1.00 0.3051 0.2883 0.3157 0.2837 0.3064
1.25 0.3258 0.3089 0.3283 0.3338 0.3245
1.50 0.3418 0.3140 0.3565 0.3401 0.3485
LLaMA3.2 (1B) 1.00 0.3413 0.3653 0.3359 0.3523 0.3472
1.25 0.3843 0.3540 0.3897 0.3956 0.3990
1.50 0.3906 0.4217 0.4112 0.4369 0.4082
LLaMA2 (7B) 1.00 0.4221 0.3842 0.4209 0.4228 0.4265
1.25 0.5118 0.4482 0.5189 0.4853 0.4899
1.50 0.5614 0.5240 0.5800 0.5370 0.5631
LLaMA3 (8B) 1.00 0.2879 0.2761 0.3001 0.3077 0.2946
1.25 0.3628 0.3266 0.3817 0.3544 0.3232
1.50 0.4099 0.4024 0.3965 0.3864 0.3851
Qwen3 (8B) 1.00 0.3514 0.3439 0.3443 0.3481 0.3514
1.25 0.4045 0.3645 0.4007 0.3788 0.3805
1.50 0.4196 0.3885 0.4415 0.4373 0.4188
Model Bits (l = 1, P = 1) (l = 1, P = 2) (l = 2, P = 1) (l = 8, P = 1) (l = 16, P = 1)
Tiny LLaMA (1.1B) 1.00 0.5517 0.5452 0.5631 0.5811 0.5615
1.25 0.6153 0.5832 0.6137 0.6268 0.6431
1.50 0.6447 0.6268 0.6502 0.6556 0.6436
Qwen3 (0.6B) 1.00 0.5424 0.5337 0.5332 0.5403 0.5392
1.25 0.5522 0.5337 0.5615 0.5560 0.5669
1.50 0.5620 0.5495 0.5642 0.5686 0.5734
LLaMA3.2 (1B) 1.00 0.5865 0.5778 0.6023 0.5838 0.5909
1.25 0.6110 0.6143 0.6181 0.6436 0.6208
1.50 0.6257 0.6186 0.6333 0.6404 0.6398
LLaMA2 (7B) 1.00 0.6306 0.5997 0.6225 0.6230 0.6192
1.25 0.6844 0.6561 0.7002 0.6681 0.6670
1.50 0.7182 0.6980 0.7160 0.7116 0.7051
LLaMA3 (8B) 1.00 0.5277 0.5234 0.5207 0.5294 0.5245
1.25 0.5963 0.5473 0.6192 0.5887 0.5653
1.50 0.6317 0.6088 0.6153 0.6273 0.6148
Qwen3 (8B) 1.00 0.5756 0.5577 0.5734 0.5490 0.5664
1.25 0.5898 0.5647 0.5979 0.5849 0.5903
1.50 0.6175 0.5876 0.5903 0.6153 0.6050

πŸ“Š λ…Όλ¬Έ μ‹œκ°μžλ£Œ (Figures)

Figure 1



Figure 2



A Note of Gratitude

The copyright of this content belongs to the respective researchers. We deeply appreciate their hard work and contribution to the advancement of human civilization.

Start searching

Enter keywords to search articles

↑↓
↡
ESC
⌘K Shortcut