Highly Efficient and Effective LLMs with Multi-Boolean Architectures

arXiv stat.ML / 4/22/2026

💬 OpinionDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper addresses the challenge that naive post-training weight binarization for LLMs often causes large performance drops, while training-aware binarization can be complex and inefficient.
  • It proposes a framework that represents LLMs using multi-kernel Boolean parameters, enabling direct fine-tuning in the Boolean domain without relying on full-precision latent weights.
  • By removing the need for latent weights, the approach aims to reduce model complexity both during fine-tuning and at inference time.
  • Experiments across multiple LLMs indicate the method outperforms recent ultra low-bit quantization and binarization techniques.
  • Overall, the work suggests a new direction for efficiently adapting and deploying LLMs using Boolean-domain training with improved effectiveness.

Abstract

Weight binarization has emerged as a promising strategy to reduce the complexity of large language models (LLMs). Existing approaches fall into post-training binarization, which is simple but causes severe performance loss, and training-aware methods, which depend on full-precision latent weights, adding complexity and limiting efficiency. We propose a novel framework that represents LLMs with multi-kernel Boolean parameters and, for the first time, enables direct finetuning LMMs in the Boolean domain, eliminating the need for latent weights. This enhances representational capacity and dramatically reduces complexity during both finetuning and inference. Extensive experiments across diverse LLMs show our method outperforms recent ultra low-bit quantization and binarization techniques.