预印本 / 版本 1

LanguageFold: A Bio-inspired Hierarchical Sparse Attention Mechanism for Large Language Models

本文是预印本,尚未经过同行评审认证。

作者

分类
关键词
Biology4AI; natural language processing; large language model; attention mechansim

摘要

Large language models predominantly rely on the Transformer architecture, whose self-attention mechanism incurs a quadratic computational cost O(N2) with respect to input length, leading to significant memory and computation bottlenecks when processing ultra-long contexts. This work proposes LanguageFold, a hierarchical sparse attention mechanism inspired by the Self-Returning Random Walk model of genome folding (Huang et al. 2020). LanguageFold decomposes global attention into dynamically constructed tree attention with a theoretical scaling of O(NlogN). Preliminary experiments on prompt-based generation and the DROP reading comprehension benchmark indicate that this tree-structured attention enables efficient language processing while preserving accuracy and enhancing structural interpretability. These results highlight the promise of genome-inspired attention mechanisms for optimizing the scalability of large language models.

指标

收藏: 2
查看次数: 201
下载次数: 9

下载次数

已发布

2026-01-29

如何引用

Wu, Y., & Huang, K. (2026). LanguageFold: A Bio-inspired Hierarchical Sparse Attention Mechanism for Large Language Models. 浪淘沙预印本平台. https://doi.org/10.65215/LTSpreprints.2026.01.28.000108

利益冲突声明

作者声明无任何需要披露的利益冲突。