Attention Editing: A Versatile Framework for Cross-Architecture Attention Conversion
arXiv:2604.05688v1 Announce Type: new Abstract: Key-Value (KV) cache memory and bandwidth increasingly dominate large language model inference cost in long-context and long-generation regimes. Architectures such …
Zhen Cheng, Hao-Bo Yang, Wan-Yi Huang, Jin-Long Li
5 views