MHLA: Restoring Expressivity of Linear Attention via Token-Level Multi-Head Paper ⢠2601.07832 ⢠Published 20 days ago ⢠51
mzwing/Seikaijyu_RWKV7-2.9B-v3-UnlimitedRP-mini-novel-chat-preview-GGUF 3B ⢠Updated Apr 4, 2025 ⢠520 ⢠2