attention
PulseAugur coverage of attention — every cluster mentioning attention across labs, papers, and developer communities, ranked by signal.
-
Rhamba framework integrates attention and Mamba for fMRI self-supervised learning
Researchers have developed Rhamba, a novel framework for self-supervised learning on resting-state fMRI data. This framework combines region-aware masking with hybrid Attention-Mamba architectures to improve the analysi…
-
Switch Attention dynamically routes between full and sliding window attention
Researchers have introduced Switch Attention (SwiAttn), a novel hybrid transformer architecture designed to address the computational bottleneck of standard full attention mechanisms in long-context language modeling. S…
-
Beyond Linearity in Attention Projections: The Case for Nonlinear Queries
Researchers are exploring the fundamental mechanisms behind transformer attention, with new papers analyzing its gradient flow structure and dynamics. One study interprets attention as a gradient flow on a unit sphere, …
-
Eugene Yan shares guide to running weekly AI paper club for learning communities
Eugene Yan details a successful weekly paper club that has met for 18 months, discussing at least 80 AI-related papers. The club focuses on foundational concepts, models, training, and inference techniques within machin…
-
Mamba model offers Transformer-level performance with faster inference and longer context
Mamba, a new State Space Model (SSM), presents an alternative to the dominant Transformer architecture in AI. It aims to match Transformer performance and scaling laws while efficiently handling extremely long sequences…