Abstract
A novel attention mechanism, Adaptive Filter Attention, uses a learnable dynamics model to compute attention weights, incorporating pairwise uncertainties and simplifying to dot-product attention under certain conditions.
We introduce Adaptive Filter Attention (AFA), a novel attention mechanism that incorporates a learnable dynamics model directly into the computation of attention weights. Rather than comparing queries and keys directly, we model the input sequence as discrete observations of a linear stochastic differential equation (SDE). By imposing a linear dynamics model with simultaneously diagonalizable state matrices and noise covariances, we can make use of a closed-form solution to the differential Lyapunov equation to efficiently propagate pairwise uncertainties through the dynamics. Attention naturally arises as the maximum likelihood solution for this linear SDE, with attention weights corresponding to robust residual-based reweightings of the propagated pairwise precisions. Imposing an additional constraint on the state matrix's eigenvalues leads to a simplified variant with the same computational and memory complexity as standard attention. In the limit of vanishing dynamics and process noise, and using a small-angle approximation, we recover ordinary dot-product attention.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- WeightFlow: Learning Stochastic Dynamics via Evolving Weight of Neural Network (2025)
- Universal Learning of Nonlinear Dynamics (2025)
- Efficient Attention Mechanisms for Large Language Models: A Survey (2025)
- Differentiable Adaptive Kalman Filtering via Optimal Transport (2025)
- Transition Models: Rethinking the Generative Learning Objective (2025)
- Uncertainty Propagation Networks for Neural Ordinary Differential Equations (2025)
- Flow Matching-Based Generative Modeling for Efficient and Scalable Data Assimilation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper