Local Scale Equivariance with Latent Deep Equilibrium Canonicalizer
Abstract
A deep equilibrium canonicalizer (DEC) enhances local scale equivariance in deep networks, improving performance and consistency on ImageNet.
Scale variation is a fundamental challenge in computer vision. Objects of the same class can have different sizes, and their perceived size is further affected by the distance from the camera. These variations are local to the objects, i.e., different object sizes may change differently within the same image. To effectively handle scale variations, we present a deep equilibrium canonicalizer (DEC) to improve the local scale equivariance of a model. DEC can be easily incorporated into existing network architectures and can be adapted to a pre-trained model. Notably, we show that on the competitive ImageNet benchmark, DEC improves both model performance and local scale consistency across four popular pre-trained deep-nets, e.g., ViT, DeiT, Swin, and BEiT. Our code is available at https://github.com/ashiq24/local-scale-equivariance.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- CLIPSym: Delving into Symmetry Detection with CLIP (2025)
- Test-Time Canonicalization by Foundation Models for Robust Perception (2025)
- π3: Scalable Permutation-Equivariant Visual Geometry Learning (2025)
- Rotation Equivariant Arbitrary-scale Image Super-Resolution. (2025)
- Neural Tangent Knowledge Distillation for Optical Convolutional Networks (2025)
- Extending Foundational Monocular Depth Estimators to Fisheye Cameras with Calibration Tokens (2025)
- L-SR1: Learned Symmetric-Rank-One Preconditioning (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper