Vision Transformers for Image Classification: A Comparative Survey
Transformers were initially introduced for natural language processing, leveraging the self-attention mechanism. They require minimal inductive biases in their design and can function effectively as set-based architectures. Additionally, transformers excel at capturing long-range dependencies and en...
Saved in:
Main Authors: | Yaoli Wang, Yaojun Deng, Yuanjin Zheng, Pratik Chattopadhyay, Lipo Wang |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2025-01-01
|
Series: | Technologies |
Subjects: | |
Online Access: | https://www.mdpi.com/2227-7080/13/1/32 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
ADAPTIVE VISION AI
by: V. Vodyanitskyi, et al.
Published: (2024-12-01) -
3D landmark‐based face restoration for recognition using variational autoencoder and triplet loss
by: Sahil Sharma, et al.
Published: (2021-01-01) -
Editorial: Computer vision and image synthesis for neurological applications
by: Chenglong Zou, et al.
Published: (2025-02-01) -
Assessing bias and computational efficiency in vision transformers using early exits
by: Seth Nixon, et al.
Published: (2025-01-01) -
Classroom Behavior Recognition Using Computer Vision: A Systematic Review
by: Qingtang Liu, et al.
Published: (2025-01-01)