Leveraging two-dimensional pre-trained vision transformers for three-dimensional model generation via masked autoencoders
Abstract Although the Transformer architecture has established itself as the industry standard for jobs involving natural language processing, it still has few uses in computer vision. In vision, attention is used in conjunction with convolutional networks or to replace individual convolutional netw...
Saved in:
Main Authors: | Muhammad Sajid, Kaleem Razzaq Malik, Ateeq Ur Rehman, Tauqeer Safdar Malik, Masoud Alajmi, Ali Haider Khan, Amir Haider, Seada Hussen |
---|---|
Format: | Article |
Language: | English |
Published: |
Nature Portfolio
2025-01-01
|
Series: | Scientific Reports |
Subjects: | |
Online Access: | https://doi.org/10.1038/s41598-025-87376-y |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Two-Dimensional Ferroelectric Materials: From Prediction to Applications
by: Shujuan Jiang, et al.
Published: (2025-01-01) -
Blink Detection Using 3D Convolutional Neural Architectures and Analysis of Accumulated Frame Predictions
by: George Nousias, et al.
Published: (2025-01-01) -
A novel system for markerless intra-operative bone and bone fragment tracking
by: Marcus Tatum, et al.
Published: (2025-12-01) -
Two- and Three-Dimensional Culture Systems: Respiratory In Vitro Tissue Models for Chemical Screening and Risk-Based Decision Making
by: Joanne Wallace, et al.
Published: (2025-01-01) -
Latent space improved masked reconstruction model for human skeleton-based action recognition
by: Enqing Chen, et al.
Published: (2025-02-01)