A hybrid model based on transformer and Mamba for enhanced sequence modeling

Abstract State Space Models (SSMs) have made remarkable strides in language modeling in recent years. With the introduction of Mamba, these models have garnered increased attention, often surpassing Transformers in specific areas. Nevertheless, despite Mamba’s unique strengths, Transformers remain e...

Full description

Saved in:
Bibliographic Details
Main Authors: Xiaocui Zhu, Qunsheng Ruan, Sai Qian, Miaohui Zhang
Format: Article
Language:English
Published: Nature Portfolio 2025-04-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-87574-8
Tags: Add Tag
No Tags, Be the first to tag this record!