LNLF-BERT: Transformer for Long Document Classification With Multiple Attention Levels

Transformer-based models, such as Bidirectional Encoder Representations from Transformers (BERT), cannot process long sequences because their self-attention operation scales quadratically with the sequence length. To remedy this, we introduce the Look Near and Look Far BERT (LNLF-BERT) with a two-le...

Full description

Saved in:
Bibliographic Details
Main Authors: Linh Manh Pham, Hoang Cao the
Format: Article
Language:English
Published: IEEE 2024-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10744540/
Tags: Add Tag
No Tags, Be the first to tag this record!