Addressing Activation Outliers in LLMs: A Systematic Review of Post-Training Quantization Techniques
Large Language Models (LLMs) have transformed natural language processing, yet their deployment remains challenging due to substantial computational, memory, and energy demands. Post-training quantization has emerged as a key strategy for enabling efficient inference, particularly in resource-constr...
Saved in:
| Main Authors: | Patrik Czako, Gabor Kertesz, Sandor Szenasi |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10994764/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Efficient Deep Learning Model Compression for Sensor-Based Vision Systems via Outlier-Aware Quantization
by: Joonhyuk Yoo, et al.
Published: (2025-05-01) -
Mitigating Quantization Errors Due to Activation Spikes in Gated Linear Unit-Based Large Language Models
by: Jaewoo Yang, et al.
Published: (2025-04-01) -
Systematic Analysis of Retrieval-Augmented Generation-Based LLMs for Medical Chatbot Applications
by: Arunabh Bora, et al.
Published: (2024-10-01) -
Efficient LLMs Training and Inference: An Introduction
by: Rui Li, et al.
Published: (2025-01-01) -
ResDecode: Accelerating Large Language Models Inference via Residual Decoding Heads
by: Ziqian Zeng, et al.
Published: (2025-06-01)