Efficient LLMs Training and Inference: An Introduction
ChatGPT was released in late November 2022, making a significant impact globally. Following this release, numerous domestic and international open-source projects for large model training emerged, including Alpaca, BOOLM, LLaMA, ChatGLM, DeepSpeedChat, and ColossalChat. Both academia and industry ha...
Saved in:
| Main Authors: | Rui Li, Deji Fu, Chunyu Shi, Zhilan Huang, Gang Lu |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10756602/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Addressing Activation Outliers in LLMs: A Systematic Review of Post-Training Quantization Techniques
by: Patrik Czako, et al.
Published: (2025-01-01) -
Long-context inference optimization for large language models: a survey
by: TAO Wei, et al.
Published: (2025-01-01) -
AsymGroup: Asymmetric Grouping and Communication Optimization for 2D Tensor Parallelism in LLM Inference
by: Ki Tae Kim, et al.
Published: (2025-01-01) -
Hive: A secure, scalable framework for distributed Ollama inference
by: Domen Vake, et al.
Published: (2025-05-01) -
ResDecode: Accelerating Large Language Models Inference via Residual Decoding Heads
by: Ziqian Zeng, et al.
Published: (2025-06-01)