A benchmark for evaluating crisis information generation capabilities in LLMs
Introduction. Large language models (LLMs) have become increasingly significant in crisis information management due to their advanced natural language processing capabilities. This study aims to develop a comprehensive evaluation benchmark to assess the effectiveness of LLMs in generating crisis i...
Saved in:
| Main Authors: | Ruilian Han, Lu An, Wei Zhou, Gang Li |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
University of Borås
2025-03-01
|
| Series: | Information Research: An International Electronic Journal |
| Subjects: | |
| Online Access: | https://publicera.kb.se/ir/article/view/47518 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
RVBench: Role values benchmark for role-playing LLMs
by: Ye Wang, et al.
Published: (2025-08-01) -
Systematic Analysis of Retrieval-Augmented Generation-Based LLMs for Medical Chatbot Applications
by: Arunabh Bora, et al.
Published: (2024-10-01) -
TSTBench: A Comprehensive Benchmark for Text Style Transfer
by: Yifei Xie, et al.
Published: (2025-05-01) -
LLMs in the Generation of Seismic Alert Communiqués
by: Oscar Peña-Cáceres, et al.
Published: (2025-05-01) -
Measuring and Improving the Efficiency of Python Code Generated by LLMs Using CoT Prompting and Fine-Tuning
by: Ramya Jonnala, et al.
Published: (2025-01-01)