A comparative analysis of encoder only and decoder only models for challenging LLM-generated STEM MCQs using a self-evaluation approach
Large Language Models (LLMs) have demonstrated impressive capabilities in various tasks, including Multiple-Choice Question Answering (MCQA) evaluated on benchmark datasets with few-shot prompting. Given the absence of benchmark Science, Technology, Engineering, and Mathematics (STEM) datasets on Mu...
Saved in:
| Main Authors: | Ghada Soliman, Ph.D., Hozaifa Zaki, Mohamed Kilany |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Elsevier
2025-03-01
|
| Series: | Natural Language Processing Journal |
| Subjects: | |
| Online Access: | http://www.sciencedirect.com/science/article/pii/S294971912500007X |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Exploration of the Efficiency of SLM-Enabled Platforms for Everyday Tasks
by: Volodymyr Rusinov, et al.
Published: (2025-04-01) -
Effects of a long term faculty development program on improvement in quality of MCQs: an impact evaluation study
by: Rukhsana Ayub, et al.
Published: (2025-04-01) -
LLM-Based Doppelgänger Models: Leveraging Synthetic Data for Human-Like Responses in Survey Simulations
by: Suhyun Cho, et al.
Published: (2024-01-01) -
Comparative analysis of NLP-driven MCQ generators from text sources
by: Asmae Azzi, et al.
Published: (2025-12-01) -
COLLUSION DETECTION SOFTWARE IN ONLINE MULTIPLE CHOICE EXAMINATIONS – A REVIEW
by: Srinivasa Jayachandra
Published: (2022-12-01)