LLM-Enhanced Chinese Morph Resolution in E-Commerce Live Streaming Scenarios

E-commerce live streaming in China has become a major retail channel, yet hosts often employ subtle phonetic or semantic “morphs” to evade moderation and make unsubstantiated claims, posing risks to consumers. To address this, we study the Live Auditory Morph Resolution (LiveAMR) task, which restore...

Full description

Saved in:
Bibliographic Details
Main Authors: Xiaoye Ouyang, Liu Yuan, Xiaocheng Hu, Jiahao Zhu, Jipeng Qiang
Format: Article
Language:English
Published: MDPI AG 2025-06-01
Series:Entropy
Subjects:
Online Access:https://www.mdpi.com/1099-4300/27/7/698
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:E-commerce live streaming in China has become a major retail channel, yet hosts often employ subtle phonetic or semantic “morphs” to evade moderation and make unsubstantiated claims, posing risks to consumers. To address this, we study the Live Auditory Morph Resolution (LiveAMR) task, which restores morphed speech transcriptions to their true forms. Building on prior text-based morph resolution, we propose an LLM-enhanced training framework that mines three types of explanation knowledge—predefined morph-type labels, LLM-generated reference corrections, and natural-language rationales constrained for clarity and comprehensiveness—from a frozen large language model. These annotations are concatenated with the original morphed sentence and used to fine-tune a lightweight T5 model under a standard cross-entropy objective. In experiments on two test sets (in-domain and out-of-domain), our method achieves substantial gains over baselines, improving <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><msub><mi>F</mi><mrow><mn>0.5</mn></mrow></msub></semantics></math></inline-formula> by up to 7 pp in-domain (to 0.943) and 5 pp out-of-domain (to 0.799) compared to a strong T5 baseline. These results demonstrate that structured LLM-derived signals can be mined without fine-tuning the LLM itself and injected into small models to yield efficient, accurate morph resolution.
ISSN:1099-4300