Multilingual capabilities of GPT: A study of structural ambiguity.

This study examines the multilingual capabilities of GPT, focusing on its handling of syntactic ambiguity across English, Korean, and Japanese. We investigate whether GPT can capture language-specific attachment preferences or if it relies primarily on English-centric training patterns. Using ambigu...

Full description

Saved in:
Bibliographic Details
Main Authors: Myung Hye Yoo, Joungmin Kim, Sanghoun Song
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2025-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0326943
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850117572582703104
author Myung Hye Yoo
Joungmin Kim
Sanghoun Song
author_facet Myung Hye Yoo
Joungmin Kim
Sanghoun Song
author_sort Myung Hye Yoo
collection DOAJ
description This study examines the multilingual capabilities of GPT, focusing on its handling of syntactic ambiguity across English, Korean, and Japanese. We investigate whether GPT can capture language-specific attachment preferences or if it relies primarily on English-centric training patterns. Using ambiguous relative clauses as a testing ground, we assess GPT's interpretation tendencies across language contexts. Our findings reveal that, while GPT (GPT-3.5-turbo, GPT-4-turbo, GPT 4o)'s performance aligns with native English speakers' preferred interpretations, it overgeneralizes this interpretation in Korean and lacks clear preferences in Japanese, despite distinct attachment biases among native speakers of these languages. The newer, smaller-scale models-o1-mini and o3-mini-further reinforce this trend by closely mirroring English attachment patterns in both Korean and Japanese. Overall results suggest that GPT's multilingual proficiency is limited, likely reflecting a bias toward high-resource languages like English, although differences in model size and tuning strategies may partially mitigate the extent of English-centric generalization. While GPT models demonstrate aspects of human-like language processing, our findings underscore the need for further refinement to achieve a more nuanced engagement with linguistic diversity across languages.
format Article
id doaj-art-87ef9bbacf2e4dd18b916f1a36a7480c
institution OA Journals
issn 1932-6203
language English
publishDate 2025-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj-art-87ef9bbacf2e4dd18b916f1a36a7480c2025-08-20T02:36:04ZengPublic Library of Science (PLoS)PLoS ONE1932-62032025-01-01207e032694310.1371/journal.pone.0326943Multilingual capabilities of GPT: A study of structural ambiguity.Myung Hye YooJoungmin KimSanghoun SongThis study examines the multilingual capabilities of GPT, focusing on its handling of syntactic ambiguity across English, Korean, and Japanese. We investigate whether GPT can capture language-specific attachment preferences or if it relies primarily on English-centric training patterns. Using ambiguous relative clauses as a testing ground, we assess GPT's interpretation tendencies across language contexts. Our findings reveal that, while GPT (GPT-3.5-turbo, GPT-4-turbo, GPT 4o)'s performance aligns with native English speakers' preferred interpretations, it overgeneralizes this interpretation in Korean and lacks clear preferences in Japanese, despite distinct attachment biases among native speakers of these languages. The newer, smaller-scale models-o1-mini and o3-mini-further reinforce this trend by closely mirroring English attachment patterns in both Korean and Japanese. Overall results suggest that GPT's multilingual proficiency is limited, likely reflecting a bias toward high-resource languages like English, although differences in model size and tuning strategies may partially mitigate the extent of English-centric generalization. While GPT models demonstrate aspects of human-like language processing, our findings underscore the need for further refinement to achieve a more nuanced engagement with linguistic diversity across languages.https://doi.org/10.1371/journal.pone.0326943
spellingShingle Myung Hye Yoo
Joungmin Kim
Sanghoun Song
Multilingual capabilities of GPT: A study of structural ambiguity.
PLoS ONE
title Multilingual capabilities of GPT: A study of structural ambiguity.
title_full Multilingual capabilities of GPT: A study of structural ambiguity.
title_fullStr Multilingual capabilities of GPT: A study of structural ambiguity.
title_full_unstemmed Multilingual capabilities of GPT: A study of structural ambiguity.
title_short Multilingual capabilities of GPT: A study of structural ambiguity.
title_sort multilingual capabilities of gpt a study of structural ambiguity
url https://doi.org/10.1371/journal.pone.0326943
work_keys_str_mv AT myunghyeyoo multilingualcapabilitiesofgptastudyofstructuralambiguity
AT joungminkim multilingualcapabilitiesofgptastudyofstructuralambiguity
AT sanghounsong multilingualcapabilitiesofgptastudyofstructuralambiguity