How social reinforcement learning can lead to metastable polarisation and the voter model.

Previous explanations for the persistence of polarization of opinions have typically included modelling assumptions that predispose the possibility of polarization (i.e., assumptions allowing a pair of agents to drift apart in their opinion such as repulsive interactions or bounded confidence). An e...

Full description

Saved in:
Bibliographic Details
Main Authors: Benedikt V Meylahn, Janusz M Meylahn
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2024-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0313951
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849772971343740928
author Benedikt V Meylahn
Janusz M Meylahn
author_facet Benedikt V Meylahn
Janusz M Meylahn
author_sort Benedikt V Meylahn
collection DOAJ
description Previous explanations for the persistence of polarization of opinions have typically included modelling assumptions that predispose the possibility of polarization (i.e., assumptions allowing a pair of agents to drift apart in their opinion such as repulsive interactions or bounded confidence). An exception is a recent simulation study showing that polarization is persistent when agents form their opinions using social reinforcement learning. Our goal is to highlight the usefulness of reinforcement learning in the context of modeling opinion dynamics, but that caution is required when selecting the tools used to study such a model. We show that the polarization observed in the model of the simulation study cannot persist indefinitely, and exhibits consensus asymptotically with probability one. By constructing a link between the reinforcement learning model and the voter model, we argue that the observed polarization is metastable. Finally, we show that a slight modification in the learning process of the agents changes the model from being non-ergodic to being ergodic. Our results show that reinforcement learning may be a powerful method for modelling polarization in opinion dynamics, but that the tools (objects to study such as the stationary distribution, or time to absorption for example) appropriate for analysing such models crucially depend on their properties (such as ergodicity, or transience). These properties are determined by the details of the learning process and may be difficult to identify based solely on simulations.
format Article
id doaj-art-47b38f98bdfb43d485d7acb6ab06d755
institution DOAJ
issn 1932-6203
language English
publishDate 2024-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj-art-47b38f98bdfb43d485d7acb6ab06d7552025-08-20T03:02:11ZengPublic Library of Science (PLoS)PLoS ONE1932-62032024-01-011912e031395110.1371/journal.pone.0313951How social reinforcement learning can lead to metastable polarisation and the voter model.Benedikt V MeylahnJanusz M MeylahnPrevious explanations for the persistence of polarization of opinions have typically included modelling assumptions that predispose the possibility of polarization (i.e., assumptions allowing a pair of agents to drift apart in their opinion such as repulsive interactions or bounded confidence). An exception is a recent simulation study showing that polarization is persistent when agents form their opinions using social reinforcement learning. Our goal is to highlight the usefulness of reinforcement learning in the context of modeling opinion dynamics, but that caution is required when selecting the tools used to study such a model. We show that the polarization observed in the model of the simulation study cannot persist indefinitely, and exhibits consensus asymptotically with probability one. By constructing a link between the reinforcement learning model and the voter model, we argue that the observed polarization is metastable. Finally, we show that a slight modification in the learning process of the agents changes the model from being non-ergodic to being ergodic. Our results show that reinforcement learning may be a powerful method for modelling polarization in opinion dynamics, but that the tools (objects to study such as the stationary distribution, or time to absorption for example) appropriate for analysing such models crucially depend on their properties (such as ergodicity, or transience). These properties are determined by the details of the learning process and may be difficult to identify based solely on simulations.https://doi.org/10.1371/journal.pone.0313951
spellingShingle Benedikt V Meylahn
Janusz M Meylahn
How social reinforcement learning can lead to metastable polarisation and the voter model.
PLoS ONE
title How social reinforcement learning can lead to metastable polarisation and the voter model.
title_full How social reinforcement learning can lead to metastable polarisation and the voter model.
title_fullStr How social reinforcement learning can lead to metastable polarisation and the voter model.
title_full_unstemmed How social reinforcement learning can lead to metastable polarisation and the voter model.
title_short How social reinforcement learning can lead to metastable polarisation and the voter model.
title_sort how social reinforcement learning can lead to metastable polarisation and the voter model
url https://doi.org/10.1371/journal.pone.0313951
work_keys_str_mv AT benediktvmeylahn howsocialreinforcementlearningcanleadtometastablepolarisationandthevotermodel
AT januszmmeylahn howsocialreinforcementlearningcanleadtometastablepolarisationandthevotermodel