• Aucun résultat trouvé

Introducing the Argumentation Framework Within Agent-Based Models to Better Simulate Agents' Cognition in Opinion Dynamics: Application to Vegetarian Diet Diffusion

N/A
N/A
Protected

Academic year: 2021

Partager "Introducing the Argumentation Framework Within Agent-Based Models to Better Simulate Agents' Cognition in Opinion Dynamics: Application to Vegetarian Diet Diffusion"

Copied!
30
0
0

Texte intégral

(1)

HAL Id: hal-03234980

https://hal.inrae.fr/hal-03234980

Submitted on 25 May 2021

HAL is a multi-disciplinary open access

archive for the deposit and dissemination of

sci-entific research documents, whether they are

pub-lished or not. The documents may come from

teaching and research institutions in France or

abroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, est

destinée au dépôt et à la diffusion de documents

scientifiques de niveau recherche, publiés ou non,

émanant des établissements d’enseignement et de

recherche français ou étrangers, des laboratoires

publics ou privés.

Introducing the Argumentation Framework Within

Agent-Based Models to Better Simulate Agents’

Cognition in Opinion Dynamics: Application to

Vegetarian Diet Diffusion

Patrick Taillandier, Nicolas Salliou, Rallou Thomopoulos

To cite this version:

Patrick Taillandier, Nicolas Salliou, Rallou Thomopoulos. Introducing the Argumentation Framework

Within Agent-Based Models to Better Simulate Agents’ Cognition in Opinion Dynamics: Application

to Vegetarian Diet Diffusion. Journal of Artificial Societies and Social Simulation, SimSoc Consortium,

2021, 24 (2), �10.18564/jasss.4531�. �hal-03234980�

(2)

Introducing the Argumentation

Framework within Agent-Based Models

to Better Simulate Agents’ Cognition in

Opinion Dynamics: Application to

Vegetarian Diet Diffusion

Patrick Taillandier

1,2,3

, Nicolas Salliou

4

, Rallou

Thomopoulos

5

1National Research Institute for Agriculture, Food and the Environment (INRAE) University of Toulouse, Castanet-Tolosan, Auzeville BP 52627, France

2UMMISCO, IRD, Sorbonne University, 32 Avenue Henri Varagnat, 93143 Bondy Cedex, France 3WARM, Thuyloi University, No. 175 Tay Son, Dong Da, Hanoi, Vietnam

4ETH Zürich, Stefano-Franscini-Platz 5, Zürich, 8093, Switzerland

5IATE, University of Montpellier, INRAE, Institut Agro, 2 place Pierre Viala, 34060, Montpellier, France

Correspondence should be addressed to patrick.taillandier@gmail.com Journal of Artificial Societies and Social Simulation24(2) 6, 2021 Doi: 10.18564/jasss.4531 Url: http://jasss.soc.surrey.ac.uk/24/2/6.html Received: 06-03-2020 Accepted: 12-03-2021 Published: 31-03-2021

Abstract:This paper introduces a generic agent-based model simulating the exchange and the diffusion of pro

and con arguments. It is applied to the case of the diffusion of vegetarian diets in the context of a potential emer-gence of a second nutrition transition. To this day, agent-based simulation has been extensively used to study opinion dynamics. However, the vast majority of existing models have been limited to extremely abstract and simplified representations of the diffusion process. These simplifications impairs the realism of the simulations and disables the understanding of the reasons for the shift of an actor’s opinion. The generic model presented here explicitly represents exchanges of arguments between actors in the context of an opinion dynamic model. In particular, the inner attitude towards an opinion of each agent is formalized as an argumentation graph and each agent can share arguments with other agents. Simulation experiments show that introducing attacks be-tween arguments and a limitation of the number of arguments mobilized by agents has a strong impact on the evolution of the agents’ opinion. We also highlight that when a new argument is introduced into the system, the quantity and the profile of the agents receiving the new argument will impact the evolution of the overall opinion. Finally, the application of this model to vegetarian diet adoption seems consistent with historical food behaviour dynamics observed during crises.

Keywords:Opinion Dynamics, Agent-Based Simulation, Argumentation Framework, Vegetarian Diets

Introduction

1.1 Historically, societies moving towards higher income per capita increased their consumption of animal

prod-ucts, and known as the nutrition transition model (Popkin 1993). This trend is valid worldwide and knows little variation between countries and cultures (Sans & Combris 2015). Currently, billions are moving towards more animal product consumption, especially as the economies of China and India are growing. However, in many developed countries, a potential new trend is emerging towards a second diet transition with the stabilization and/or the decline of animal product consumption (Vranken et al. 2014). Many factors may influence this process like animal welfare, negative environmental impacts of livestock production as well as health concerns (Godfray et al. 2018; Poore & Nemecek 2018). One hypothesis to explain such a trend would be the diffusion and wider

(3)

adoption of vegetarian diets, ranging from flexitarianism (or semi-vegetarian) to veganism (Beardsworth & Keil 1991). Choosing a vegetarian diet usually stems in ethical, environmental and/or health concerns (Ruby 2012). By reducing the share of animal products consumed per capita, such a transition towards more plant-based diets would reduce harm to the environment, health risks and animal sufferings. Therefore, this transition to-wards more sustainable diets is desirable but many hindrances exist at the individual level (Stoll-Kleemann & Schmidt 2017), such as lack of awareness about these diets (Macdiarmid et al. 2016), health concerns (Herzog 2011), etc.

1.2 We focus on the role of arguments in changing people’s opinions on diets by raising awareness on animal

prod-ucts consumption. While a recent study (Scalco et al. 2019) modelled the influence of colleagues and household members on meat consumption, the relation between argument acquisition at the individual level and opinion diffusion has never been explored for vegetarian diets. Our assumption is that the diffusion of ethical, health and environmental arguments in favour of such diets probably fuels the vegetarian diet adoption process.

1.3 Agent-based modelling is a classical approach to study opinion dynamics as it takes into account the

hetero-geneity of actors and the impact of interactions between them. Among existing approaches, the most popular one - known as the bounded confidence model - uses a numerical value to represent the opinion towards an option (Deffuant et al. 2000; Hegselmann & Krause 2002). The opinion of each agent is updated by averaging a set of agent opinions. Typically, in the classic bounded confidence model, when two agents with respective opinions x and x0meet, they adjust their opinions conditionally upon their difference of opinions being smaller

in magnitude than a threshold d (i.e. the opinions of the agents are modified if |x − x0| < d):

x = x + µ × (x0− x)

x0= x0+ µ × (x − x0) (1)

with µ the parameter of speed of convergence.

1.4 In the past fifteen years, many studies have proposed to enrich this generic model, for example by taking into

account fixed uncertainties, by integrating multi-dimensional opinion dynamics (Lorenz 2003; Urbig & Malitz 2005), by studying the behaviour of the model when adding extremists (Mathias et al. 2016) or contrasting effects (Jager & Amblard 2005; Huet et al. 2008).

1.5 These models are very relevant to study social influence. However, most of them remain theoretical and only

a few have been applied to real case-studies using data and validated (Flache et al. 2017). Another drawback is the difficulty to understand the inner motivation underlying the change in an agent’s opinion. Indeed, as the opinion is usually integrated in a single numerical value, the reasons why the agent has changed his/her opinion is unknown.

1.6 To integrate the inner motivation underlying change, a relevant framework is the argumentation model (Besnard

& Hunter 2008). Argumentation deals with situations where information contains contradictions because it originates from several sources or corresponds to several points of view that possibly have different priori-ties. It is a reasoning model based on the construction and evaluation of interacting arguments. It has been formalized both in philosophy as well as in computer science (Rescher 1997) and applied to various domains including non-monotonic reasoning (Dung 1995a), decision making (Thomopoulos 2018) or negotiation (Kraus et al. 1998). The argumentation model framework introduced in Dung (1995a) consists of a set of arguments and binary relations expressing conflicts among arguments. An argument gives a reason for believing a claim, or for doing an action. Historically, the typical field of application of argumentation in computer science was the legal domain (Prakken & Sartor 2015). More recently, several studies proved its relevance in social-related concerns, medicine, food systems, supply chains, policies and controversies, especially for decision-making purposes (Thomopoulos 2018).

1.7 Mäs & Flache (2013) proposed a different theory to represent argument exchanges. Their model, itself inspired

by earlier Persuasive Argument Theory (PAT), is based on the Argument-Communication Theory of Bi-polarization (ACTB) that adds to existing models the communication of arguments. In this model, arguments are abstracted by a numerical value between -1 (con argument) and 1 (pro argument). The agent’s opinion is the average value of the arguments that the agent considers relevant. Agents disregard pieces of information not communicated in recent interactions (and consider them as not relevant). An experiment showed that the model enables to reproduce bi-polarization (i.e., the development of increasingly antagonistic groups Esteban & Ray 1994) with-out explicitly representing negative influence. A recent extension of this model has been proposed in Banisch & Olbrich (2021) to take into account different issues on which opinions can be formed (each issue is linked to a subset of arguments).

1.8 Another work using explicit arguments is Stefanelli & Seidl (2017). To do so, they collected empirical data

through questionnaires. This date plays a very important role as each agent has a set of different types of argu-ments and will form an opinion according to the valence of the arguargu-ments and the importance it gives to them.

(4)

Changes in opinion will happen by interacting with other agents in their social network and through comparison of argument values. Therefore, there is no explicit exchange of arguments as such in this model: the interaction results in the adaptation (or not) of the agents’ argument scales, i.e., the valence and importance that the agent gives to a type (benefit, risk or process) of arguments. The adaption of the agent’s argument scale is computed according to the position of each argument in its own continuum of social judgment.

1.9 Another approach by Wolf et al. (2015) does not directly use arguments but the closely related concept of "need".

In this model about electric cars is also based on empirical data and each agent assigns a weight to each iden-tified need (e.g., safety, comfort, costs) for each possible action (e.g. using an electric vehicle). These weights change during the simulation through interactions with other agents.

1.10 Although these models show interesting results, we emphasize that they are silent about argumentative

rea-soning and do not explicitly formalize the tensions between arguments. In this context, Friedkin et al. (2016) deals with statements rather than arguments and introduces the notion of logical constraints associated with statements. This means that if an agent believes that a specific statement is true, then he/she will automatically impact his/her beliefs regarding other statements linked to the previous one.

1.11 Regarding argumentation, several studies have proposed the use of the system introduced by Dung (1995a) in

opinion dynamic models. For example, in Gabbriellini & Torroni (2014), all agents reason from the same set of arguments and exchange attacks between arguments. The exchange of attacks is carried out during a dialogue phase. During this phase, an agent who is about to receive an attack he/she is disagreeing with can either accept this new attack (if the agent formulating the attack is trustworthy, which is a stochastic process), reply with a counter-attack, or end the exchange. If the agent decides to reply, the agent proposing the first attack can in turn reply. It goes on until an agent accepts an attack or ends the dialog. This model provides a very interesting basis for integrating Dung’s argumentation system into a model of opinion diffusion. Nevertheless, it is based on the hypothesis of a common set of arguments for all agents. Unlike Mäs & Flache (2013), there are no arguments but attacks exchanged here, which can be questioned. Moreover, an opinion in Gabbriellini & Torroni (2014)’s model is an argumentation framework. The model is thus a discrete opinion dynamics model, with no numeric update, which differs from our approach.

1.12 The link between arguments and opinions was explored in Villata et al. (2017) through an empirical study of

emotions. Explanation and reasoning theories were also proposed in cognitive psychology Williams & Lom-brozo (2010).

1.13 A recent study worth mentioning (Butler et al. 2019a) focuses on collective decision-making processes and

pro-poses to combine a deliberative process using Dung’s system of argumentation with a process of interpersonal influence (Deffuant et al. 2000). In this work, each argument is modelled by a real number between −1 and 1representing the support of the argument for a principle (e.g., "protect the environment"). Each agent has an opinion about the principle modelled and this opinion evolves through a group deliberative process during which agents exchange arguments and have a direct influence through pair interactions.

1.14 A last study using Dung’s system of argumentation is the model proposed by Butler et al. (2019b). Similarly to

Butler et al. (2019a), this model combines dyadic interactions (pair-wise interactions) with collective deliber-ation. One of its major contributions is the introduction of the notion of argumentative epistemic vigilance, i.e., the possibility for agents to reject an argument in case of "message-source" discrepancy. Indeed, when an agent receives an argument from another agent, he/she can invalidate it either by asserting the existence of an argument that attacks the first argument, or by pointing out that the argument he/she has received is not consistent with the opinion of the sender.

1.15 Our proposal is in line with the studies of Butler et al. (2019a,b) except that we simply focus on the "daily"

exchange of arguments using a general process of evolution of opinion close to Mäs & Flache (2013) and we do not investigate group deliberation. One of the originality of our paper go further by eliciting the content of arguments. To do so, we use the system introduced by Dung (1995a) as a basis and enrich it with detailed descriptions of the arguments. Finally, as in Stefanelli & Seidl (2014), the collection of empirical data plays an important role in the design of the model.

1.16 Section 2 presents the empirical approach we start from. Section 3 describes the generic model that we

pro-pose. Section 4 explores the model behaviour relating to the impact of the attacks in the argumentation graph and to the impact of the number of arguments known by the agents. Section 5 presents the application of this model to study the evolution of the vegetarian diet. Finally, Section 6 concludes and presents some perspec-tives of this work.

(5)

Empirical Approach

2.1 In this section, the first steps to build well-informed and populated models are instantiated on the case of

di-etary changes towards vegetarian diets. They are guided by the questions: What knowledge does literature provide about the case? What data should be collected in order to fuel the model scenarios?

Literature overview

2.2 Changes in individual behaviours have been extensively studied in the case of addictions (Prochaska et al. 1992).

Results tend to demonstrate that behavioural changes possibly follow a series of stages. However, in the case of food choices, changes in dietary patterns are quite specific and their stages are less well-identified (Povey et al. 1999). Food choices are complex, dynamic, and change over the course of a person’s life. They are determined by a wide variety of factors. In their review, Vabo & Hansen (2014) try to address these factors. Three fundamen-tal groups stand out almost systematically: factors related to the characteristics of the food (organoleptic char-acteristics, nutritional content, function, etc.), factors related to the consumer (physiological and psychological aspects) and environmental factors (economic, cultural, social context, etc.). These factors have a double influ-ence on food choices, by building the food preferinflu-ences of individuals, particularly during their childhood, and by influencing choices (through price, health, practicality, sensory appeal, mood, etc.).

Data acquisition

2.3 In order to get insights about the role of arguments in following a vegetarian diet, we conducted a survey and

built an argument database. These are published in Salliou et al. (2019) and described hereafter.

2.4 First, 1714 French citizens were surveyed and asked about their actual diet, as well as expressing on a five degree

Lickert-scale their agreement with 16 key arguments about animal product consumption. These 16 arguments were extracted from the participatory online platform Kialo which allows users to co-construct argument hi-erarchies about any topic. We considered these arguments as central as they are the main and first degree arguments over a hierarchy of more than 2000 arguments expressed by over 1,400 participants about the topic of "humans should stop eating meat" (Kialo 2021). The analysis of the survey reveals that 40% of respondents would ideally have a lower meat consumption than their current diet. This finding backs the assumption that conditions for dietary change towards more plant-based diets are significant, which supports the objective of the simulation.

2.5 Secondly, we constructed a database composed of 145 arguments obtained from google search about

vegetar-ian diets. An analysis is provided in Salliou & Thomopoulos (2018). Our sources of arguments are newspapers, grey literature and top ten google research (“vegetarian diet”; “vegan diet”; “vegetalism argument”). The lat-ter inquiry added to the pool popular scientific papers, webmedia articles and blog posts. We read thoroughly each source and extracted all arguments as expressed by their authors. For each argument we attributed a criterion (“Nutritional”; “Economic”; ”Environmental”; “Anthropological”; ”Ethical”; ”Health” or ”Social”) and noted the stakeholder expressing this argument (“Journalist”; ”Scientist”, ”Philosopher”; ”Blogger”, etc.). We also indicated for each argument whether it was pro (+) or con (-) vegetarian diets (see Table 1 for example of arguments). From this list of arguments base we built an argumentation network (Dung 1995b) by establishing "attack" relationships between arguments (Figure 1). An attack happens when an argument challenges another argument. For example, the argument "a vegan diet is healthy" is attacked by the argument "vegans have B12 vitamin deficiency". As arguments rarely mention explicitly which arguments they attack, they were elicited by us. We did not check whether the attack is legitimate or not. Arguments, connected by these "attacks" relation-ships, form an argumentation network.

(6)

Id Type Statement Rationale Criterion Actor Source type 1 - Vegan diet is deficient inB12 vitamin Vegetal proteins do notcontain B12 vitamin Nutritional Jounalist Newspaper 15 - Plant proteins trigger al-lergies Plant-based food aremore regularly allergic Nutritional Innovationcluster Powerpoint 23 + Vegetarian diet is goodfor health Diabetes, cancer andcoronary risks are

reduced Health Scientists

Scientific paper 56 + Stop eating animalsdoes not mean animal

extinction

Deforestation for the cultivation of animal feed provokes species extinctions

Environmental Bloggerpro-vegan Blog post

59 + Animals suffer wheneaten, not plants A nervous system isneeded to suffer, which

plants do not have Ethical

Blogger

pro-vegan Blog post Table 1: Example of arguments collected towards vegetarian option (extracted from Salliou & Thomopoulos 2018)

Figure 1: Graphical Representation of a sample of arguments and attacks about Reduced Meat Consumption: each number corresponds to one argument. Apart from the black node, which is common to several sources, each source is represented by one node color (extracted from Salliou & Thomopoulos 2018)

Generic Model

Main concepts

3.1 The idea behind this model is to explicitly represents agents’ own mental deliberation process from arguments

towards an opinion, through the use of the argumentation framework. We use the argumentation framework of Dung (1995a) (Definition 1) complemented with a structured description of arguments extending those intro-duced in Bourguet et al. (2013); Thomopoulos et al. (2018) (Definition 2).

Definition 1. Dung’s argumentation graph.An argumentation graph is a pair (A, R) where A is a set of

(7)

Definition 2. Argument.We define an argument by a tuple a = (I; O; T ; S; R; C; A; T s), with: • I: the identifier of the argument;

• O: the option that is concerned by the argument;

• T: the type of the argument: pro, con or neutral towards the option; • S: the statement of the argument, i.e., its conclusion;

• R: the rationale underlying the argument, i.e., its hypothesis;

• C: the importance of each criterion (e.g., "Nutritional", "Economic", "Environmental") on which the argument relies on;

• A: the agent who proposes the argument; • Ts: the type of source the argument comes from.

Example 1. An example of argument for the vegetarian diet context is ("1", "adoption of the vegetarian diet", "con", "Vegan diet is deficient in B12 vitamin", "Vegetable proteins do not contain B12 vitamin", "criterion ’Nutritional’ with a importance of 1.0", "journalist of ’Canard Enchainé’", "Newspaper").

3.2 We thus consider that each agent is characterized by a set of attributes:

• argumentation graph: a directed graph that represents a Dung’s argumentation system. Each node is an argument, and each edge represents an attack from an argument to another one. The weight of an edge represents the strength of the attack for the agent. The interested reader can refer to Yun et al. (2018) for different ways to define attacks.

• criterion importance: for each criterion that arguments rely on, a score (numerical value between 0 and 1) represents the importance of this criterion for the agent. As opinions are formed from a cognitive and an affective part (Bergman 1998), criterion are used to evaluate the affective preference of arguments. • opinion: a numerical value that corresponds to the opinion of the agent. A value higher than 0 means that

the agent is in favour of the option, a value lower than 0 means that the agent is against the option, and if the value is 0, the agent is neutral towards the option.

• behaviour: a nominal value that corresponds to the behaviour resulting from the agent’s opinion. Exam-ples of possible values in the food diet application are omnivorous, flexitarian, vegetarian or vegan. There is thus a mapping between the opinions, defined on a numerical domain, and the set of behaviours, which label predefined consumer profiles.

3.3 We define the notion of strength of an argument for an agent.

Definition 3. Argument strength.Let us consider an agent j, the strength of an argument a is defined as follows: strength(j, a) = X

c∈CRIT

jc× ac (2)

with CRIT , the set of criteria, acthe importance of the criterion c for the argument a (see Definition 2), and jcthe

importance of c for the agent j.

3.4 From the notion of strength, we define the notion of value for a set of arguments.

Definition 4. Value of a set of arguments.

Let us consider a set of arguments A for an agent j, the value of this set of arguments is computed as follows:

value(j, A) = P a∈A strength(j, a) × type(a) P a∈A strength(j, a) (3) with: type(a) =    −1 if type of a = con 0 if type of a = neutral 1 if type of a = pro

(8)

3.5 Using the notion of strength, we also define the notion of simplified argument graph:

Definition 5. Simplified argumentation graph.Let us consider an agent j, (A, R) an argumentation graph and

(a, a0) ∈ R. The simplified argumentation graph (A, R0)obtained from (A, R) is defined by: (a, a0) ∈ R0if and

only if:

• (a, a0) ∈ Rand

• if (a0, a) ∈ Rthen strength(j, a) ≥ strength(j, a0).

3.6 This means that if an argument a attacks an argument a0and if a0attacks a, only the attack that has for origin

the argument with the highest strength is kept in the simplified graph. If the arguments have the same strength, both attacks are kept.

3.7 Finally, we define the notion of preferred extension.

Definition 6. Preferred extension.Let an argumentation system (A, R) and B ⊆ A. Then:

• B is conflict-free if and only if 6 ∃ai, aj∈ B such that (ai, aj) ∈ R;

• B defends an argument ai∈ B if and only if for each argument aj∈ A, if (aj, ai) ∈ R, then ∃ak ∈ B such

that (ak, aj) ∈ R;

• a conflict-free set B of arguments is admissible if and only if B defends all its elements. A preferred extension is a maximal (with respect to set inclusion) admissible set of arguments.

Dynamics

3.8 We made the choice to use the same general model as the one proposed in Mäs & Flache (2013). A simulation

step corresponds to the exchange of an argument between two agents, i.e., an agent gives one of his/her argu-ments to another agent. When an agent learns a new argument, the oldest argument is removed from his/her argumentation graph. This forgetting process, already defined in Mäs & Flache (2013), was introduced to take into account the limitation of human cognition and memory. We also integrated the use of an argument, i.e., giving it to another agent, triggers the agent to remember it. The given argument is automatically considered as the agent’s most recent argument. Similarly, an agent who receives an argument he/she already has will not add it again in his/her argumentation graph, but will consider this argument as the most recent among his/her arguments. The effect of this mechanism is that some arguments may be forgotten by the entire population of agents. Thus, for example, if all agents tend to converge towards the same opinion, most of the arguments against that opinion will be forgotten.

3.9 Concerning the choice of the agent to exchange arguments with, we used the same partner selection method as

Mäs & Flache (2013). In each simulation step, an agent chosen randomly (uniform distribution) selects another agent. The probability that the second agent is chosen as an interaction partner depends on the similarity between the two agents in terms of opinion.

3.10 Let i and j be 2 agents, the similarity between i and j is:

Similarity(i, j) = 1

2(2 − |i.opinion − j.opinion|) (4)

3.11 And the probability for an agent i to select j for partner considering N the set of all possible partners is:

P robai(j) =

(Similarity(i, j))h P

k∈N

(Similarity(i, k))h (5)

with h, the strength of homophily.

3.12 For the choice of the argument to be given, our hypothesis is that an agent will give an argument that seems

relevant to him/her and that allowed him/her to form his/her opinion. In other words, it means picking an argu-ment belonging to the set of arguargu-ments in the preferred extension maximizing the absolute value of opinion as defined in Equation 3. The agent will choose a random argument in this set.Our proposal to randomly choose

(9)

which argument to give to another agent is due to the dependence of such action on external factors not rep-resented in the model such as the course of the discussion between individuals, the profile of the other, etc. Other choices could have been made such as giving the argument with the highest strength, the argument with the highest chance of convincing the other, etc. We discuss this point in the perspectives of the article.

3.13 Note that in the case of an argumentation graph without attacks, there will be only one preferred extension

which will be composed of all the arguments considered by the agent. We thus find ourselves in the same case as the ACTB model where the agent chooses an argument at random among all the arguments at his/her dis-posal. In the other cases, the ranking between several co-existing preferred extensions is known as the “ranking semantics” problem (Yun et al. 2020). We made a modelling choice stating that an agent chooses the preferred extension with the highest value, which expresses the motivation to favour the most adamant view stemming from the extensions.

3.14 Once an agent receives a new argument (and at the initialization of the model), the agent deliberates using

his/her argument graph to make his/her opinion. Contrary to Mercier & Sperber (2011), who state that individ-uals will be strongly critical of any new argument challenging their own opinion, we assumed no such psycho-logical reactance (Brehm 1966). The deliberation is composed of 3 steps:

1. simplifying the argumentation graph according to the weights of the edges (see Section 3.5);

2. computing the set of preferred extensions from the simplified argumentation graph (see Section 3.7); 3. computing the opinion from the preferred extensions: for each extension, the agent computes its value

using Equation 3, then returns the extension with the maximal absolute value. If several extensions have the same absolute value, then the agent randomly selects one of these extensions.

3.15 If we consider an argumentation graph with no attack and that all the criteria have the same importance for all

the agents, then, we are in the exact context of the ACTB model, when all relevant arguments have the same persuasiveness (i.e., all arguments are equally weighted in the calculation of the opinion). In this case, the evaluation of an opinion of an agent j with a set of arguments A can be directly computed by:

opinion(j) = P a∈A type(a) Card(A) (6) with: type(a) =    −1 if type of a = con 0 if type of a = neutral 1 if type of a = pro

3.16 Convergence towards a steady state can only be achieved if none of the agents can change their opinion no

matter what happens in terms of exchanging arguments. The definition of such a steady state depends on the strength of homophily h. Indeed, the model relies on a stochastic choice of agents to exchange arguments (see Equation 5): if h = 0, it means that all agents can exchange arguments with all other agents even if they have a very different opinion; if h > 0, it means that all the agents can exchange arguments with all other agents unless they have a completely different opinion (i.e., if one of the agents has a −1 opinion and the other has a 1opinion). Thus, in the first case, to be sure to obtain a stable state, all the agents must have the same opinion (−1 or 1) and arguments of a homogeneous type (all pro or all con). In the case of h > 1, the first condition can be relaxed: all agents must have arguments of homogeneous type (all pro or all con) but their opinion can be either −1 or 1. Indeed, in this case, agents will only exchange arguments with agents who already have the same opinion as them and the new arguments brought, in accordance with the opinion of both agents, will not have an impact on the result of the opinion calculation.

Implementation

3.17 The model was implemented with the GAMA platform (Taillandier et al. 2019). GAMA provides modellers with a

dedicated modelling language which is easy to use and learn. It also allows them to naturally integrate GIS data and includes an extension dedicated to generating a spatialized and structured synthetic population (Chapuis et al. 2018), which is particularly interesting for building empirically grounded models. The main components of the model (arguments, argumentative agents, etc.) were implemented as a plugin for the GAMA platform. The interest of making a plug-in is to facilitate the reuse of these elements in other models. Thus, a modeller wishing to use them will just have to import the plugin and she/he will be able to directly use all these functions. This is particularly interesting for non simple functions such as the calculation of preferred extensions which

(10)

is based on the JArgSemSAT Java library (Cerutti et al. 2017). The plugin was designed to be as modular as possible allowing the modellers to customize all the existing functions (for example, the computation of the argument strength). It was developed under the GPL-3 licence, and is available on Github (Github 2021). It can be directly downloaded and installed from GAMA 1.8.1 from the GAMA experimental p2 update site.

Model Exploration

4.1 In this section, we explore the impact of different parameters on the simulation results, respectively the number

of attacks in the global argumentation graph, the strength of homophily (h), and the number of arguments per agent. The values of the parameters used for the 3 experiments are given in the Table 2.

Experiment number of attacks h number of arguments

per individual Experiment 1: influence of argument attacks 0, 100, 200, 300, 500, 1000 10 10 Experiment 2: Influence of homophily strength 300 0, 1, 5, 10, 50, 100 10 Experiment 3: influence of

the number of arguments 300 10 1, 3, 7, 10, 30, 60

Table 2: Parameter values used for the experiments

4.2 For this exploration, we carried out experiments using conditions close to those used in Mäs & Flache (2013) to

study the dynamics of bi-polarization.

4.3 We simulated a population of 100 agents, homogeneous in the sense that criteria have the same importance

for all agents. Each agent is initialized with a set of 10 arguments chosen at random among 60 arguments (30 pro and 30 con). At the beginning of the simulation, for each agent, an oldness value between 1 (recent) and 10 (old) is assigned to each argument: one argument has a oldness of 1, another of 2, another of 3 and so on up to 10. We also considered that 300 attacks link the arguments. The attacks are randomly generated between two arguments with different conclusions. Indeed, we consider that a pro argument can only attack a con argument and vice versa. Since we have 30 pro and 30 con arguments, the maximum number of attacks is 1800.

4.4 Concerning the strength of homophily, we set the value of h at 10 for our experiments.

4.5 We studied the change of agents’ opinions over 1,000,000 simulation events. This number was chosen to

max-imsze the chances of reaching a steady state. As the model is stochastic, we ran the simulation 100 times per value of parameters.

4.6 In terms of outputs, we analysed the average distribution of opinions for the 100 repetitions, the number of

stable states obtained and the average number of states to reach such a stable state, and finally the evolution of the polarization of the agents’ opinion.

4.7 Concerning the polarization at time t, we use the following equation to estimate it:

Pt= 1 |N |(|N | − 1) i∈N,j∈N X i6=j (dij,t− γt)2 (7)

where N is the set of agents, dij,tthe distance between the opinions of agent i and agent j at time t computed

as dij,t= |opinioni,t− opinionj,t|and γtthe mean opinion distance among all the agents at time t.

Influence of argument attacks

4.8 In order to evaluate the impact of considering attacks in the model that are not taken into account in the ACTB

model, we ran the model using the same conditions as the one presented in the previous section (same param-eter values, same number of iteration, and same number of replication). The only difference is that we vary the number of attacks in the global argumentation graph. We tested 6 values for the number of attacks: 0, 100, 200, 300, 500, and 1000.

(11)

Figure 2: Evolution of the polarization of opinions according the the number of attacks (h = 10; number of argu-ments = 10): x-axis, time step; y-axis, mean value of polarization for the 100 repetitions.

(12)

Figure 3: Impact of the number of attacks on the opinions of agents (h = 10; number of arguments = 10): x-axis, value of opinion; y-axis, number of agents per opinion values.

4.9 Figure 2 and Figure 3 show the results obtained for different number of attacks. As showed, the attacks between

arguments have a strong impact on the result. Indeed, in the case where no attack is taken into account, no phenomenon of bi-polarization of opinion is visible and there is no convergence towards an unique opinion, whereas it is very marked as soon as the number of attacks exceeds 300. This result can be explained by the fact that the larger the number of attacks, the smaller the number of arguments that the agent considers relevant, because the attacked arguments that are not defended are not retained in his/her preferred extensions (see Section 3.7). In addition, the arguments that are relevant for the agent often supports the same conclusion, leading to a polarization of the agent’s opinion. It can also be observed that for 300 attacks, the agents’ opinion tends to converge towards a bipolarization, whereas when the number of attacks increases (500 or 1000) the agents’ opinion tends to converge towards a single value. In fact, the higher the number of attacks, the greater the chance of giving an argument that attacks the other arguments, and therefore the greater the chance that the other agent changes his or her opinion for an opinion close to that of the one who gave the argument, which ultimately leads to a reinforcing effect as a consensus begins to emerge.

Influence of homophily strength

4.10 We tested the model with the following values for h: 0, 1, 5, 10, 50, and 100. A value of 0 means that the agent

(13)

4.11 As shown in Figure 4, when h = 0, the polarization value quickly converges towards 0, which means that all

agents converge towards the same opinion. We can see in Figure 5 that opinion of agents converges towards the two extreme categories ([−1, −0.75[ and [0.75, 1]), i.e. at the end of the simulation, either all agents have an opinion between [−1, −0.75], or all agents have an opinion between [0.75, 1]. Indeed, as already mentioned in the previous experiment, a high number of attacks (300 in this experiment) means that the number of argu-ments that the agent considers as relevant is low. This is because attacked arguargu-ments that are not defended are not in the agent’s preferred extensions and thus agents often have a rather polarized opinion. As in the case where h = 0, agents can give arguments to all other agents with the same probability, even to those with a very different opinion. The higher the number of agents sharing the same opinion, the faster they are able to con-vince agents with a different opinion to converge towards their opinion. This creates a reinforcing phenomenon leading to a fast convergence towards a uniform opinion (polarization = 0). This phenomenon can be observed in the polarization chart. Note that in Figure 5, for h = 0, the agents’ opinions seem to be bipolarized (half of the agents with an opinion in the interval [−1, −0.75[ and the other half with an opinion in the interval [0.75, 1]. In reality, this is an effect due to the aggregation of the 100 simulations: over the 100 simulations, if in 50 the agents converge to an opinion value of −1 and in 50 to an opinion value of 1, on average the agents will be evenly distributed between these extreme intervals. The fact that the standard deviation is very high is a good indication of this type of phenomenon.

4.12 As the value of h increases, the agents tend to converge more and more towards higher values of polarization

(and an increasingly smaller standard deviation in the distribution of opinions - Figure 5), up to a certain level (above 50). Indeed, as shown by Mäs & Flache (2013), the increase in the value of h will lead to the observation of a phenomenon of bipolarization: The higher the value of h, the less agents will interact with agents with very different opinions and therefore try to convince them. Talking only to agents with similar opinions will also mean that the pool of arguments to which they will be subjected will be smaller, and agents will mostly receive arguments that are consistent with their opinion, leading to a phenomenon of reinforcement of their opinion to an extreme. From a certain level of h, the agents will only exchange arguments with agents already having a very close opinion to them, which explains why several clusters may appear and therefore the polarization value obtained is lower for h = 100 than for h = 50.

Figure 4: Evolution of the polarization of opinions according the the homophily strength (number of attacks = 300; number of arguments = 10): x-axis, time step; y-axis, mean value of polarization for the 100 repetitions.

(14)

Figure 5: Impact of the homophily strength(h) on the opinions of agents (number of attacks = 300; number of arguments = 10): x-axis, value of opinion; y-axis, number of agents per opinion values.

Influence of the number of arguments per individual

4.13 In order to assess the impact of the number of arguments per individual, we carried out an experiment using

the same conditions as the previous experiment. The only difference is that we vary the number of arguments known by each agent (10 in the previous experiment). We tested 6 values for the number of arguments: 1, 3, 7, 10, 30, and 60 (i.e. every agent knows all the arguments).

4.14 Figure 6 and Figure 7 show that the number of arguments known by each agent has a strong impact on the

result.

4.15 In the case of a single argument, a perfect bi-polarization is observed, which is normal: each agent has a single

argument and builds its decision from it. As we have as many pro as con arguments, each agent has a 1 chance out of 2 to have one of these two types of arguments and thus to have an opinion totally pro (opinion of 1) or con (opinion of -1) the option.

4.16 It is also expected that when all agents have all 60 arguments, as all agents have the same criteria values, all

agents will have the same preferred extensions and the same absolute values for them. With 300 attacks, the chances of having unattacked arguments are quite low, so the chances of getting homogeneous extensions (only arguments with the same conclusion) are high. Therefore, very often, two extensions will appear with

(15)

contradictory opinions (-1 and 1). Each agent will therefore have an equal chance of having at the end a totally pro (opinion of 1) or con (opinion of -1) opinion to the option. Very often, but not for every simulation, the polarization value will be close to 1, which explains the average polarization value of 0.82.

4.17 With 3 arguments, the global opinion tends to converge to a single value (polarization value close to 0, and

opin-ions mostly between [-1,-0.75[ and [0.75, 1.0]). This result can be explained by the fact that with 3 arguments, only 2 configurations exist: three homogeneous arguments, two arguments with the same conclusion and one with a different conclusion. The number of possible opinion values is therefore limited: -/+1 (3 arguments with the same conclusion in the preferred extension), -/+0.33 (2 arguments with the same conclusion and 1 argu-ment with a different conclusion in the preferred extension), 0 (1 arguargu-ment for and 1 arguargu-ment against in the preferred extension). In this context, the reception of an argument will have a profound impact on the agent’s opinion. Once the number of agents with more arguments pro (or con) is greater than the number of agents with arguments con (or pro), the overall opinion irremediably converges towards a single opinion.

4.18 In the case of 7 arguments, the results are more contrasted, due to the greater number of possible combinations.

The value of polarization is higher because the agents’ opinions are more distributed in the opinion space and because a greater number of simulations converge towards a bipolarization, which results in an increase in the value of the polarization. This phenomenon is further reinforced by 10 arguments where the number of simulations converging towards a bipolarization increases.

4.19 Finally, in the case of 30 arguments, as in the case of 60 arguments, with 300 attacks and many arguments,

the chances of having unattacked arguments are quite low, so the chances of obtaining homogeneous exten-sions (only arguments with the same conclusion) are high. In this context, the simulations converge either to a single value of opinion (for most of them) or to a strong bipolarization, which explains this low mean value of polarization.

Figure 6: Evolution of the polarization of opinions according the the number of arguments per individual (h = 10; number of attacks = 300): x-axis, time step; y-axis, mean value of polarization for the 100 repetitions.

(16)

Figure 7: Impact of the number of arguments per individual on the opinions of agents (h = 10; number of attacks = 300): x-axis, value of opinion; y-axis, number of agents per opinion values.

Application to Vegetarian Diet Diffusion

5.1 The previous model was instantiated in the context of the vegetarian diet diffusion. Although the survey shows

that there is an important difference between the opinion (answer to the question "Ideally, what diet would you like to have in the future?") and the practice (answer to the question "What is your current diet?"), in this series of experiments we chose to focus above all on the evolution of agents’ opinion and not on their practice. More particularly, we wanted to show how the introduction of a new argument impacts more or less the general opinion of agents.

5.2 For this series of experiments, we used the 145 arguments collected in Salliou et al. (2019).

5.3 We used the results of the survey to generate the agents: we created from the survey a population of 1714 agents.

For each of them, we defined the opinion option –omnivorous (40%), flexitarian (49.1%), vegetarian (8.3%) and vegan (2.6%)– from the answer to the question "Ideally, what diet would you like to have in the future?". These proportions illustrate the share of opinions in the sample and not their declared diet in practice, which have much lower proportions of individuals following one of the vegetarian diets.

5.4 For each agent, we drew a set of initial arguments according to the type of arguments (pro) and (con) depending

(17)

was chosen because research on human cognitive capabilities tends to show that humans can process and re-call about 7 pieces of information (Miller 1956; Mäs & Flache 2013). We experimentally checked the coherence of this number, by observing the number of arguments spontaneously provided by 16 individuals asked to list pro and con arguments concerning the consumption of animal food products: 3 to 13 arguments were sponta-neously given, with an average of 6,2. The number of pro arguments per agent was defined randomly (uniform distribution) using the following intervals:

• omnivorous: [5-7] • flexitarian: [3-4] • vegetarian: [1-2] • vegan: [0-1]

5.5 For each agent, the number of con arguments is 7 - number of pros arguments.

5.6 For the criterion importance values, we used the survey to determine the relative importance of the different

criteria. More precisely, we used the answer to the question concerning the degree of agreement with argu-ments: as each argument is linked to a criterion, we used the answer (Lickert-scale, i.e. value between 1 and 5) to give a value to the linked criterion. Figure 8 shows the distribution of scores given for each criterion. Table 3 shows the mean value for each criterion per category of people. We can observe that omnivores tend to give more value than others to health, anthropological and nutritional criteria while vegetarians and even more so vegans tend to give more value than others to ethical and environmental criteria.

Figure 8: For each criterion, number of agents that have a value of 1 (less important), 2, 3, 4 and 5 (most impor-tant).

Criterion Omnivorous Flexitarian Vegetarian Vegan

Health 3.56 2.95 2.36 1.71

Ethical 2.64 3.36 4.03 4.54

Anthropological 3.92 3.47 2.68 1.59

Environmental 3.03 3.68 4.11 4.57

Nutritional 3.93 3.48 2.61 2.13

Table 3: Mean value for each criterion per category (Omnivorous, Flexitarian, Vegetarian, Vegan)

5.7 Figure 9 shows the distribution of opinions at initialization. A first observation is that the stochasticity is low:

despite the random drawing of the arguments, the standard deviation obtained for each group is very small.

5.8 Another observation is that the distribution of opinions appears to be consistent with the data. For example,

10.9% of the respondents defined vegetarianism or veganism as the ideal diet which is consistent with the per-centage of 10.5% for the agents with an opinion higher than 0.5. Similarly, if we take all the agents with a neg-ative opinion about vegetarianism (opinion below 0), we get 44.8%, which is close to the 40% of people who stated that their ideal diet is omnivorous.

(18)

Figure 9: Initial opinions of the agents: x-axis, value of opinion; y-axis, percentage of agents per opinion values.

5.9 We used the homophily rule for partner selection with a value of 10 for h and we still used the same rule as

in Section 4.8 for the choice of the argument to be given (one of the preferred extensions that maximises the absolute value of the opinion).

Evolution of the system without introducing new arguments

5.10 We have studied the normal evolution of the system when no new arguments are introduced. As the simulation

is stochastic, we performed 30 repetitions of the simulation. During 2,000,000 simulation steps, we analysed the evolution of opinion and the polarization of opinions.

5.11 Figure 11 shows the evolution of the agents’ opinions for the 2,000,000 simulation steps. The general opinion

tends to converge towards a mean value of 0.1, which shows that the vegetarian diet tends to be better accepted by the agent population. At the end of the 2,000,000 simulation steps, we can observe by comparing the dis-tribution obtained in Figure 9 and Figure 10 that the main evolution is the increase of agents with an extremist opinion, especially pro-extremist opinions (opinion >=0.75), that can also be observed in Figure 12 with an in-crease of the polarization from 0.2 to 0.45. The inin-crease in this number of extremes could already be observed in the experiment presented in Section 4.10. When h has a value of 10, there is indeed a tendency towards polarization and the extremization of opinions.

(19)

Figure 10: Opinions of the agents after 2,000,000 simulation steps (mean of the 30 replications): x-axis, value of opinion; y-axis, percentage of agents per opinion values.

Figure 11: Evolution of the general opinion. X-axis: simulation step; Y-axis : value of opinion for the 30 replica-tions.

(20)

Figure 12: Evolution of the polarization. X-axis: simulation step; Y-axis : value of polarization for the 30 replica-tions.

Impact of the number of the agents to whom the new argument is given

5.12 The goal of this experiment is to analyse the impacts of giving a new argument to different numbers of agents.

We tested a general scenario where a new argument is introduced at the initialization of the simulation to cer-tain percentage of the agent population. The new argument is pro vegetarian diet and attacks 5 arguments con vegetarian diet (randomly selected from arguments that concerns the same criterion). The argument is given to randomly selected agents regardless of their opinion and with a random criterion concerned by the argument. As the simulation is stochastic, we ran 30 replications for each configuration. For each configuration, we evalu-ated, during 200,000 simulation steps, the mean opinion, the value of polarization, and the number of agents that have the new argument in their argumentation graph.

% of informed agents init opinion final opinion Polarization final % of informed agents

0% -0.204 (0.005) 0.046 (0.035) 0.203 (0.004) 0.0% (0.0%)

10% -0.183 (0.006) 0.072 (0.028) 0.2 (0.004) 9.28% (3.1%)

20% -0.162 (0.008) 0.081 (0.035) 0.198 (0.004) 12.4% (2.5%)

50% -0.097 (0.017) 0.084 (0.041) 0.186 (0.005) 12.1% (3.4%)

100% 0.011 (0.033) 0.085 (0.038) 0.154 (0.005) 13.5% (2.3%)

Table 4: For different percentage of agents receiving the argument at the simulation initialization, mean value of opinion at step 0 and 200,000 and number of agents still having the argument in their argumentation network after 200,000 simulation steps. Mean value on the 30 replications (with the standard deviation).

(21)

Figure 13: Evolution of the general opinion for different percentage of agents receiving the argument. X-axis: simulation step; Y-axis : mean value of opinion for the 30 replications.

Figure 14: Evolution of the polarization for different percentage of agents receiving the argument. X-axis: simu-lation step; Y-axis : mean value of polarization for the 30 replications.

(22)

Figure 15: Evolution of the number of agents with the new argument for different percentage of agents receiv-ing the argument. X-axis: simulation step; Y-axis : mean number of agents with the new argument for the 30 replications.

5.13 Table 4 shows the results obtained after 200,000 simulation steps for different percentages of agents receiving

the argument at the initialization of the simulation, Figure 13 the evolution of the general opinion for these values, Figure 14 the evolution of the polarization and Figure 14 the evolution of the number of agents who have the new argument in their argumentation graph.

5.14 A first observation is that for all proportion values, the polarization increases (Figure 14). These values evolve

in a similar way whatever the proportion value.

5.15 Another observation, which was expected, is that the introduction of the new argument has a significant impact

on the initial opinion of the agents (p − value by Wilcoxon test lower than 1.0e − 14 between all proportion values). Thus, the impact of the introduction of the argument is all the stronger the higher the number of agents is concerned. It can also be observed that while introducing an argument has a significant impact on the opinion results obtained after 200,000 steps of simulations (p-value with Wilcoxon test lower than 0.01 for all proportion combinations), introducing it to 10% of the agents or more has no significant impact. Indeed, the rate of 10% triggered the most significant change in the evolution of global opinion: when the rate increases, the evolution of global opinion tends to decrease. This can also be observed on the evolution curves: at the beginning of the simulation, the general opinion tends to increase in all cases, but when the argument is introduced to all the agents, the overall opinion tends to decrease rapidly (the higher the rate of introduction, the greater the drop) before increasing slowly again afterwards.

5.16 This increase at the beginning and this fall can be explained by the forgetting process: at the beginning of the

simulation, the agents tend to keep the new argument, considered as recent. However, after a while, the agents who can mobilize this argument (mainly omnivores) tend to forget it, which impacts their opinion. This phe-nomenon is visible on Figure 15: for a proportion of 1.0, at the beginning of the simulation, all the agents have the argument that has been introduced, but after a certain number of simulation steps, the number of agents with this argument decreases rapidly until converging to a number close to 230 agents. Also, for the case of a proportion of 0.5, the number of agents having the argument increases at the beginning: as the argument is re-cent, no agent forgets it, but as the agents who have this new argument in their preferred extension propagate it, the number of agents having the argument increases. Then, as in the case of the 1.0 proportion, once the argument becomes older, agents start to forget it.

(23)

5.17 This result could potentially explain part of the evolution of meat consumption in the context of the mad cow

disease (or BSE for Bovine spongiform encephalopathy) in the 90’s. Godfray et al. (2018) show how the European consumption of meat dramatically fell in the beginning of the 1990’s when many consumers discovered the disease through considerable media coverage (Ashworth & Mainland 1995). The precise percentage of European consumers who received information conveying arguments against beef consumption due to health risks is unknown. However, we think it’s a fair assumption to consider this percentage was very high. By the mid 90’s, the crisis ended and meat consumption started to steadily grow again but at a lower growth rate than before (Godfray et al. 2018). In that sense, Figure 13 shows that when all agents received the argument the general opinion partly recovers after the initial introduction of the new argument. Under the assumption that opinion partly translates into behaviour (Bleda & Shackley 2012), our model seems able to reproduce some observed behaviour in diet change under a wide introduction of an argument.

5.18 This phenomenon also exists when the argument is introduced to a smaller number of agents, but it will quickly

be disseminated by those agents whose opinion is favorable to the argument to other agents who are also potentially favourable to it (due to the homophily rule).

Impact of the profile of the agents to whom the new argument is given

5.19 The goal of this experiment is to analyse if the introduction to certain agent profiles (vegan, vegetarian...)

im-pacts more or less the general opinion of agents. We tested a general scenario with a new argument introduced at the initialization of the simulation to 20% of the agent population (344 agents). The new argument is pro veg-etarian diet and attacks 5 arguments con vegveg-etarian diet (randomly selected from arguments that concerns the same criterion). This scenario was tested with 4 profiles of agents receiving the argument: no specific profile, i.e. the argument is given to 344 randomly selected agents regardless of their opinion; 344 agents with lowest opinion value (i.e. omnivorous); 344 with the more neutral value - opinion closest to 0.0 (i.e. flexitarian); 344 agents with highest opinion value (i.e. vegetarian or vegan agents). We also tested the 3 criteria concerned by the argument that were the most represented in the argumentation network: nutritional, health and ethical. As the simulation is stochastic, we run 30 replications for each configuration. For each configuration, we evalu-ated, during 200,000 simulation steps, the mean opinion, the value of polarization, and the number of agents that have the new argument in their argumentation graph.

Profile Criterion init opinion final opinion Polarization final % of informed agents

no argument - -0.202(0.01) 0.05(0.03) 0.48 (0.017) 0% (0%) no profile Nutritional -0.16 (0.01) 0.097(0.03) 0.47 (0.013) 13.83% (3.1%) Health -0.162 (0.01) 0.094(0.02) 0.47 (0.013) 14.1% (3.2%) Ethical -0.159 (0.01) 0.098(0.02) 0.47 (0.014) 14.83% (3.6%) Con Nutritional -0.135 (0.01) 0.091(0.03) 0.47 (0.015) 11.97% (2.5%) Health -0.143 (0.01) 0.088(0.04) 0.47 (0.016) 13.17% (2.8%) Ethical -0.158 (0.01) 0.087(0.03) 0.47 (0.017) 14.44% (3.2%) Neutral Nutritional -0.157 (0.01) 0.091(0.03) 0.47 (0.014) 11.6% (3.1%) Health -0.166 (0.01) 0.097(0.03) 0.47 (0.013) 13.83% (3.1%) Ethical -0.157 (0.01) 0.099(0.03) 0.47 (0.015) 12.9% (3.2%) Pro Nutritional -0.181 (0.01) 0.08(0.03) 0.48 (0.013) 9.7% (3%) Health -0.185 (0.01) 0.079(0.03) 0.47 (0.016) 9.2% (3.5%) Ethical -0.178 (0.01) 0.095(0.03) 0.47 (0.016) 11.01% (3.9%)

Table 5: For each profile of agents receiving the argument and for each criterion concerned by the argument, mean value of opinion at step 0 and 200,000 and number of agents still having the argument in their argumen-tation network after 200,000 simulation steps. Mean value on the 30 replications (with the standard deviation).

(24)

Figure 16: Evolution of the general opinion for different profiles of agents receiving the argument and different criteria concerned by the argument. X-axis: simulation step; Y-axis: mean value of opinion for the 30 replica-tions.

Figure 17: Evolution of the polarization for different profiles of agents receiving the argument and different criteria concerned by the argument. X-axis: simulation step; Y-axis: mean value of polarization for the 30 repli-cations.

(25)

Figure 18: Evolution of the number of agents with the new argument for different profiles of agents receiving the argument and different criteria concerned by the argument. X-axis: simulation step; Y-axis: mean number of agents with the new argument for the 30 replications.

5.20 Table 5 shows the results obtained after 200,000 simulation steps for different profiles of agents receiving the

argument and for different criteria concerned by the argument, Figure 16 the evolution of the general opinion for these values, Figure 17 the evolution of the polarization and Figure 17 the evolution of the number of agents having the new argument in their argumentation graph. An initial observation is that regardless of the audi-ence that receives the argument and the criterion to which it relates, the introduction of the argument always allows a significant increase in the value of the initial and final opinions (p-value with Wilcoxon’s test lower than 0.001for all combinations of profile and criteria). An interesting point is that for some of the combinations the introduction of the new argument allowed a more rapid evolution of the general opinion than without it. This means that not only did the introduction of an argument have an immediate effect (higher value of opinion in the initial state), but that having this new argument made it possible to convince other agents more quickly.

5.21 Another phenomenon is that whatever the profile of the agent receiving the argument and the criterion

con-cerned, the number of agents having it in their argumentation graph evolves in a similar way: as observed in Section 5.12, initially, the argument spreads and therefore the number of agents having it in their possession increases. Then, the agents who do not use the argument start to forget the argument and the number of agents having this argument decreases until a level lower than the initial number of agents having the argument.

5.22 Finally, a general observation is that the introduction of the new argument, whatever the profile of the agents

or the criterion concerned, has no real impact on polarization. After 200,000 simulation steps, all combinations of parameters converge to a polarization value of 0.48.

5.23 Concerning the impact of the criterion concerned and of the profile of agents receiving the argument, results

show that depending of the profile concerned, the criterion concerned does not have the same impact. Indeed, the impact (opinion value) of an argument diffused to con agents is slightly greater when it concerns the nutri-tional criterion whereas the impact is significantly greater for pro agents when it concerns the ethical criterion. Among the three agent profiles (neutral, pro, con), it is when the argument is given to neutral agents that it has the most impact. It is a hesitant public about its attitude towards vegetarian diets and can therefore easily be convinced to switch to a stricter vegetarian diet.

5.24 To interpret these observations, its’ important to stress that the population was initialized based on the actual

survey results. In the case of omnivores for instance, the value of ethical criterion is low for most of the agents, while it is very high for most of vegetarian/vegan agents. In reality, omnivorous are probably protected from

(26)

ethical attacks as they somehow own a "normative argument" as their behaviour is in line with the social norm. Such argument is not present in the database of arguments for the very reason that the benefit of social norms for individuals is to avoid to loose time and energy to wonder about the legitimacy of a behaviour and to state its existence (Epstein 2006). In a similar fashion, many nutritional arguments are against vegan diets, claiming the nutritional balance is questionable. Yet the nutritional criterion has among the highest importance values for omnivorous agents, which leads nutritional arguments to destabilize the argumentation system for them.

5.25 Important differences can also be observed for the evolution of the number of agents having the argument in

their argumentation graph. It is when the argument is not given to any particular agent profile (no profile) that the argument propagates best at the beginning of the simulation: indeed, the homophily rule makes that agents rather exchange arguments with like-minded agents. Therefore, when the argument is given to agents with a similar profile (i.e. close value of opinion), the argument will circulate less among other types of agents. The fact that the argument circulates less when it is given to pro agents can be explained by the fact that the argument will attack con arguments but will not be attacked: it will therefore often end up in the preferred extensions of the agents having it. But this extension will often be composed of a larger number of arguments in the case of pro agents as they will often have a set of pro arguments that are not attacked. This is not the case for con agents, who will often see their con argument attacked by this new argument and therefore have less argument in their preferred extension. This situation gives a better chance to pass on this argument, which explains the better diffusion of the argument.

Conclusion and Perspectives

6.1 The paper presented a generic opinion dynamic model implemented with the GAMA platform based on the

use of formal argumentation. The use of the model was illustrated through an application about the diffusion of a favourable opinion of vegetarian diet. The carried out experiment shows the possibilities offered by the framework. We plan to go further in the analysis of the model, in particular in terms of analysis of the steady-states. To that end, we intend to draw on the work of Camargo (2020) which proposes methods for analysing the steady-state of the model proposed by Banisch & Olbrich (2021).

6.2 Like the studies of Gabbriellini & Torroni (2014); Butler et al. (2019a), this study contributes to bridge formal

ar-gumentation and agent-based models of opinion dynamics. Our goal for the future is to continue to strengthen this bridge. Thus, we plan to enrich the way arguments are evaluated. In the current version, the evaluation of arguments depends on the criteria concerned by the argument and on the importance of these criteria for the agent. Other factors can impact the perception of an argument, and among them, the source of the argu-ment (Pornpitakpan 2004). As an example, the profusion of fake news from dubious source can impact people differently. Our model should soon be able to take this difference of perception into account. We also plan to use an approach similar to Banisch & Olbrich (2021) to take into account different attitudes on different issues. For example, for the dissemination of the vegetarian diet, instead of considering a single issue on vegetarian diets, we could imagine having four different issues related to different types of diets (omnivorous, flexitarian, vegetarian, vegan) with a specific subset of arguments for each. Indeed, some arguments may only concern one specific diet and not the others.

6.3 Another enrichment that we plan to add is a mechanism to enable the evolution of the criterion of importance

for the agents. Indeed, these values are not fixed for life but can evolve after a particular event and from the influence of others. We also plan to enrich the argument protocol. In that sense, we could adapt the type of ar-guments exchanged depending on the opinions hold by both parties. For example, a flexitarian would exchange a pro-vegetarian argument to an omnivore, but would exchange a pro-omnivore argument to a vegetarian. In this context, it would also be interesting to go further by taking inspiration from the studies of Gabbriellini & Torroni (2014); Butler et al. (2019a) that integrate real exchanges of argument and where an agent does not sim-ply give an argument to another agent without the latter being able to respond. In this sense, the notion of trust introduced by Gabbriellini & Torroni (2014) is interesting, as well as the notions of tables for deliberations (Butler et al. 2019a) and epistemic vigilance introduced (Butler et al. 2019b).

6.4 A last perspective concerns the link between the model and the BEN (Behaviour with Emotions and Norms)

agent architecture (Taillandier et al. 2016; Bourgais et al. 2017, 2020). Indeed, in addition to the BDI (Belief Desire Intention) reasoning engine, the BEN architecture introduces numerous concepts that could be interesting for our work such as the personality of agents based on the classic OCEAN model (McCrae & John 1992) and the social relation between agents evaluated according to 5 dimensions (liking, dominance, solidarity, familiarity and trust).

Figure

Figure 1: Graphical Representation of a sample of arguments and attacks about Reduced Meat Consumption:
Table 2: Parameter values used for the experiments
Figure 2: Evolution of the polarization of opinions according the the number of attacks ( h = 10; number of argu- argu-ments = 10): x-axis, time step; y-axis, mean value of polarization for the 100 repetitions.
Figure 3: Impact of the number of attacks on the opinions of agents ( h = 10; number of arguments = 10): x-axis, value of opinion; y-axis, number of agents per opinion values.
+7

Références

Documents relatifs

In the high risk cohort of patients with node-positive disease and ER-negative primary tumors, the cumulative incidence of CNS metastases at any time was significantly higher than

In Step 3 - Goal Modeling, the goal model that is generated by the Translation process is evaluated by the stakeholders.. These models can be used as a discussion means to

Nous souscrivons donc aux recommandations européennes les plus récentes concernant le traitement anticoagulant et antiagrégant des patients atteints de fibrillation auriculaire (et

Dans le cadre du transfert, ouvrir les portes de l'atelier lors des audits constituait une occasion de présenter un atelier flambant neuf, répondant à des exigences Qualité

When persons with or without national affiliations migrate between regions, countries, states, and continents, then the forms of this movement, the attributions

dataset (293 species for which ≥15 plant species were assayed) with absHRB and relative host range breadth (relHRB) variables included as

Aag-initiated base excision repair promotes ischemia reperfusion injury in liver, brain, and kidney.. The MIT Faculty has made this article

Mean hourly AC electrical energy use for the Test and Reference house on the six baseline days closest in weather to our peak load reduction days.. Also shown is the