Abstract
Achieving social order in societies of self-interested autonomous agents is a difficult problem due to lack of trust in the actions of others and the temptation to seek rewards at the expense of others. In human society, social norms play a strong role in fostering cooperative behaviour—as long as the value of cooperation and the cost of defection are understood by a large proportion of society. Prior work has shown the importance of both norms and metanorms (requiring punishment of defection) to produce and maintain norm-compliant behaviour in a society, e.g. as in Axelrod’s approach of learning of individual behavioural characteristics of boldness and vengefulness. However, much of this work (including Axelrod’s) uses simplified simulation scenarios in which norms are implicit in the code or are represented as simple bit strings, which limits the practical application of these methods for agents that interact across a range of real-world scenarios with complex norms. This work presents a generalisation of Axelrod’s approach in which norms are explicitly represented and agents can choose their actions after performing what-if reasoning using a version of the event calculus that tracks the creation, fulfilment and violation of expectations. This approach allows agents to continually learn and apply their boldness and vengefulness parameters across multiple scenarios with differing norms. The approach is illustrated using Axelrod’s scenario as well as a social dilemma from the behavioural game theory literature.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
The conflict between social benefit and an individual’s self-interest is a central challenge in all social relationships as individuals may put their own interests ahead of those of the society as a whole, often leading to a suboptimal outcome for all—a situation known as a social dilemma.
Understanding how societies can solve this conflict and achieve cooperation toward the collective good is essential. In a fishing scenario, for example, it is profitable for a fisherman to catch as many fish as possible, but if everyone is selfishly doing the same thing, the fishery will eventually run out of fish.
The question of why people often do cooperate with others remains, despite the fact that individuals may benefit more from defecting than from cooperating. The classical game theory literature from the last few decades models social dilemmas using payoff matrices or trees and solution concepts such as the Nash equilibrium, while making the assumption that all agents are perfectly rational and well-informed. However, this becomes intractable to reason about for a large number of agents.
Human society suggests that cooperation can occur due to internal and social motivations such as altruism, rational expectations (e.g. focal points), social choice mechanisms (e.g. voting and bargaining) and social norms [9].
One body of prior work has focused on the role of norms, providing evidence that social norms play an important role in fostering cooperation. Social norms imply that members of society should comply with prescribed behaviour while avoiding proscribed behaviours [15]. Bicchieri explains why agents adhere to social norms, claiming that a social norm emerges as a result of our expectations of others and beliefs about their expectations [4]. Axelrod uses an evolutionary computing approach to show how agents adopt normative behaviour after learning individual parameters of boldness and vengefulness, where their boldness represents their propensity to violate norms and vengefulness represents their inclination to punish others for violating norms [3]. However, his approach is based on an implicit representation of norms. The norms themselves play no role in his simulation. Instead the agents have hard-coded logic for using the boldness and vengefulness parameters to inform decisions about whether to cooperate or defect and whether to punish defectors and agents that observe defections but do not punish the defectors. This limits its practical use for agents that interact in a variety of real-world situations with a range of different norms.
In this work, we propose a generalisation of Axelrod’s method where norms are represented explicitly and agents can choose their course of action after engaging in what-if reasoning to compare the normative outcomes of alternative actions. This approach is significant because it enables agents to continuously learn and apply their boldness and vengefulness parameters across a variety of scenarios with various norms.
The following is the structure of the paper. Axelrod’s norms and metanorms games are discussed in Sect. 2. Section 3 emphasises the use of explicit norms to encode Axelrod’s mechanism. Section 4 depicts the results of generalisation of Axelrod’s norms and metanorms games, as well as the use of boldness and vengefulness in other scenarios. The prior event calculus models of norms are described in Sect. 5. Section 6 concludes the paper.
2 Background of Axelrod’s Model
Axelrod states that “the extent to how often a given type of action is a norm depends on just how often the action is taken and just how often someone is punished for not taking it”. To understand how cooperation emerges from norms, he developed a game in which players learn the parameters of boldness and vengefulness over generations of the population and can choose to deviate from the norms and metanorms, receiving punishment for their violations [3].
2.1 The Norms Game
Axelrod’s norms game follows an evolutionary model in which successful agent strategies propagate over generations. A strategy is a pair of values representing the agent’s boldness and vengefulness. Each agent has the option of defecting by violating a norm, and there is a chance of being observed by other agents with the probability S, which is drawn individually for each agent from a uniform distribution. Each of the agents has two decisions to make (Fig. 1a).
-
Agents must decide whether to cooperate or defect based on their boldness value (B). A defecting agent (when S < B) receives a Temptation payoff (T = 3) while other agents receive a Hurt payoff (H = \(-1\)). If an agent decides to cooperate, no one’s payoff will change as a result.
-
If an agent observes others defecting (as determined by the S value), the agent decides whether to punish those defectors based on its vengefulness (V) (a probability of punishment). Punishers incur an enforcement cost (E = \(-2\)) every time they punish (P = \(-9\)) a defector.
Axelrod simulated the norms game five times with 100 generations of 20 agents.Footnote 1 Between generations, the utilities of each agent are used to evolve the population of agents. Agents with scores greater than the average population score plus one standard deviation are reproduced twice in a new generation. Agents with a score less than the average population score minus one standard deviation are not reproduced. Other agents are only reproduced once.Footnote 2 The initial values of B and V are chosen at random from a uniform distribution of eight values ranging from 0/7 to 7/7, with the numerator represented as a 3 bit string. During reproduction each bit has a 1% chance of being flipped as a mutation.
2.2 The Metanorms Game
Axelrod found that norms alone were not sufficient to sustain norm compliance in society. He therefore introduced a metanorm to reinforce the practice of punishing defectors. The metanorms game includes punishment for those agents who do not punish defectors after observing them defect (Fig. 1b). Metapunishers incur a meta-enforcement cost (\(E^\prime =-2\)) every time they metapunish (\(P^\prime =-9\)).
3 Generalising Axelrod’s Metanorms Game Using Explicit Norms
To generalise Axelrod’s metanorms game, we provide an explicit representation of norms and a mechanism that can compare alternative actions to determine which will lead to a norm violation. The expectation event calculus (EEC) [5], a discrete event calculus extension, provides this capability.
3.1 The Expectation Event Calculus
The event calculus (EC) consists of a set of predicates that are used to encode information about the occurrence of events and dynamic properties of the state of the world (known as fluents), as well as a set of axioms that interrelate these predicates [14]. This logical language supports various types of reasoning. In this work, we use it for temporal projection. This takes as input a narrative of events that are known to occur (expressed using \( happensAt (E, T)\), where E is an event and T is a time point) and a domain-specific set of clauses defining the conditions under which events will initiate and terminate fluents (expressed using the predicates \( initiates (E,F,T)\) and \( terminates (E,F,T)\)). The EC axioms are then used to infer what fluents hold at each time point. By default, fluents are assumed to have inertia, i. e. they hold until explicitly terminated by an event.
The EC, in general, assumes that time is dense, and time points are ordered using explicit ‘<’ constraints. In this work, we use the discrete event calculus (DEC), which assumes that time points are discrete and identified by integers [11].
The expectation event calculus (EEC) [5] is an extension of the DEC that includes the concepts of expectation, fulfilment, and violation. Expectations are constraints on the future, expressed in a form of linear temporal logic, that the agent wishes to monitor. Expectations are free from inertia and instead are automatically progressed from one state to the next, which means they are partially evaluated and re-expressed in terms of the next time point. During progression, if they evaluate to true or false, a fulfilment or violation is generated.
Figure 2 illustrates temporal projection in the EEC. In addition to the standard features of the DEC, there are two special kinds of fluents: \( exp\_rule \) and \( exp \). A conditional rule to create expectations is expressed by an \( exp\_rule ( Cond , Exp )\) fluent. Here, \( Cond \) is a condition on the past and/or present, while \( Exp \) represents the future expectation. \( Exp \) will be expected if \( Cond \) holds, in which case an \( exp ( Exp )\) fluent is created. In our implementation of the EEC, the condition can test for fluents holding, the occurrence of events (expressed using \( happ (E)\)) and the presence of a symbolic label L in a state (using the expression @L). Complex expressions involving conjunctions and linear temporal logic operators such as next, eventually, always and never can also be used. Labels are associated with time points using \( label(L, T) \) declarations, and are not required to be unique. To distinguish between basic events in the narrative and the inferred fulfilment and violation events, we use the predicate happensAtNarrative to declare the narrative events.
In contrast to the earlier approach of Cranefield [5], we represent fulfilments and violations as events rather than fluents, denoted \( fulf ( Cond , Exp , T , Res )\) and \( viol ( Cond , Exp , T , Res )\), where \( Cond \) and \( Exp \) are the condition and expectation of an expectation rule that was triggered at time T to create the expectation, and \( Res \) is the residual expectation (after being progressed zero or more times since its creation) at the time of its fulfilment or violation.Footnote 3
Our EEC implementation includes a what-if predicate that accepts two alternative event lists (\(E_1\) and \(E_2\)) as arguments and infers the fluents that would hold and the events (including violation and fulfilment events) that would occur if \(E_1\) or, alternatively, \(E_2\) occurred at the current time point. It returns the fluents and events that would occur if the events in \(E_1\) are performed but not those in \(E_2\), and those that would occur if the events in \(E_2\) occur but not those in \(E_1\). This can be used as a basic form of look-ahead to assist an agent in deciding between two alternative sets of actions. In particular, in this work we consider options that are singleton lists and use the what-if predicate to compare which (if any) of two actions will cause one or more expectation violations.
3.2 Modelling Axelrod’s Scenario with the EEC
We model time as a repeated cycle of steps and associate an EEC label with each step. We use the event calculus initiates and terminates clauses to define the effects of events that update an agent’s S value, give payoff to an agent as the outcome of all agents’ cooperate or defect actions, and punish and metapunish agents.
We use the EEC within a simulation platform [6] that integrates Repast Simphony [12] with the EEC through queries to SWI Prolog. This includes an institutional model in which agents take on roles by asserting to the EEC narrative that certain institution-related events have occurred, such as joining an institution and adding a role. Each role has an associated set of conditional rules. A rule engineFootnote 4 is run at the start of selected simulation steps when agents must choose an action and these role-specific rules recommend the actions that are relevant to the agent’s current role based on queries to the EEC, e.g. to check the current step’s label and the fluents that currently hold. Then the agent can run scenario-specific code to select one of the actions to perform.Footnote 5
In contrast to Axelrod’s implicit representation of a norm and metanorm, our explicit representation of a norm implies that three norms are required. In the metanorms game, each action choice is governed by a norm. As there are three choice-points, there are three norms that we model using \({exp\_rule}\) fluents.
As the first-order norm described above is likely to be insufficient to motivate selfish agents to follow the norm and cooperate with others, a second-order norm is required.
This second-order norm is triggered by a sawViolation fluent, which is created when a violation of the first-order norm occurs and a defector is observed. The following initiates clause creates this fluent.
The condition for this clause first determines which agent is responsible for the unfulfilled expectation, then generates a possible observer different from the violator and compares that agent’s S value with a random number to determine whether or not the violation has been observed.
In our application, the violated expectation will include an instantiation of one of the event terms defect(A), punish(A) or metapunish(A), and we can use these to identify the responsible agent. We therefore define the responsible predicate in Prolog as follows.
To encourage the punishment of second-order norm violators, a third-order norm, is required.
Figure 3a, b illustrate the differences between our implementations of the metanorms game with implicit and explicit norms. Figure 3a makes hard-coded action choices, following Axelrod’s algorithm. However, in Fig. 3b, whenever there is an action choice to be made, the two action choices are compared using what-if reasoning that is informed by one of the three norms. For the decision between cooperation or defection, if only one of the action choices will cause norm violation, the agent’s boldness parameter is used to decide whether the violating option (cooperation) is chosen. For decisions between (meta)punishment or no punishment, the vengefulness parameter (V) is used: the violating option is chosen with probability \(1-V\).
The box labelled “Cooperate or defect” indicates a time step in which an agent must decide whether to cooperate or defect based on whether it has a high boldness value (Fig. 3a). In Fig. 3b, the first-order norm’s exp_rule expressing the first-order will have been created in the initial time step and triggered once for each agent, resulting in exp fluents stating that each agent should never defect. Therefore, an agent can use the what-if mechanism to compare the outcomes of the two alternative actions, cooperate or defect. If the agent’s boldness value exceeds S, the agent will violate the first-order norm by defecting; otherwise, the agent will cooperate.
Figure 3a shows how the punishment step of the implicit norm simulation hard-codes the decision to punish each observed defector with a probability given by an agent’s vengefulness parameter. In contrast, the explicit norm representation simulation cycle in Fig. 3b shows how what-if reasoning informed by the explicit second-order norm detects that failure to punish will cause a norm violation. That norm-violating option is then chosen with probability \(1-V\).
In the metapunishment step of the implicit norm simulation, if an agent chose not to punish an observed defector in the punishment step, then every agent with sufficiently high vengefulness will metapunish that agent (Fig. 3(a)). However, when using explicit norms, (Fig. 3b), what-if reasoning detects that failure to metapunish will cause a violation of the explicit third-order norm, and this option will be chosen with probability \(1-V\).
Figure 4 depicts in more detail our use of explicit norms with agents that are aware of norm violations. We have three norms represented by exp_rule fluents, which are triggered at different time steps. After triggering, each exp_rule fluent creates an expectation. The EEC initially clauses generated the exp_rule fluents for the first-order norm (\(N_1\)), second-order norm (\(N_2\)), and third-order norm (\(N_3\)). Each agent has two roles: temptation role and possible punisher role. Table 1 shows what actions an agent can take in the simulation when assigned to a specific role for each step. The temptation role specifies that an agent can choose to cooperate or defect at the cooperate or defect step. At the punishment step an agent with the possible punisher role can choose to punish or do not punish, and an agent with the possible punisher role can choose to metapunish or not to metapunish at the metapunishment step. At the initial time step of the simulation, both roles (temptation role and possible punisher role) are activated for each agent.
The EEC what-if predicate is used to consider two options: cooperate or defect, punish or do not punish, metapunish or do not metapunish (depending on the current step in the simulation cycle), and determine whether one option produces a violation while the other does not. The non-violating option is then chosen (or a random choice if there is no violation). If both options result in a violation, the cost of each violation is calculated (using domain-specific knowledge) and the less costly option is chosen. If the costs are the same, a random selection is made.
At the simulation’s final step, regenerate_population, successful agents are replicated and mutated to form a new generation of the same size [7]. In this simulation, folded outlined arrows represent iteration: one for the three norms and their corresponding expectations within one generation, and the other for 100 generations of simulation.
4 Experiments
Experiment 1: Generalisation of Axelrod’s metanorms game This experiment illustrates that our generalised metanorms game with explicit norms generates punishment and metapunishment events that sustain norm-compliant behaviour. We used 20 agents, 100 generations and 100 runs. Figure 5a shows a scatter plot of the mean boldness and vengefulness values at the end of each run, and we observe that mean boldness is always low and mean vengefulness ranges from high to average.
Figure 5b depicts the vector representation of the same data set.Footnote 6 Vectors show how boldness and vengefulness change across generations in the population. The results show that what-if reasoning with explicit norms causes the first-order norm to be largely upheld in the society due to low boldness being maintained.
Experiment 2: Using the boldness and vengefulness in another scenario Klein [10] introduced a scenario that we refer to as the plain-plateau scenario in our previous work [13]. The scenario depicts a society in which people have the option of living on a river plain with easy access to water, otherwise, they can live on a plateau. Flooding is a risk for river-plain residents. When the government has complete discretionary power, it is in the government’s best interests to compensate citizens whose homes have been flooded by taxing citizens who live on the plateau, creating a prisoner’s dilemma situation. In our previous work, we experimented with the use of social norm-based expectations to achieve coordination where citizen agents are hard-coded to prefer actions that will result in no violation.
Figure 6a illustrates this prior work. Each agent has either a plain dweller or a plateau dweller role, and in each simulation cycle there are two choices: agents can stay on the plateau or move to the plain. In this scenario, we assume there exists a norm that no one should live in the plain and a metanorm stating that plateau dwellers should punish those who live on the plain.
Figure 6b illustrates the application of our generalisation of the metanorms game to this scenario.Footnote 7 This simulates an agent finding itself in a new scenario after having already evolved its personality with respect to norms, and illustrates the generality of our approach using explicit norms and what-if reasoning. In the plain-dweller role, the EEC what-if predicate is used to consider two options: move to the plateau or stay in the plain; similarly, in the plateau-dweller role, the what-if mechanism is used to consider either move plain or stay plateau. When agents with high boldness in the plain-dweller role choose to stay in the plain, they violate the norm. When other vengeful agents observe violators, they punish them, unless insufficient vengefulness causes them to violate the metanorm.
We simulated a group of agents who encounter the plain-plateau scenario after evolving their boldness and vengefulness parameters in the Axelrod scenario. From the first run of Experiment 1, we randomly sampled six (boldness, vengefulness) pairs from personalities of the twenty agents at the end of that run. This resulted in six agents with a boldness of zero.Footnote 8
We used the expectation-aware action-selection mechanism generalised version of Axelrod’s metanorms game with the input norms replaced with those shown at the top of Fig. 6. As before, boldness was used as the probability of choosing a violating action over a non-violating one, and \(1 - \textrm{vengefulness}\) as the probability of choosing to violate an expectation to punish. As all six agents had a boldness of zero, they all opted to live on the plateau (which counts as cooperation in this scenario) whenever they had an opportunity to change their location. Therefore, there were no norm violations and no expectations to punish were created.
While this cooperative behaviour was not surprising given the lack of boldness of the simulated agents, this was an emergent outcome from the transfer of attitudes to normative behaviour learned in the previous scenario, a uniform action-selection mechanism that can be used across scenarios, and the ability to provide new norms as symbolic inputs to inform this mechanism.
5 Prior Event Calculus Models of Norms
This section of the paper reviews some research on the use of event calculus in autonomous agent reasoning to examine the effects of norms.
Alrawagfeh [1] suggests formalising prohibition and obligation norms using event calculus and offers a method for BDI agents to reason about their behaviour at runtime while taking into account the norms in effect at the time and previous actions. Norms are represented by EC rules that initiate fluents with special meanings. The introduced fluents represent punishments for breaking a prohibition norm or failing to fulfil obligation norms, or the rewards for fulfilling obligation norms. The normative reasoning strategy assists agents in selecting the most profitable plan by temporarily asserting to the event calculus the actions that each plan would generate and considering the punishments and/or rewards it would trigger.
In Alrawagfeh’s work, norms cannot be changed dynamically without changing the event calculus rule base, because they are defined by EC initiates clauses. In contrast, in our approach, EC rules can be instantiated automatically from \( exp\_rule \) fluents, which can be changed dynamically by events.
Alrawagfeh has no representation of active norms, violations or fulfilments: only punishments and rewards. In our work, expectation creation, fulfilment, and violation are represented as events, and the what-if predicate compares alternative events to track expectation creation, fulfilment, and violation. We do not assume that rewards and/or punishments will always follow violations and fulfilments; these could be defined by separate \( exp\_rules \) or EC initiates clauses.
Hashmi et al. [8] propose a number of new EC predicates to allow them to model different types of obligation that occur in legal norms. In particular, they introduce a deontically holds at predicate that ensures an obligation enters into force at the same time that the triggering event occurs. In contrast, our approach using the EEC does not necessitate the introduction of a new type of EC predicate in order to initiate a deontic predicate. An \( exp\_rule \) or an expectation can be created with a standard initiates clause and an \( exp \) fluent is created by an \( exp\_rule \) in the state where the condition of the rule becomes true. The EEC does, however, include additional axioms to handle the progression of expectations.
Alrawagfeh and Hashmi et al. both use standard EC, whereas we use discrete EC because this work involves discrete time simulations.
6 Conclusion and Future Work
In previous work [13], we used the EEC what-if mechanism for choosing actions in the presence of expectations. However, we assumed that all agents are compliant and will always choose a non-violating action if possible. The current work removes this assumption, but it also makes the following significant standalone contribution: it generalises Axelrod’s metanorms game to use explicitly represented norms. This allows the metanorms game to be used across multiple scenarios.
Applying our generalised version of Axelrod’s metanorms game to varying scenarios will require changing the mechanism for evolving boldness and vengefulness parameters. Strategy evolution through population regeneration is not realistic for agents that continually evolve their boldness and vengefulness as they move between different scenarios. Therefore, in future work we will investigate the use of a pairwise comparison approach where an agent may adopt another agent”s strategy based on a comparison of their respective fitnesses, e.g. by using the Fermi process [2].
Notes
- 1.
Axelrod used five runs of a hundred generations to simulate the norms and metanorms games. However, we follow the recommendation of [7] and use 100 runs.
- 2.
Axelrod does not state how he maintains a fixed population size after applying these reproduction rules. We follow the approach of [7] involving random sampling when the new population is too large, and random replication when the population is too small.
- 3.
There is also an extended version of the \( exp \) fluent with these four arguments—the version used in this paper has only the residual expectation as its argument.
- 4.
- 5.
At present we assume there are no more than two relevant actions.
- 6.
Populations with similar average levels of boldness and vengefulness are grouped together to create each vector. The end point of each arrow shows the average levels of these features one generation later [3].
- 7.
We do not include a third-order norm for the plain-plateau scenario as this was not part of Klein’s model [10].
- 8.
Run 1 of Experiment 2 ended with only one agent with a non-zero boldness value: 4/7.
References
Alrawagfeh, W.: Norm representation and reasoning: a formalization in event calculus. In: International Conference on Principles and Practice of Multi-Agent Systems, pp. 5–20. Springer (2013)
Altrock, P.M., Traulsen, A.: Fixation times in evolutionary games under weak selection. New J. Phys. 11(1), 013012 (2009)
Axelrod, R.: An evolutionary approach to norms. Am. Polit. Sci. Rev. 80(4), 1095–1111 (1986)
Bicchieri, C.: The Grammar of Society: The Nature and Dynamics of Social Norms. Cambridge University Press (2005)
Cranefield, S.: Agents and expectations. In: International Workshop on Coordination, Organizations, Institutions, and Norms in Agent Systems, pp. 234–255. Springer (2013)
Cranefield, S., Clark-Younger, H., Hay, G.: A collective action simulation platform. In: Multi-Agent-Based Simulation XX: 20th International Workshop, MABS 2019, Montreal, QC, Canada, May 13, 2019, Revised Selected Papers 20, pp. 69–80. Springer (2020)
Galan, J.M., Izquierdo, L.R.: Appearances can be deceiving: Lessons learned re-implementing Axelrod’s evolutionary approach to norms. J. Artif. Soc. Soc. Simul. 8(3) (2005)
Hashmi, M., Governatori, G., Wynn, M.T.: Modeling obligations with event-calculus. In: Rules on the Web. From Theory to Applications: 8th International Symposium, RuleML 2014, Co-located with the 21st European Conference on Artificial Intelligence, ECAI 2014, Prague, Czech Republic, August 18–20, 2014, pp. 296–310. Springer (2014)
Holzinger, K.: The problems of collective action: A new approach. MPI Collective Goods Preprint No. 2003/2, SSRN (2003). https://doi.org/10.2139/ssrn.399140
Klein, D.B.: The microfoundations of rules vs. discretion. Constit. Polit. Econ. 1(3), 1–19 (1990)
Mueller, E.T.: Commonsense Reasoning. Morgan Kaufmann (2006)
North, M.J., Collier, N.T., Ozik, J., Tatara, E.R., Macal, C.M., Bragen, M., Sydelko, P.: Complex adaptive systems modeling with Repast Simphony. Complex Adapt. Syst. Model. 1(1), 3 (2013)
Sengupta, A., Cranefield, S., Pitt, J.: Solving social dilemmas by reasoning about expectations. In: Coordination, Organizations, Institutions, Norms, and Ethics for Governance of Multi-Agent Systems XIV, pp. 143–159. Springer (2022)
Shanahan, M.: The event calculus explained. In: Artificial Intelligence Today, pp. 409–430. Springer (1999)
Thøgersen, J.: Social norms and cooperation in real-life social dilemmas. J. Econ. Psychol. 29(4), 458–472 (2008)
Acknowledgements
This work was supported by the Marsden Fund Council from New Zealand Government funding, managed by Royal Society Te Apārangi.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Sengupta, A., Cranefield, S., Pitt, J. (2023). Generalising Axelrod’s Metanorms Game Through the Use of Explicit Domain-Specific Norms. In: Fornara, N., Cheriyan, J., Mertzani, A. (eds) Coordination, Organizations, Institutions, Norms, and Ethics for Governance of Multi-Agent Systems XVI. COINE 2023. Lecture Notes in Computer Science(), vol 14002. Springer, Cham. https://doi.org/10.1007/978-3-031-49133-7_2
Download citation
DOI: https://doi.org/10.1007/978-3-031-49133-7_2
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-49132-0
Online ISBN: 978-3-031-49133-7
eBook Packages: Computer ScienceComputer Science (R0)