1 Introduction

Many day-to-day economic interactions occur between total strangers. These types of interactions include consumer-to-consumer electronic commerce, peer-to-peer lending and loans, trading goods in online communities, peer-to-peer file-sharing and online real-time multi-player games. However, individuals can learn from their friends’ experiences to help them make “correct” or “informed” decisions.

For example, individuals typically have family members, friends, colleagues, business partners and neighbors (collectively referred to as friends) in the real world. Individuals may participate in online real-time multi-player games against anonymous players and face decisions concerning the amount of effort to put in and whether to continue playing, especially if they are playing in a team role. By communicating with friends, an individual can find out how much effort to devote and determine whether it is worthwhile to continue playing.

The present paper aims to capture these particular phenomena and develops a coevolutionary model of social coordination and matching in networks in which agents interact with strangers and learn from neighbors.

Formally, we consider an even number of agents embedded in an arbitrary, fixed network. Agents are matched in pairs to play a \(2\times 2\) coordination game where there is a conflict between Pareto-efficiency and risk dominance. In each period, three things, which are assumed to be independent, occur. First, all separated agents are randomly matched in pairs. Any two separated agents are matched to each other with strictly positive probability. Second, with a strictly positive probability, if each agent’s payoff is strictly less than the maximum of all neighbors’ payoffs, the agent will imitate the action choice of the most successful neighbor. Third, with a strictly positive probability, when each agent performs strictly worse than the most successful neighbor, the agent will exit the present partnership. A single member can end any partnership unilaterally. Occasionally, agents make mistakes in action revision and partnership updating, and the noise rates in these two processes may be different.

First, we explore the convergence of unperturbed dynamics. With probability one, the unperturbed dynamics will converge to one of the limit states. The underlying intuition is as follows. For simplicity, consider a connected network. Assume that there are only two agents in a partnership choosing the Pareto-efficient action; these two agents receive the highest possible payoff. If there is a unique neighbor of the initial two players of the Pareto-efficient action, then with strictly positive probability, the unique neighbor switches to the Pareto-efficient action and is matched to a player of the risk-dominant action, regardless of whether she has an opportunity for partnership updating. Note that the conflict between Pareto-efficiency and risk dominance implies that the payoff of an agent who chooses the Pareto-efficient action and plays against an agent choosing the risk-dominant action is lowest. The unique neighbor of the initial two players of the Pareto-efficient action receives the lowest possible payoff. As a result, the Pareto-efficient action fails to spread to any other agent. Now, we consider the case in which the initial two players of the Pareto-efficient action have two or more neighbors. With strictly positive probability, two neighbors of the initial two players of the Pareto-efficient action choose the Pareto-efficient action and exit from the present partnership. These neighbors they will be matched to each other in the next period. Following this logic, the Pareto-efficient action may spread to the entire population. The case of the complete network falls into the second case. The logic of the contagion of the Pareto-efficient action can be generalized to certain networks that are not complete, in which a concept of widely connected networks is introduced to identify a particular subset of these networks.

Next, we identify the stochastically stable states of perturbed dynamics when the network is connected and the population size is large. If the network is widely connected, only the limit states where all agents choose the Pareto-efficient action are stochastically stable. The underlying intuition is as follows. When the network is widely connected, in any limit state, all agents play the same action. Consider the transition from a limit state where all agents choose the risk-dominant action to one where all agents choose the Pareto-efficient action. Assume that two agents in a partnership switch from the risk-dominant action to the Pareto-efficient action by mistake. As an implication of wide connectedness, two mutants have two or more neighbors. With positive probability, in the next period, the neighbors of the mutants choose the Pareto-efficient action and are matched to each other if possible. Following this logic, the Pareto-efficient action will spread to the entire population. The discussion also suggests that to induce the reverse transition, for each partnership, one member should mutate to choose the risk-dominant action. As a result, when the population size is larger than six, in the long run, all agents will choose the Pareto-efficient action. In addition, we introduce an example in which all agents are distributed along a line. The line is not widely connected. In this example, there exist polymorphic-action limit states. When partnership mutation is much more likely than action mutation, only the limit states where all agents choose the Pareto-efficient action are stochastically stable. When partnership mutation has the same frequency as action mutation, only the limit states where all agents choose the risk-dominant action are stochastically stable.

Our paper contributes to the literature on learning dynamics in local interaction games.Footnote 1 Many existing papers model strategy revision as myopic best response and show the emergence of the risk-dominant convention [2, 7, 8, 11, 27]. However, there also exists another strand of research in which strategy revision is modeled as imitation. Eshel et al. [14] assume that agents are situated around a circle and play Prisoner’s Dilemma with their two immediate neighbors. They find that imitation of actions with the highest average payoffs might help individuals reach a cooperative outcome. Alós-Ferrer and Weidenholzer [1] assume that agents are situated around a circle and play a coordination game with their neighbors, not necessarily two immediate ones. They show that if interactions are neither global nor limited to the immediate neighbors, the Pareto-efficient convention can be uniquely selected provided that agents follow imitation rules. Although Alós-Ferrer and Weidenholzer [3] consider a coordination game and Alós-Ferrer and Weidenholzer [4] consider a minimum-effort game, they both assume that agents located within an arbitrary fixed network can observe the strategies and performances of agents beyond their interaction neighborhoods and imitate the most successful choices. Using this assumption, they derive sufficient conditions for the efficient convention to be unique selected. Cui [9] studies the case where agents located within an arbitrary, fixed network play a coordination game with their neighbors and imitate the strategies of their better-performing neighbors. For each agent, if one neighbor’s payoff from a specific interaction exceeds her average payoff per interaction, the neighbor is perceived as better performing. He shows that if each agent’s neighborhood is large, in the long run, all agents will choose the Pareto-efficient action. Cui and Wang [10] investigate the situation where in each period, a small proportion of agents located within an arbitrary fixed network are randomly chosen to play a minimum-effort game. The agents learn from their own and neighbors’ experiences and imitate the most successful choices. They show that if each agent’s neighborhood is large, in the long run, all agents choose the highest effort level. As a complement, when agents are located within an arbitrary, fixed network, both the present paper and Khan [25] show that (random) matching and imitation might help agents reach the Pareto-efficient convention.

Furthermore, our paper is related to the literature on random matching in learning dynamics in games. In the seminal paper of Robson and Vega-Redondo [29], a fixed even number of agents are randomly matched in pairs, once every period, to play a coordination game and then choose the strategy with the highest (random) average payoff. They then show that the Pareto-efficient convention is selected. Both the present paper and Khan [25] extend Robson and Vega-Redondo [29] by assuming that agents are located within an arbitrary, fixed network in which agents can only observe the action choices and payoffs of neighbors. Although both papers assume that (1) the observation network is fixed and the interaction network varies and (2) agents use imitation rules to revise actions, this paper is distinguished from Khan [25] by introducing (bounded) rationality into the evolution of the interaction network. In Khan [25], in each period, all agents are randomly matched in pairs; that is, the updating of the interaction network is totally random and does not take into account agents’ (bounded) rationality. By contrast, in our setting, each partnership will survive unless some member has a neighbor performing strictly better than herself, and only the separated agents will be randomly matched at the beginning of the next period. Because of this difference, in Khan [25, the proof of Proposition 1], the Pareto- efficient action can spread from two matched agents to the entire population, and there does not exist any polymorphic-action limit state; by contrast, in our paper, there may exist polymorphic-action limit states when the network is not widely connected.

In addition, our paper also contributes to the literature on the coevolution of games and networks in which individuals choose interaction partners in addition to the action.Footnote 2 The great majority of existing papers model action revision as the myopic best response [18, 19, 21, 31,32,33]. However, there exists another strand of research. In Fosco and Mengel [15], agents learn about action choices and link choices by imitating the successful behavior of others. The present paper falls into this second strand. In particular, we assume that based on payoff comparison with neighbors, agents decide whether to imitate neighbors’ action choices and whether to maintain their present partnerships.

The rest of this paper is organized as follows. Section 2 introduces the basic building blocks of the model. Section 3 studies the convergence of the unperturbed dynamics and introduces the definition of wide connectedness to exclude the polymorphic-action component. Section 4 examines the stochastic stability of the perturbed dynamics in widely connected networks and introduces an example (the line) to show that when the network is not widely connected, there exist polymorphic-action limit states, and the stochastically stable states may be different. Section 5 explores the properties of widely connected networks. Section 6 concludes.

2 Model

2.1 The Base Game

The base game is a symmetric \(2\times 2\) coordination game G. Each player has two actions, A and B. Let \(u(x,x^{'})>0\) be the payoff of a player choosing action x given that the other player is choosing action \(x^{'}\). The payoff function is represented by the following table:

figure a

where \(a>d\), \(b>c\), \(a>b\) and \(b+d>a+c\). Hence, G has two strict Nash equilibria, (AA) and (BB), where (AA) is Pareto efficient and (BB) is risk-dominant. In addition, it is assumed that \(b>d\). Thus, an A-player (a B-player) always strictly prefers to interact with an A-player (a B-player). In summary, we have \(a>b>d>c>0\), where \(d>c\) follows from \(b+d>a+c\).

The game G can be normalized as

figure b

where \(\alpha =(d-c)/(a-c)\) and \(\beta =(b-c)/(a-c)\). We have \(1>\beta>\alpha >0\) and \(\alpha +\beta >1\).

2.2 Local Observation

Let \({\mathbf {N}}= \{1, \ldots ,n\}\) be the set of agents where \(n\ge 4\) is an even number. For any \(i,j\in {\mathbf {N}}\), the relationship between them is captured by a binary variable, \(g_{ij}\in \{0,1\}\). If i and j are able to observe each other’s action and payoff in the last period, \(g_{ij}=1\); otherwise, \(g_{ij}=0\). Although each agent always knows her action and payoff, for clarity, it is assumed that \(g_{ii}=0\). Thus, \(g=\{(g_{ij}) _{i,j\in {\mathbf {N}}}\}\) defines an undirected graph and is referred to as an observation network over \({\mathbf {N}}\).

For any agent \(i\in {\mathbf {N}}\), the neighborhood\(N(i)=\{j\in {\mathbf {N}}:g_{ij}=1\}\) is the set of agents with whom mutual observation exists. Assume that \(N(i)\ne \emptyset \) for any \(i\in {\mathbf {N}}\). Each element in N(i) is called a neighbor. The neighborhood sized(i) is the cardinality of N(i); that is, \(d(i)=|N(i)|\). The notion of neighborhood can be extended as follows. For any nonempty proper subset \({\mathbf {N}}^{'}\subset {\mathbf {N}}\), \(N({\mathbf {N}}^{'})=\{j\in {\mathbf {N}}{\setminus } {\mathbf {N}}^{'}:\exists i\in {\mathbf {N}}^{'} \text { such that }g_{ij}=1\}\).

For any \(i,j\in {\mathbf {N}}\), a path from i to j is a sequence of distinct elements \((i_{1}, \ldots ,i_{L})\) satisfying \(i=i_{1}\), \(g_{i_{L}j}=1\) and \(g_{i_{l}i_{l+1}}=1\) for any \(1\le l\le (L-1)\). When \(i=j\), a circular path is defined. For any \({\mathbf {N}}^{'}\subset {\mathbf {N}}\), \({\mathbf {N}}^{'}\ne \emptyset \), the subgraph \(\big ({\mathbf {N}}^{'},(g_{ij}) _{i,j\in {\mathbf {N}}^{'}}\big )\) is a component if (1) for any \(i,j \in {\mathbf {N}}^{'}\), \(i\ne j\), there is a path from i to j and (2) there is no strict superset \({\mathbf {N}}^{''}\) of \({\mathbf {N}}^{'}\) such that (1) holds. Let \({\mathbf {c}}\) denote a component. Let \({\mathbf {C}}\) denote the set of components. The network g is connected if \({\mathbf {C}}\) is a singleton. The network g is complete if \(g_{ij}=1\) for any two distinct agents \(i,j\in {\mathbf {N}}\).

2.3 Matching

The following definition describes how agents are matched in pairs.

Definition 1

(Matching) The set of unordered pairs \(m=\{(i,j):i,j \in {\mathbf {N}} \text { and } i\ne j\}\) defines a matching over \({\mathbf {N}}\) if for any \(i\in {\mathbf {N}}\), there exists only one agent \(j\in {\mathbf {N}}\) such that \((i,j)\in m\) holds.

Any pair \((i,j)\in m\) will be referred to as a partnership, where agent j will be referred to as the partner of agent i under the matching m. Let \(m_{i}\) be i’s partner under m. The matching m defines a function \(f_{m}: {\mathbf {N}}\rightarrow {\mathbf {N}}\) by setting \(f_{m}(i)=m_{i}\) for any \(i\in {\mathbf {N}}\). The function \(f_{m}\) is a one-to-one correspondence, and \(f_{m}(f_{m}(i))=i\) for any \(i\in {\mathbf {N}}\). Let \({\mathbf {M}}\) denote the set of all possible matchings over \({\mathbf {N}}\).Footnote 3

2.4 Imitation Dynamics and Endogenous Matching

Time is discrete; i.e., \(t=0,1,2, \ldots \). In each period t, each agent i commits to an action \(a_{i}(t)\in \{A,B\}\). Let \({\mathbf {a}}(t)=(a_{1}(t), \ldots ,a_{n}(t))\) denote the action profile. All agents are matched in pairs to play the game G, where the matching is m(t). By interacting with her partner \(m_{i}(t)\), agent i receives a payoff \(u(a_{i}(t),a_{m_{i}(t)}(t))\). In each period t, three things, which are assumed to be independent, occur.

First, all separated agents are randomly matched in pairs. Suppose that each of the possible ways of pairing up all separated agents is chosen with strictly positive probability.

Second, with common probability \((1-p)\), each agent i receives an opportunity to revise her action. The probability p, \(0<p<1\), is an indicator of action inertia. Having the opportunity of action revision, if i’s payoff is strictly less than the maximum of all neighbors’ payoffs, i will switch to the action choice of the most successful neighbor in period \(t+1\); otherwise, i continues to play \(a_{i}(t)\). Formally,

$$\begin{aligned} a_{i}(t+1)=\left\{ \begin{array}{l@{\quad }l} a_{i}(t) &{} \text {if } i\in \displaystyle \text {arg} \max _{k\in N(i)\cup \{i\}} u(a_{k}(t),a_{m_{k}(t)}(t)) \\ a_{j_{0}}(t) &{} \text {otherwise} \\ \end{array} \right. \end{aligned}$$

where \(j_{0}\in \text {arg} \max _{k\in N(i)}u(a_{k}(t),a_{m_{k}(t)}(t))\). Note that \(0<\alpha<\beta <1\). This definition implies that if there are two or more elements in \(\text {arg} \max _{k\in N(i)\cup \{i\}}u(a_{k}(t),a_{m_{k}(t)}(t))\), these agents choose the same action.

Third, with common probability \((1-q)\), each agent i updates her partnership. The probability q, \(0<q<1\), is an indicator of partnership inertia. Having the opportunity of partnership updating, if \(u(a_{i}(t),a_{m_{i}(t)}(t))\ge \max _{k\in N(i)}u(a_{k}(t),a_{m_{k}(t)}(t))\), i is satisfied with the status quo; otherwise, i is dissatisfied. For any partnership \((i,m_{i}(t))\in m(t)\), one member’s dissatisfaction can end the partnership and can leave both i and \(m_{i}(t)\) separated. Thus, to repeat a partnership, when each member has the opportunity of partnership updating, she must have no discontent.

The behavioral rule deserves a brief discussion. There exists inertia in action revision and partnership updating, and whether to revise an action and whether to update the partnership are mutually independent. Thus, agent i may only have the opportunity to revise her action or may only have the opportunity to update her partnership. Furthermore, agents are assumed to be boundedly rational. If \(u(a_{i}(t),a_{m_{i}(t)}(t))\ge \max _{k\in N(i)}u(a_{k}(t),a_{m_{k}(t)}(t))\), agent i will neither alter the action nor update the partnership; otherwise, i will imitate the action choice of the most successful neighbor or exit from the present partnership. When the opportunity of partnership updating is available, each agent updates the partnership following the rule outlined, regardless of whether she has already altered the action. In addition, to repeat the present partnership, mutual consent of both members is required. In other words, either member can break up a partnership unilaterally. The partnership-updating mechanism is analogous to the rule of link formation and deletion in the definition of pairwise stability given by Jackson and Wolinsky [22].

The action-revision and partnership-updating process defines a Markov chain \(\{S(t)\}_{t\in {\mathbb {N}}}\) over \(\varOmega =\big (\prod _{i\in {\mathbf {N}}}\{A,B\}\big )\times {\mathbf {M}}\). We call this process unperturbed dynamics. An absorbing set is a minimal subset of \(\varOmega \) such that once the dynamics reaches the set, the probability of leaving it is zero. An absorbing state is an element that forms a singleton absorbing set. Let \({\bar{\varOmega }}\) be a union of one or more absorbing sets. The basin of attraction\(D({\bar{\varOmega }})\) is the set of states from which the unperturbed dynamics will converge to \({\bar{\varOmega }}\) with probability one.

There exist multiple absorbing sets. For instance, for any \(m\in {\mathbf {M}}\),

$$\begin{aligned} ((\underbrace{A, \ldots ,A}_{n}),m) \text { or } ((\underbrace{B, \ldots ,B}_{n}),m) \end{aligned}$$

is an absorbing state. In fact, given that all agents choose the same action, each agent receives an identical payoff and is satisfied with the status quo. Neither her action nor her partnership will be altered.

To select among all likely absorbing sets, we introduce random noises to identify the long-run equilibria [11, 24, 35]. Suppose that in each period t, when receiving the opportunity of action revision or partnership updating,

  • agent i makes mistakes in action revision with probability \(\epsilon \in (0,1)\);

  • agent i makes mistakes in partnership updating with probability \(\epsilon ^{\lambda }\), \(\lambda >0\);

  • these two processes (of making mistakes) are independent, and random noises are independent across agents and periods.

In the case of action-revision mutation, any element in \(\{A,B\}\) is randomly chosen with strictly positive probability. In the case of partnership-updating mutation, the partnership is broken despite both agents being satisfied with the partnership, or the relationship survives despite one or both agents being unsatisfied. The parameter \(\lambda \) measures the likelihood of partnership-updating mutation relative to that of action-revision mutation. The lower \(\lambda \) is, the more likely that the partnership mutation will occur. This behavior gives rise to perturbed dynamics\(\{S^{\epsilon }(t)\}_{t\in {\mathbb {N}}}\). For any \(\epsilon \in (0,1)\), the Markov chain \(\{S^{\epsilon }(t)\}_{t\in {\mathbb {N}}}\) is irreducible and has a unique invariant distribution \(\mu ^{\epsilon }\). A state \(s\in \varOmega \) is stochastically stable if \(\lim _{\epsilon \rightarrow 0}\mu ^{\epsilon }(s)>0\). Let \(\varOmega ^{*}\) denote the set of all stochastically stable states.

Jackson and Watts [21] model partner choice and action choice as independent processes where only one link may be changed in each period. Furthermore, action-revision mutation occurs with probability \(\epsilon \), and link-updating mutation occurs with probability \(\gamma \), where it is assumed that \(\gamma =f\epsilon \) for some \(f>0\). Thus, action-revision mutation and link-updating mutation vanish at the same rate. As in Jackson and Watts [21], in our model, action revision and partner updating are two independent processes. However, when \(\lambda <1\), partnership-updating mutation vanishes at a lower rate than action-revision mutation; when \(\lambda =1\), partnership-updating mutation and action-revision mutation vanish at the same rate; when \(\lambda >1\), partnership-updating mutation vanishes at a higher rate than action-revision mutation.

3 Convergence of the Unperturbed Dynamics

In this section, we examine the convergence of the unperturbed dynamics. First, we show that the unperturbed dynamics always converges to a limit state where (1) no agent has an incentive to alter her action; (2) at most one A-agent (B-agent) wants to update her partnership; and (3) there exists at most one component such that both actions coexist. Second, we identify a particular class of networks such that for each limit state the coexistence of both actions in any component is impossible.

3.1 General Networks

Consider a component \({\mathbf {c}}\) of the network g. For any \(s\in \varOmega \) and \(a\in \{A,B\}\), let \({\mathbf {I}}_{a}(s,{\mathbf {c}})\) be the set of a-players in \({\mathbf {c}}\) who are motivated to switch to the other action. Let \({\mathbf {P}}_{a}(s,{\mathbf {c}})\) be the set of a-players in \({\mathbf {c}}\) who are willing to continue to choose action a but are dissatisfied with their present partners. Recall that in the game G, \(0<\alpha<\beta <1\). We have \(i\in {\mathbf {I}}_{A}(s,{\mathbf {c}})\) if and only if i plays the game G with a B-player and receives a payoff of 0 and one neighbor plays action B, while any neighbor who plays A also gets a payoff of 0. Formally,

$$\begin{aligned} {\mathbf {I}}_{A}(s,{\mathbf {c}})= & {} \{i\in {\mathbf {c}}: a_{i}=A \text { and } a_{m_{i}}=B; \exists j_{0}\in N(i)\ a_{j_{0}}=B; \nonumber \\&\forall j\in N(i)\ a_{j}=B \text { or } a_{m_{j}}=B \}. \end{aligned}$$
(1)

We have \(i\in {\mathbf {I}}_{B}(s,{\mathbf {c}})\) if and only if one neighbor plays A and receives a payoff of 1. Formally,

$$\begin{aligned} {\mathbf {I}}_{B}(s,{\mathbf {c}})= \{i\in {\mathbf {c}}: a_{i}=B \text { and } \exists j_{0}\in N(i)\ a_{j_{0}}=a_{m_{j_{0}}}=A\}. \end{aligned}$$
(2)

Following from the discussion immediately after the behavioral rule, \(i\in {\mathbf {P}}_{a}(s,{\mathbf {c}})\) if and only if the partner plays the other action and one neighbor receives a payoff of u(aa) and is the most successful among all agents in N(i). Formally,

$$\begin{aligned} {\mathbf {P}}_{A}(s,{\mathbf {c}})= \{i\in {\mathbf {c}}: a_{i}=A \text { and } a_{m_{i}}=B; \exists j_{0}\in N(i)\ a_{j_{0}}=a_{m_{j_{0}}}=A\} \end{aligned}$$
(3)

and

$$\begin{aligned} {\mathbf {P}}_{B}(s,{\mathbf {c}})= & {} \{i\in {\mathbf {c}}: a_{i}=B,a_{m_{i}}=A; \exists j_{0}\in N(i)\ a_{j_{0}}=a_{m_{j_{0}}}=B; \nonumber \\&\forall j\in N(i)\ a_{j}=B \text { or } a_{m_{j}}=B\}. \end{aligned}$$
(4)
Fig. 1
figure 1

Illustration of \({\mathbf {I}}_{a}(\cdot ,\cdot )\) and \({\mathbf {P}}_{a}(\cdot ,\cdot )\) for any \(a\in \{A,B\}\): (1) each node represents an agent; (2) each solid line indicates a link, whereas a dotted line indicates that two end-agents are matched to each other; and (3) a blank (gray) node indicates that the agent plays action A (B)

The following example clarifies the definitions of \({\mathbf {I}}_{a}(\cdot ,\cdot )\) and \({\mathbf {P}}_{a}(\cdot ,\cdot )\) for any \(a\in \{A,B\}\).

Example 1

Assume that \({\mathbf {N}}=\{1, \ldots ,8\}\) and the network g is given by \(g_{i(i+1)}=1\) for any \(1\le i\le 7\) and \(g_{18}\)=1. Thus, g is a circle. Consider the state s illustrated by Fig. 1 where 1 and 2 play action B and are matched to each other, 5 and 6 play action A and are matched to each other, and A-players 3 and 7 are matched to B-players 4 and 8. Each agent in \(\{3,4,7,8\}\) receives a strictly lower payoff than some neighbor. Either of 7 and 8 plays the same action as the most successful neighbor and is not motivated to alter the action. As a result,

$$\begin{aligned} {\mathbf {I}}_{A}(s,{\mathbf {c}})=\{3\},\ {\mathbf {I}}_{B}(s,{\mathbf {c}})=\{4\},\ {\mathbf {P}}_{A}(s,{\mathbf {c}})=\{7\} \text { and } {\mathbf {P}}_{B}(s,{\mathbf {c}})=\{8\}. \end{aligned}$$

The following theorem explores the convergence of the unperturbed dynamics. The proof is provided in “Appendix”.

Theorem 1

The unperturbed dynamics \(\{S(t)\}_{t\in {\mathbb {N}}}\) will converge to one of the limit states with probability one. The state \(s=\big ({\mathbf {a}},m\big )\in \varOmega \) is a limit state if and only if

  1. 1.

    \({\mathbf {I}}_{A}(s,{\mathbf {c}})= {\mathbf {I}}_{B}(s,{\mathbf {c}})=\emptyset \) for any \({\mathbf {c}}\in {\mathbf {C}}\) and \(\sum _{{\mathbf {c}}\in {\mathbf {C}}} |{\mathbf {P}}_{a}(s, {\mathbf {c}})| \le 1\text { for all } a\in \{A,B\}\);

  2. 2.

    there is at most one component \({\mathbf {c}}_{AB}\) such that both actions are played by its members where if \({\mathbf {c}}_{AB}\) exists, \(|{\mathbf {P}}_{A}(s,{\mathbf {c}}_{AB})|=1\).

The unperturbed dynamics will converge to one of the limit states with certainty. Assume that there are only two A-players, \(i_{0}\) and \(j_{0}\), \(g_{i_{0}j_{0}}=0\), and that they are matched to each other. Any agent in \(N(i_{0})\) (\(N(j_{0})\)) receives a strictly lower payoff than \(i_{0}\) (\(j_{0}\)) and is motivated to switch from B to A and end the present partnership. In the next period, with strictly positive probability, all agents in \(N(i_{0})\cup N(j_{0})\) play A and are matched to each other if possible. Following this logic, the Pareto-efficient action A can spread from \(i_{0}\) and \(j_{0}\) to others. When \(|\cup _{i\in \{i_{0},j_{0}\}}\big (\{i\}\cup N(i)\big )|\) is odd, one A-player \(i_{1}\) is matched to a B-player \(j_{1}\). Furthermore, if \(N(\cup _{i\in \{i_{0},j_{0}\}}\big (\{i\}\cup N(i)\big ))\) is a singleton consisting of \(j_{1}\), A fails to spread from \(i_{1}\) to \(j_{1}\). Then, a polymorphic-action limit state is reached. In addition, for any \(a\in \{A,B\}\), the number of a-players who are discontent with their present partners never exceeds one; otherwise, these a-players will be matched to each other.

When the polymorphic-action component \({\mathbf {c}}_{AB}\) exists, there are two distinct agents \(i_{0}\) and \(j_{0}\) in \({\mathbf {c}}_{AB}\), \(g_{i_{0}j_{0}}=1\), such that \(i_{0}\) is an A-player and \(j_{0}\) is a B-player. As a B-player, \(j_{0}\) receives a payoff of \(\alpha \) or \(\beta \) that is strictly less than 1. Agent \(i_{0}\) is matched to a B-player and receives a payoff of zero, which is strictly less than \(\alpha \) or \(\beta \); otherwise, \(j_{0}\) is motivated to alter the action. This implies that \(i_{0}\) is the unique A-player who is motivated to exit from the present partnership. Note that \(i_{0}\) is not motivated to change her action. Therefore, for agent \(i_{0}\), one neighbor plays A and is matched to an A-player. That is, the number of A-players is greater than three.

Fig. 2
figure 2

Illustration of two limit states

As an illustration, the following example presents two limit states where for one of them, there is a polymorphic-action component.

Example 2

Assume that the network g consists of two components:

$$\begin{aligned} {\mathbf {c}}_{1}= & {} \big (\{1,2,3,4,5\},g_{12}=g_{23}=g_{34}=g_{45}=1\big ),\\ {\mathbf {c}}_{2}= & {} \big (\{6,7,8\}, g_{67}=g_{78}=1\big ). \end{aligned}$$

Figure 2 illustrates two limit states in g. In the left panel, all members in each component choose the same action. Only agent 3 is motivated to end the present partnership because she receives a lower payoff than one neighbor 4. In the right panel, \({\mathbf {c}}_{1}\) is the polymorphic-action component. Due to the poor performance of agent 2, the Pareto-efficient action A fails to spread to all members of \({\mathbf {c}}_{1}\). However, 2 has no incentive to alter action because the most successful neighbor 1 chooses A. Only 2 and 3 are willing to end the present partnerships. Following the partnership-updating mechanism, they continue to be matched.

3.2 Wide Connectedness

When the network g is complete, there is no polymorphic-action limit state. If both actions coexist, the discussion following Theorem 1 implies that there exist two A-players, denoted by \(i_{0}\) and \(j_{0}\), who are matched to each other. Note that a link exists between any two distinct agents. Each B-player has one neighbor \(i_{0}\) or \(j_{0}\) who receives the highest possible payoff by choosing A and is motivated to switch to A. This yields a contradiction.

A fascinating question that arises is whether this conclusion can be extended to more general networks? Fortunately, a positive answer can be given. We now present a particular class of networks that exclude polymorphic-action limit states.

Definition 2

(Wide connectedness) A component \({\mathbf {c}}=\big ({\mathbf {N}}^{'},(g_{ij}) _{i,j\in {\mathbf {N}}^{'}}\big )\) is widely connected if either \(|{\mathbf {N}}^{'}|=2\) or for any proper subset \({\mathbf {N}}^{''}\subset {\mathbf {N}}^{'}\), \(N({\mathbf {N}}^{'}{\setminus } {\mathbf {N}}^{''})\) is not a singleton when \({\mathbf {N}}^{''}\ge 2\). The network g is widely connected if g is connected and the unique component is widely connected.

Consider a component \({\mathbf {c}}=\big ({\mathbf {N}}^{'}, (g_{ij}) _{i,j\in {\mathbf {N}}^{'}}\big )\) where \(|{\mathbf {N}}^{'}|\ge 3\). Informally, wide connectedness implies that for any partition \(\{{\mathbf {N}}^{'}_{1}, {\mathbf {N}}^{'}_{2}\}\) of \({\mathbf {N}}^{'}\), when \(|{\mathbf {N}}^{'}_{1}|\ge 2\), there are at least two agents in \({\mathbf {N}}^{'}_{1}\) who have at least one link with agents in \({\mathbf {N}}^{'}_{2}\). Formally,

$$\begin{aligned} \sum _{i\in {\mathbf {N}}^{'}_{1}}\max _{j\in {\mathbf {N}}^{'}_{2}}g_{ij}\ge 2. \end{aligned}$$

Furthermore, when \({\mathbf {N}}^{'}_{1}={\mathbf {N}}^{'}{\setminus }\{i\}\) for any \(i\in {\mathbf {N}}^{'}\), wide connectedness implies that i has at least two neighbors.

The following example shows that the circle is widely connected.

Example 3

Assume that agents are spatially distributed around a circle. Formally, for any agent \(i\in {\mathbf {N}}=\{1, \ldots ,n\}\), \(N(i)=\{i-1,i+1\}\) where \(i\pm 1\) is understood as modulo n. As an illustration, see Fig. 1. This network is widely connected. For any nonempty proper subset \({\mathbf {N}}^{'}\subset {\mathbf {N}}\), \(N({\mathbf {N}}{\setminus } {\mathbf {N}}^{'})\) contains two or more agents.

The following example illustrates a network that is connected but not widely connected.

Example 4

Assume that there are 2m, \(m\ge 2\), agents; that is, \(n=2m\). The network g is defined as follows: \(g_{ij}=1\) for any distinct agents \(i,j\in \{1, \ldots ,m\}\) and \(i,j\in \{m+1, \ldots ,2m\}\), and \(g_{m(m+1)}=1\). This network is not widely connected. In fact, \(N({\mathbf {N}}{\setminus }\{1, \ldots ,m\})=N(\{m+1, \ldots ,2m\})=\{m\}\).

Examples 3 and 4 show that wide connectedness has no relationship with the minimum degree among all agents. Example 3 illustrates a widely connected network where \(d(i)=2\) for any \(i\in {\mathbf {N}}\), whereas Example 4 presents a connected but not widely connected network where \(\min _{i\in {\mathbf {N}}}d(i)\), which is \(m-1\), can be as large as possible when m goes to infinity.

Informally, when there are two members in a component \({\mathbf {c}}\) who play A and are matched to each other, wide connectedness of \({\mathbf {c}}\) implies that the Pareto-efficient action A is able to spread from these two agents to all members in \({\mathbf {c}}\). The following theorem shows that wide connectedness excludes the polymorphic-action limit states. The proof is provided in “Appendix”.

Theorem 2

If a component is widely connected, then all members choose the same action in any limit state.

If a component \({\mathbf {c}}\) is widely connected, for any limit state, the coexistence of both actions is prevented. The intuition is as follows. For a limit state s, when \({\mathbf {c}}\) is the polymorphic-action component, (1) at least two members play action A, and (2) \({\mathbf {P}}_{A}(s,{\mathbf {c}})\) is a singleton. Following from Definition 2, the set of B-players in \({\mathbf {c}}\) has two or more neighbors from the set of A-players in \({\mathbf {c}}\). However, the neighborhood of the set of B-players in \({\mathbf {c}}\) coincides with \({\mathbf {P}}_{A}(s,{\mathbf {c}})\). This result contradicts the fact that \({\mathbf {P}}_{A}(s,{\mathbf {c}})\) is a singleton.

As an immediate result of Theorem 2, we have the following corollary.

Corollary 1

Assume that each component of the network g is widely connected. Then, in any limit state, there exists no component such that both actions coexist.

Wide connectedness denies the possibility of the coexistence of both actions in any component for each limit state. In other words, when the network is widely connected, in any limit state, all members of each component choose the same action.

4 Stochastic Stability of the Perturbed Dynamics

In this section, we identify the stochastically stable states of the perturbed dynamics. Without loss of generality, assume that the network is connected. First, we study the case of widely connected networks. Then, we explore an example where agents are distributed along a line, i.e., a network that is not widely connected.

For clarity, we introduce the following notation. For any \(a\in \{A,B\}\), let \(\varOmega ^{a}\subset \varOmega \) denote the set of states where all agent play action a. Formally,

$$\begin{aligned} \varOmega ^{a}= \{((\underbrace{a, \ldots ,a}_{n}),m): m\in {\mathbf {M}}\} \text { for any } a\in \{A,B\}. \end{aligned}$$

Each element in \(\varOmega ^{a}\) is a limit state of the unperturbed dynamics \(\{S(t)\}_{t\in {\mathbb {N}}}\), irrespective of the architecture of the network g. Let \(\varOmega ^{AB}\subset \varOmega \) be the set of limit states where both actions A and B coexist.

4.1 Perturbed Dynamics in Widely Connected Networks

Given that the network g is widely connected, the following theorem characterizes the stochastically stable states of the perturbed dynamics.

Theorem 3

Assume that the network g is widely connected. Consider the perturbed dynamics \(\{S^{\epsilon }(t)\}_{t\in {\mathbb {N}}}\), \(\epsilon \in (0,1)\). Then,

$$\begin{aligned} \varOmega ^{*}=\left\{ \begin{array}{l@{\quad }l} \varOmega ^{A}\cup \varOmega ^{B} &{} \text {if } n=4;\\ \varOmega ^{A} &{} \text {otherwise}. \\ \end{array} \right. \end{aligned}$$

Proof

Following from Corollary 1, the set of all limit states of the unperturbed dynamics \(\{S(t)\}_{t\in {\mathbb {N}}}\) is \(\varOmega ^{A}\cup \varOmega ^{B}\).

First, the following lemma investigates the transition from \(\varOmega ^{B}\) to \(\varOmega ^{A}\). The proof is in “Appendix”.\(\square \)

Lemma 1

Assume that g is widely connected. Starting from any limit state in \(\varOmega ^{B}\), after two action mutations followed by the unperturbed dynamics, \(\{S^{\epsilon }(t)\}_{t\in {\mathbb {N}}}\) will arrive at a limit state in \(\varOmega ^{A}\) with positive probability.

Then, the following explores how to traverse all limit states in \(\varOmega ^{a}\) for any \(a\in \{A,B\}\). The proof is in “Appendix”.

Lemma 2

Assume that g is widely connected. For any two distinct limit states \(s, s^{'}\in \varOmega ^{a}\), \(a\in \{A,B\}\), there is a sequence \((s_{0}, \ldots ,s_{\ell })\) of distinct elements in \(\varOmega ^{a}\) such that (1) \(s_{0}=s\) and \(s_{\ell }=s^{'}\); and (2) for any \(0\le \iota \le (\ell -1)\), to transit from \(s_{\iota }\) to \(s_{\iota +1}\), two partnership mutations are required when \(\lambda \le 1\), and two action mutations are required when \(\lambda >1\).

Following from Lemmas 1 and 2, for any \(s\in \varOmega ^{A}\), when \(\lambda \le 1\), the minimum resistance among all s-trees is

$$\begin{aligned} 2 \lambda (|{\mathbf {M}}|-1)+2+2 \lambda (|{\mathbf {M}}|-1) \end{aligned}$$

where the first term is the resistance of transition from other limit states in \(\varOmega ^{B}\) to one state \(s^{'}\) in \(\varOmega ^{B}\); the second term is the resistance of transition from \(s^{'}\) to a state in \(\varOmega ^{A}\); and the third term is the resistance of transition from other limit states in \(\varOmega ^{A}\) to s; similarly, when \(\lambda >1\), the minimum resistance among all s-trees is \(2 (|{\mathbf {M}}|-1)+2 |{\mathbf {M}}|\).Footnote 4 However, the Proof of Lemma 1 shows that to transit from a limit state in \(\varOmega ^{A}\) to any limit state in \(\varOmega ^{B}\), 0.5n action mutations of resistance 0.5n are required. As a result, if \(n\ge 6\), the set of stochastically stable equilibria is \(\varOmega ^{A}\); if \(n=4\), all limit states are stochastically stable. \(\square \)

Consider the transition from \(\varOmega ^{B}\) to \(\varOmega ^{A}\). When one agent and her partner play action A by mistake, they both receive the highest possible payoff of 1 and are never motivated to switch to action B or end the partnership between them. Then, in the next period, with strictly positive probability, all neighbors of these two agents play action A and are matched to one another if possible, where at most one A-player is matched to a B-player. Note that the network g is widely connected. This widely connected network implies that when the set of B-players is not empty, some B-players observe A-players who are matched to A-players and are motivated to switch to A. Following the same logic, the Pareto-efficient action A can spread from two agents in a partnership to the entire population. Therefore, two action mutations can induce the transition from \(\varOmega ^{B}\) to \(\varOmega ^{A}\). The reasoning also implies that to transit from \(\varOmega ^{A}\) to \(\varOmega ^{B}\), there should be at least one B-player in any partnership, and thus at least 0.5n action mutations are required. According to the radius–coradius theorem in Ellison [12], the radius of \(\varOmega ^{A}\) is at least 0.5n, and its coradius is 2. If \(n\ge 6\), in the long run, only limit states from \(\varOmega ^{A}\) will be observed with strictly positive probability.

Next, consider the transition from \(((a, \ldots ,a),m)\) to \(((a, \ldots ,a),m^{'})\) for any \(a\in \{A,B\}\), \(m\ne m^{'}\), provided that \(\lambda \le 1\). When agents i and \(m^{'}_{i}\), \(m_{i}\ne m^{'}_{i}\), receive the opportunity of partnership updating, they end their present partnerships by mistake. In the next period, with strictly positive probability, i and \(m^{'}_{i}\) are matched to each other. After two partnership mutations, another limit state \(s^{''}\in \varOmega _{a}\) is arrived at where, in comparison with s, \(s^{''}\) has at least one more partnership in common with \(s^{'}\). Thus, any two limit states in \(\varOmega _{a}\) are accessible through transitions between elements in \(\varOmega _{a}\), where two partnership mutations are required for each transition. According to the least-resistance tree construction, when \(n=4\), all limit states are stochastically stable.

4.2 Perturbed Dynamics in a Line

When the network g is connected but not widely connected, there may exist polymorphic-action limit states. The transition from \(\varOmega ^{A}\) to \(\varOmega ^{B}\) may proceed by passing through polymorphic-action limit states. In this subsection, we introduce an example to explore the perturbed dynamics in nonwidely connected networks provided that n is large.

Suppose that all agents are distributed along a line\(g^{l}\). Specifically, \(g^{l}_{i(i+1)}=1\) for any \(1\le i\le n-1\) and \(g^{l}_{ij}=0\) otherwise. Note that there is only one link between \(\{1, \ldots ,i\}\) and \(\{i, \ldots ,n\}\) for any \(1\le i\le n\). The network \(g^{l}\) is not widely connected.

As in the case of widely connected networks, for any \(a\in \{A,B\}\) and \(m\in {\mathbf {M}}\), \(((a, \ldots ,a),m)\) is a monomorphic-action limit state. As an illustration, see Fig. 3.

Fig. 3
figure 3

An illustration of two monomorphic-action limit states

In addition to the monomorphic-action limit states in \(\varOmega ^{A}\cup \varOmega ^{B}\), there exist polymorphic-action limit states. Following from Theorem 1, for each polymorphic-action limit state, there exists only one A-player \(i_{0}\) such that (1) one neighbor plays B and one neighbor plays A and (2) \(i_{0}\) is matched to a B-player. The set of A-players is either \(\{1,2, \ldots , i_{0}\}\) or \(\{i_{0}, \ldots ,n\}\). Each A-player except \(i_{0}\) is matched to an A-player; otherwise, there is another A-player besides \(i_{0}\) who is motivated to end the present partnership, or \(i_{0}\) is motivated to switch to B. As an illustration, see Fig. 4.

Fig. 4
figure 4

An illustration of two polymorphic-action limit states

In the remaining part of this subsection, we identify the stochastically stable states of the perturbed dynamics in the line. For clarity, we focus on the case where in any polymorphic-action limit state the set of A-players is \(\{1,2, \ldots , i_{0}\}\).

First, as in the case of widely connected networks (Lemma 2), any two distinct limit states in \(\varOmega ^{a}\), \(a\in \{A,B\}\), are accessible via a path of elements in \(\varOmega ^{a}\) where each transition between two neighboring elements requires two action or partnership mutations, and they have the same stochastic stability.

Second, consider the transition from \(\varOmega ^{A}\) to \(\varOmega ^{AB}\). Suppose that in the initial state, \(n-1\) and n are matched to each other. As an illustration, see the top panel of Fig. 3. Assume that n switches from A to B by mistake. Then, a polymorphic-action limit state with one single B-player is reached.

Third, consider the transition from \(\varOmega ^{B}\) to \(\varOmega ^{AB}\). In any polymorphic-action limit state, there exist two A-players who are matched to each other. The inequality \(0<\alpha<\beta <1\) implies that at least two action mutations are required to induce the transition from \(\varOmega ^{B}\) to \(\varOmega ^{AB}\). Suppose that in the initial state, \(m(i)=i+1\) for any odd i. Assume that 1 and 2 switch from B to A by mistake. In the next period, 1 and 2 receive the highest possible payoff of 1. Then, 3 will play A by imitating the action choice of the most successful neighbor 2. A polymorphic-action limit state with three A-players is reached. The discussion also shows that to leave the basin of attraction of\(\varOmega ^{B}\), two action mutations are required. In addition, two action mutations by 1 and n can induce the transition from \(\varOmega ^{B}\) to \(\varOmega ^{A}\) provided that in the initial state, \(m(i)=n+1-i\) for any i.

Next, consider the transition from \(\varOmega ^{AB}\) to \(\varOmega ^{A}\). Suppose that in the initial state, \(m(i)=i+1\) for any odd i. As an illustration, see the bottom panel of Fig. 4. Assume that \(i_{0}-1\) ends the present partnership by mistake. In the next period, with positive probability, \(i_{0}-1\) and \(i_{0}\) are matched to each other and receive the highest possible payoff of 1, while \(i_{0}+1\) and \(i_{0}-2\) are matched to each other. Then, \(i_{0}+1\) will switch to A because the most successful neighbor \(i_{0}\) chooses A, which implies that \(i_{0}+2\) will switch to A too. A polymorphic-action limit state with two more A-players will be reached. This discussion suggests that any polymorphic-action limit state is connected to a limit state in \(\varOmega ^{A}\)via a path of elements in\(\varOmega ^{AB}\), where each transition requires one single partnership mutation.

Now, consider the transition from \(\varOmega ^{AB}\) to \(\varOmega ^{B}\). Suppose that in the initial state, \(m(i)=i+1\) for any odd i. As an illustration, see the bottom panel of Fig. 4. There are two ways to complete the transition. In the first, assume that \(i_{0}-1\) switches from A to B by mistake. Then, \(i_{0}\) will switch to B because both neighbors play B. After one single action mutation, a polymorphic-action limit state with two more B-players will be reached. In the second, assume that A-player \(i_{0}-2\) ends the partnership with A-player \(i_{0}-1\) by mistake and that B-player \(i_{0}+2\) ends the partnership with B-player \(i_{0}+3\) by mistake. In the next period, with strictly positive probability, A-player \(i_{0}-1\) (\(i_{0}-2\)) is matched with B-player \(i_{0}+3\) (\(i_{0}+2\)). Then, \(i_{0}\) will switch to B because the most successful neighbor \(i_{0}+1\) chooses B, which implies that \(i_{0}-1\) will switch to B too. After two partnership mutations, a polymorphic-action limit state with two moreB-players will be reached. Note that when there is only one B-player in the initial state, only the first route is possible.

When \(\lambda <1\), the partnership mutation is more likely than the action mutation. The above discussion shows that in this case, the resistance of the transition from a polymorphic-action limit state to another polymorphic-action limit state with more A-players is \(\lambda \), whereas the resistance of the transition in the reverse direction is \(\min \{2\lambda , 1\}\). Note that \(\min \{2\lambda , 1\}>\lambda \). Therefore, for any \(s\in \varOmega ^{A}\) and \(s^{'}\in \varOmega ^{B}\), the least resistance of all s-trees is strictly less than the least resistance of all \(s^{'}\)-trees when n is large. The set of stochastically stable states is \(\varOmega ^{A}\) when \(\lambda <1\).

When \(\lambda =1\), the partnership and action mutations occur with the same frequency. In this case, both the resistance of the transition from a polymorphic-action limit state to another polymorphic-action limit state with more A-players and the resistance of the transition in the reverse direction are 1. However, leaving the basin of attraction of \(\varOmega ^{A}\) requires one single action mutation, whereas leaving the basin of attraction of \(\varOmega ^{B}\) requires two action mutations. Thus, the set of stochastically stable states is \(\varOmega ^{B}\) when \(\lambda =1\).

5 Discussion

Wide connectedness plays an important role in the previous two sections. Now, we examine the properties of widely connected networks.

To facilitate the analysis, we first present the following definition.

Definition 3

(Tree) The network g is a tree if g is connected and there is no circular path.

An equivalent definition is as follows. A network g is a tree if for any \(i,j\in {\mathbf {N}}\), when \(g_{ij}=1\), \(g-ij\) is no longer connected, where \(g-ij\) is the network obtained from g by deleting the link between i and j. Let \(g^{t}\) denote a tree.

The line \(g^{l}\) given in Sect. 4.2 is a tree. Recall that \(g^{l}\) is not widely connected. The following lemma shows that this conclusion can be extended to any tree.

Lemma 3

Any tree is not widely connected.

The proof is trivial. For any \(i,j\in {\mathbf {N}}\), if \(g^{t}_{ij}=1\), the network \(g^{t}-ij\) consists of two components with one component \(({\mathbf {N}}^{'},(g_{ij}) _{i,j\in {\mathbf {N}}^{'}})\) satisfying \(|{\mathbf {N}}^{'}|\ge 2\). It is straightforward to conclude that either \(N({\mathbf {N}}{\setminus } {\mathbf {N}}^{'})=\{i\}\) or \(N({\mathbf {N}}{\setminus } {\mathbf {N}}^{'})=\{j\}\) holds. Definition 2 implies that \(g^{t}\) is not widely connected.

Example 3 shows that the circle is widely connected. One may question that the tree is not widely connected because of the nonexistence of a circular path. To address this issue, consider the following example.

Fig. 5
figure 5

A polymorphic-action limit state in a network containing a circular path that is not widely connected

Example 5

Consider a society comprising six agents. Assume that the network g is specified as \(g_{i(i+1)}=1\) for any \(i=1, \ldots ,5\); \(g_{13}=1\) and \(g_{ij}=0\) otherwise. Agents 1, 2, 3 and the links between them define a circular path. It is straightforward to determine that \(N(\{4,5,6\})= \{i\in \{1,2,3\}:\exists j\in \{4,5,6\} \text { s.t. } g_{ij}=1 \}=\{3\}\). That is, g is not widely connected. Figure 5 illustrates a polymorphic-action limit state in g.

Example 5 presents a connected network that is neither a tree nor widely connected. This network implies that there is no dichotomy between widely connected networks and trees.

However, the following lemma shows that if there is a circular path passing through all agents, g is widely connected. The proof is trivial, and we omit it here.

Lemma 4

If there exists a circular path passing through all elements of \({\mathbf {N}}\), g is widely connected.

Note that any network containing a circular path passing through all elements of \({\mathbf {N}}\) is denser than the circle. Lemma 4 can also be extended further. The proof is trivial, and we omit it here.

Lemma 5

Consider two networks g and \(g^{'}\) where for any \(i,j\in {\mathbf {N}}\), \(g_{ij}=1\) implies that \(g^{'}_{ij}=1\). Then, if g is widely connected, \(g^{'}\) is widely connected.

Lemma 4 only provides a sufficient condition for wide connectedness. To clarify this point, consider the following example.

Example 6

Assume that the network is a core-periphery network \(g^{cp}\). Formally, there is a partition \(\{{\mathbf {N}}_{1},{\mathbf {N}}_{2}\}\) of \({\mathbf {N}}\) such that \(N(i)={\mathbf {N}}{\setminus } \{i\}\) for any \(i\in {\mathbf {N}}_{1}\) and \(N(j)={\mathbf {N}}_{1}\) for any \(j\in {\mathbf {N}}_{2}\). That is, \({\mathbf {N}}_{1}\) (\({\mathbf {N}}_{2}\)) is the set of core (periphery) agents.Footnote 5 For any core-periphery network \(g^{cp}\), there is no circular path passing through all agents provided that \(|{\mathbf {N}}_{1}|<|{\mathbf {N}}_{2}|\). However, it is straightforward to see that when \(2\le |{\mathbf {N}}_{1}|<|{\mathbf {N}}_{2}|\), \(g^{cp}\) is widely connected.Footnote 6 Figure 6 illustrates a limit state where all agents play A.

Fig. 6
figure 6

A limit state where all agents play action A in a core-periphery network

6 Concluding Remarks

We consider the coevolution of social coordination and matchings. Agents are located within an arbitrary, fixed network and are matched in pairs to play a \(2\times 2\) coordination game. In each period, by comparing their own and neighbors’ payoffs, agents decide whether to imitate neighbors’ actions and whether to maintain their present partnerships. There is inertia in both action revision and partnership updating. For any partnership, one member’s unwillingness can end it and can separate the two members. All separated agents are randomly matched in pairs at the beginning of the next period. Occasionally, agents make mistakes in action revision and partnership updating.

We first characterize the limit states of the unperturbed dynamics. To exclude the possibility that agents in a component play different actions in a limit state, we introduce the definition of wide connectedness. Wide connectedness of a component implies that for any proper subset of two or more members, two or more elements have links to elements in the complement of the subset. A network is widely connected if it is connected and the unique component is widely connected. Given that the network is widely connected, if the size of the population is four, all monomorphic limit states are stochastically stable; if the size is strictly larger than four, only the limit states where all agents choose the Pareto-efficient action are stochastically stable. We also introduce an example, the line, to show that when the network is not widely connected, the likelihood of partnership mutation relative to that of action mutation matters, and stochastically stable states may be different.

When the network is widely connected, the Pareto-efficient action can spread from two agents in a partnership to the entire population (Lemma 1 in the Proof of Theorem 3). Therefore, the main results can be generalized to the case that agents are matched in pairs to play a 2-person symmetric game where there is a Nash equilibrium strictly Pareto-dominating any other strategy profile. One example is a 2-person minimum-effort game. Another example is a 2-person symmetric game satisfying that there exists a strictly dominant strategy such that when all players play the strictly dominant strategy, each player receives the highest possible payoff.

There are several natural extensions to the research presented here. First, as in Alós-Ferrer and Weidenholzer [3, 4], agents are assumed to be optimistic and to focus on the highest observed payoffs only. It would be desirable to consider other measures of the performance of each action; for instance, agents focus on the lowest observed payoffs of each action only. Second, agents are assumed to be matched in pairs to play a 2-person coordination game. We would be encouraged to consider the case in which agents are matched in groups to play an n-person symmetric game in which \(n\ge 3\). In this case, any proper subgroup of two or more agents may be tempted to regroup to play a “smaller” game. Third, it would be interesting to assess whether or not the main results are supported by laboratory experiments.