Keywords

1 Introduction

Human and artificial agents take into account not only shared social norms, but also complex institutional systems. We are often faced with systems of this kind in our daily life (the legal system, but also the prescriptions of an institutionalised religion, or the regulations of a company, a condominium, a regulated market, a teaching institution, a sociotechnical infrastructure such as an airport or a harbour, etc.). Most norms in such systems are created by norm-creating acts of the regulators of such systems (public or private authorities), and their content is to a large extent dependent upon the discretionary choice of the regulator. Moreover such norms regulate very specific and differentiated situations, with which most agents do not have previous acquaintance. Thus the precise content of such norms cannot be derived from shared values and attitudes, nor can be induced from social behaviour. Moreover, an institutional system can contain a huge number of such norms, up to hundreds of thousand, so that it exceeds the storage capacity of the human mind. Such norms are also subject to frequent change and to multiple interpretations, so that even if they could be stored in a single repository, the repository would soon become useless unless continuously updated.

This means that for bounded agents, the way of learning the content of a complex institutional system must differ from the way of learning social norms.

When we learn social norms we permanently store them in our memory, as the content of appropriate normative beliefs and goals, so that they can directly govern our behaviour. On the contrary, we do cannot learn and store in our memory most norms included in large normative systems. We rather possess some ideas about the existence of such a system and the ways to identify its content. When needed, we collect some fragmentary information about the system and combine this information with the relevant facts, both tasks being often delegated to experts. On the basis of this information we can conclude that the system requires us to perform certain actions.

When referring to a large normative system N an agent usually does not immediately find an answer to the question “What ought I to do?” (as it usually happens when applying a shared social norm the agent is endorsing). The agent rather needs to asks itself (or the appropriate expert) “What does N require from me?,” i.e., “What ought I do to according to N?” Agents may have to deal with different normative systems and distinguish the requests provided by each one of them.

I will propose an argumentation approach, which takes a normative system and the relevant facts as inputs, in order to deliver such answers.

2 Preliminary Notions: Actions, Obligations, Norms

For reasoning with normative systems, we need some basic notions. First, a way of expressing action and obligations is required. For actions I will use the simple E operator of Pörn (1977) (on the E operator see also Sergot 2001), though other action logics, such as STIT (Belnap et al. 2001), would be appropriate as well.

Definition 1 (Actions).

Let proposition E j ϕ describe agent j’s action consisting in the production of state of affairs ϕ, where “ϕ” is any proposition. Thus E j ϕ means “j brings it about that ϕ”. The non-accomplishment of an action is therefore described by ¬E j ϕ, i.e., j does not bring about that A.

For simplicity, when an agent brings about its own action, I will not repeat the agent’s name in the action’s result. Thus, for expressing the idea that John smokes (John brings it about that he smokes) rather than writing E John Smoke(John), I will write E John Smoke.

As an example of an action-proposition, consider the following

$$\displaystyle{ \mathsf{E}_{John}Damaged(Tom) }$$

which means “John brings it about that Tom is damaged”, or more simply “John damages Tom” while the following

$$\displaystyle{ \neg \mathsf{E}_{John}Damaged(Tom) }$$

means “John does not bring it about that Tom is damaged”, or “John does not damage Tom”. I shall adopt the logic of E, which is a classical modal logic (if A and B are logically equivalent, then E x A ↔ E x B) including the axiom schema:

$$\displaystyle{ \mathsf{E}_{x}\phi \Rightarrow \phi }$$

meaning that if the state of affairs ϕ is realised though an action, then it is the case that ϕ. For instance, the fact that Tom makes it so that Ann suffers damage, obviously entails that Ann suffers damage:

$$\displaystyle{ \mathsf{E}_{Tom}Damaged(Ann) \Rightarrow Damaged(Ann) }$$

Besides an action logic E, I need a deontic logic to express obligations.

Definition 2 (Obligations and prohibitions).

Let O denote obligation. O E j ϕ means “it is obligatory that j brings it about that ϕ”. Similarly O¬E j ϕ means “it is obligatory that j does not bring about that ϕ”, or “it is forbidden that j brings about that ϕ”.

For instance, the following means “it is obligatory that John makes it so that Tom is compensated”, or more simply, “it is obligatory that John compensates Tom”,

$$\displaystyle{ O\mathsf{E}_{John}Compensated(Tom) }$$

while the following means “it is forbidden that John damages Tom”.

$$\displaystyle{ O\neg \mathsf{E}_{John}Damages(Tom) }$$

As usual, I take permission to be the negation of prohibitions. I will not endorse here a particular deontic logic, since the following considerations may apply to different deontic logics. The reader may assume, for instance, standard deontic logic, which is a normal modal logic including, besides all tautologies of propositional logic, definition Df P. P ϕ ↔ ¬O¬ϕ, axiom K. O(ϕ ⇒ ψ) ⇒ (O ϕ ⇒ O ψ), axiom D. O ϕ ⇒ P ϕ and the necessitation rule, N according to which if ϕ is a theorem, so is O ϕ. We may on the other hand distinguish obligations directly established by the norms in the normative system of the system and derived obligation extracted from such norms, e.g., indicating what is necessary for complying with them (van der Torre and Hansen 2008). This too would be consistent with our framework, but we cannot explore it here.

For representing legal contents, we need norms, which can be viewed a kind of defeasible conditional.

Definition 3 (Norm).

A norm has the form

$$\displaystyle{ A \Rightarrow B }$$

where A is the antecedent condition, B the ensuing normative conclusion, and ⇒ expresses a defeasible unidirectional connection, according to which antecedent A triggers conclusion B. In the norm the antecedent A is a proposition and consequent B is any kind of deontic or constitutive normative qualification.

Thus, a norm A ⇒ B captures the unidirectional defeasible connection between an antecedent (possibly empty) fact and the normative consequent that is generated by that fact: normative effect B is triggered when the antecedent condition A holds. We write A ⇏ B for the statement that the norm’s antecedent fails to support the norm’s conclusion, so that the norm cannot be applied in valid inferences. Arguments establishing that A ⇏ B undercut the use of the norm A ⇒ B in valid inferences.

Here is an example of two deontic norms, the first stating that it is forbidden to cause damage to others, and the second that who causes damage to another has the obligation to compensate the latter (in the following when obvious I drop the requirement xy):

$$\displaystyle{ \begin{array}{ll} &x\neq y \Rightarrow O\neg \mathsf{E}_{x}Damaged(y) \\ &x\neq y \wedge \mathsf{E}_{x}Damaged(y) \Rightarrow O\mathsf{E}_{x}Compensated(y)\end{array} }$$

The following is an example of a constitutive norm, saying that if we injure a person (make so that someone is injured), we cause damage to that person (injuring counts as damaging):

$$\displaystyle{ \mathsf{E}_{x}Injured(y) \Rightarrow \mathsf{E}_{x}Damaged(y) }$$

Note that I do not distinguish deontic conditionals and constitutive or counts-as conditionals, since both are modelled as defeasible conditionals (Searle 1995; Jones and Sergot 1996).

I assume an argumentation system as defined in Prakken (2010). Such a system includes two sets of inference rules, strict and defeasible inference rules, which have to be applied to a knowledge base of premises.

Strict inference rules have the form [ϕ 1, … ϕ n ] ↦ ψ. The conclusion ψ of the strict rule holds without exceptions when all its antecedent conditions ϕ 1, … ϕ n hold; therefore the application of the rule to derive ψ cannot be challenged unless at least one antecedent condition in ϕ 1, … ϕ n is also challenged.

Defeasible inference rules have the form \([\phi _{1},\ldots \phi _{n}] \rightsquigarrow \psi\); the conclusion ψ of the defeasible rule holds only presumptively (with the possibility of exceptions) when all its antecedent conditions ϕ 1, … ϕ n hold; therefore the application of the rule to derive ψ can be challenged also without challenging the antecedent conditions, i.e. by rebutting the rule’s conclusion or by undercutting the rule’s application.

Rules of both kinds can applied to a knowledge base of premises, these being formulas in a logical language.

Here I shall just introduce the main idea of an argumentation system in an informal way. The model I propose is inspired by Prakken (2010), to which I refer for a detailed presentation, though my account will depart from it in some aspects, to provide a simpler framework.

Premises, i.e., formulas in the underlying logical language \(\mathcal{L}\) are basic arguments. Further arguments can constructed by applying inference rules to the conclusions of arguments already available: thus given arguments A 1, … A n with conclusion ϕ 1, … ϕ n , through an inference rule [ϕ 1, … ϕ n ] ↦ ψ we can obtain argument B s  = { A 1, … A n ψ}, while through an inference rule \([\phi _{1},\ldots \phi _{n}] \rightsquigarrow \psi\) we can obtain argument \(B_{d} =\{ A_{1},\ldots A_{n} \rightsquigarrow \chi \}\). For instance, given premises a and a ⇒ b and inference rule \([\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi\), we can construct arguments A 1 = { a}, A 2 = { a ⇒ b}, and \(A_{3} =\{ A_{1},A_{2} \rightsquigarrow b\}\), i.e., \(\{\{a\},\{a \Rightarrow b\} \rightsquigarrow b\}\).

Arguments may be defeated (rebutted or undercut) by counterarguments: rebutting takes place when an argument having a conclusion ψ through a defeasible rule (as its ultimate conclusion, or the conclusion of one of its subarguments) faces a non weaker counterargument having the complementary conclusion \(\overline{\psi }\); undercutting takes place when an argument including a defeasible rule \([\phi _{1},\ldots \phi _{n}] \rightsquigarrow \psi\), having name r (we assume that each rule has an unique name) has a counterargument with conclusion ¬r (the negation of a rule-name being understood as the denial of the rule’s applicability). An argument is justified, with regard to a knowledge base, if all of its of its defeaters are overruled, being defeated by further justified arguments.Footnote 1

Here I shall not view neither facts nor norms are inference rules of the argumentation system, but rather as premises for it, i.e., as part of its knowledge base (see Prakken and Sartor 2013). Thus, I shall assume a general pattern for building strict inference rules out of modus ponens entailments, and similarly, a general pattern for building defeasible inference rules and undercutters out of defeasible rules (such as, norms). For my purpose, I do not need to address preferences between rules. Therefore the following characterisation of a normative argumentation system will suffice.

Definition 4 (Argumentation system).

An argumentation system S is a tuple N S  = (L, R s , R d ) where

  • \(\mathcal{L}\) is a logical language (here including, in particular, the constructs for propositional logic, action and deontic logic, and the conditional symbols ⇒ , and ⇏ ).

  • R s is the set of all strict inference rules, including

    • Strict modus ponens inference rules: all rules corresponding to the schema [ϕ, ϕ ⇒ ψ] ↦ ψ for any ϕ and ψ in \(\mathcal{L}\);

    • Specification: all rules corresponding to the schema [ϕ] ↦ ϕ[t], where t is a substitution of variables in ϕ with terms in \(\mathcal{L}\);

    • Logical axioms: all rules corresponding to the schema [] ↦ ϕ, where ϕ is any theorem of propositional logic or other deductive logical systems to be used (here action logic E, and standard deontic logic D}

  • R d is the set of all defeasible inference rules, including

    • Defeasible modus ponens inference rules: all rules corresponding to the schema \([\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi\) for any ϕ and ψ in \(\mathcal{L}\).

    • Defeasible undercutting inference rules: all rules corresponding to the schema \([\phi \not\Rightarrow \psi ] \rightsquigarrow \neg `[\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi '\) where \(`[\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi '\) is the name for the inference rule \([\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi\)

We read the name of an inference rule as the assertion that the rule is applicable, and so the negation the name of an inference rule is the assertion that the rule is inapplicable; the defeasible undercutting inference schema says that if a norm fails to support its conclusion, then the inference rule based on that norm is inapplicable.

The logical-axioms inference-schema allows any theorem of the deductive logics being used (e.g. O ϕ → P ϕ from deontic logic) to be introduced in any argument, as an unchallengeable premise. We can now define the idea of a normative knowledge base.

Definition 5 (Normative Knowledge Base).

A normative knowledge base K is a tuple (C, N), of two sets of premises:

  • a set C of contextual circumstances,

  • a set N of norms (a normative system)

By contextual circumstances I mean the propositions describing the relevant facts of the case, such as the fact that John damages Tom, the amount of the damage, whether John intended to cause the damage or was careless, etc. This notion of a fact is a relative one, since certain normative rules (the constitutive ones) may establish under what conditions a certain qualification is satisfied, so that an apparently factual qualification becomes a normative outcome. Consider for instance legal rules establishing what counts as negligence in road traffic, or in medical practice. For our purpose however, we do not need to address this issue, since we view as contextual circumstances all relevant true propositions that are not established through the application of the normative system we are considering (i.e., that are not established as being the consequent of a norm whose antecedent condition is satisfied).

Finally we define the notion of an entailment with regard to a normative knowledge base and a normative argumentation system.

Definition 6 (Defeasible entailment).

We shall say that a normative knowledge base K = (C, N) defeasibly entails A,and write K ∣​ ∼ A, to mean that knowledge base K enables us to construct a justified argument for A, using the inference rules in argumentation system S.

For instance, given knowledge base K 1 = ({a}, {a ⇒ b}), we can construct an undefeated (indeed unattached) argument A 3 for b. Thus we may say that K 1 ∣​ ∼ b.

Let us now consider knowledge base

$$\displaystyle{K_{2} = (\{a,c,d\},\{a \Rightarrow b,c \Rightarrow \neg b,d \Rightarrow (c\not\Rightarrow \neg b)\})}$$

This knowledge base enables the construction of argument A 3 for b, as above.

K 2 also enables the construction of arguments A 4 = { c}, A 5 = { c ⇒ ¬b} and \(A_{6} =\{ A_{4},A_{5} \rightsquigarrow \neg b\}\), the latter being a rebutting counterargument to A 3.

However K 2 also provides for the construction of arguments A 6 = { d}, A 7 = { d ⇒ (c ⇏ ¬b)}, \(A_{8} =\{ A_{6},A_{7} \rightsquigarrow c\not\Rightarrow \neg b\}\) and \(A_{9} =\{ A_{8} \rightsquigarrow \neg `([c,c \Rightarrow \neg b] \rightsquigarrow \neg b)'\).} The last arguments undercuts A 5, so that A 3 is freed from its only attacker and is thus justified.

The following example shows how from a norm and an instance of its antecedent we can defeasibly derive an instance of the conditional’s consequent.

$$\displaystyle{ \begin{array}{l} \{\mathsf{E}_{Tom}Damaged(John),\mathsf{E}_{x}Damaged(y) \Rightarrow O\mathsf{E}_{x}Compensated(y)\}\,\mid \! \sim \, \\ O\mathsf{E}_{Tom}Compensated(John)\end{array} }$$

To execute this inference we just need to instantiate the pattern \([\phi,\phi \Rightarrow \psi ] \rightsquigarrow \psi\) into the defeasible inference rule

$$\displaystyle{ \begin{array}{ll} [\mathsf{E}_{Tom}Damaged(John),\mathsf{E}_{x}Damaged(y) \\ \Rightarrow O\mathsf{E}_{x}Compensated(y)] \rightsquigarrow O\mathsf{E}_{x}Compensated(y) \end{array} }$$

and apply it to the fact and rule above.

3 Relativised Obligations and Permissions

In addressing compliance we have to connect a normative system N and the factual circumstances C relevant to N’s application, in the context of a given argumentation system. Here I am only interested in obligations and institutional facts that are generated by norms in N, when applied to facts in C. Thus we can assume that C contains (or entails) all factual literals (atomic propositions or negations of them) which are true in the real or hypothetical situation in which the norms have to be applied, without considering how the truth of such literals can be established. For simplicity’s sake we can limit C to the factual literals that are relevant to the application of norms in N, matching literals in the antecedent of a norm in N. When the considered factual circumstances are those that hold in the real world (rather than in a merely possible situation), i.e., they are the truths relevant to the application of N in the case at hand, I shall denote them through the expression T N .

I will now introduce the notion of a relativised obligation, namely, a way of expressing the fact that an obligation holds with regard to a normative system and a set of circumstances. A relativised obligation sentence does not express a norm, but it expresses an assertion about the implications of norms (normative systems) and circumstances (in the terminology of Alchourrón 1969 and Alchourrón and Bulygin 1971 such assertions are called “normative propositions”).

Definition 7 (Relativised sentences and obligations).

We say that any sentence ϕ holds relatively to normative system N and circumstances C, and write [ϕ] C, N iff (C, N) ∣​ ∼ ϕ

$$\displaystyle{ [\phi ]_{C,N}\stackrel{\mathrm{def}}{=}(C,N)\,\mid \! \sim \,\phi }$$

Let the expression \(\mathcal{A}_{j}\) cover both E j A and ¬E j A and let \(\overline{\mathcal{A}_{j}}\) denote the complement of \(\mathcal{A}_{j}\) (\(\overline{\mathcal{A}_{j}}\) stands for ¬E j ϕ if \(\mathcal{A}_{j} = \mathsf{E}_{j}\phi\); it stands for E j ϕ if \(\mathcal{A}_{j} = \neg \mathsf{E}_{j}\phi\)). Then we can say that it is obligatory to do (or not to do) an action relatively to a certain set of circumstances and a normative system, if such circumstances and system entail that the action ought (or ought not) to be done.

$$\displaystyle{ \mathbb{O}_{C,N}\mathcal{A}_{x}\stackrel{\mathrm{def}}{=}(C,N)\,\mid \! \sim \, O\mathcal{A}_{x} }$$

When we are referring to the true relevant circumstances of the real world, denoted as T N , rather than to circumstances of hypothetical situations, we simply write [ϕ] N , or \(\mathbb{O}_{N}\mathsf{E}_{x}A\).

$$\displaystyle{ \begin{array}{c} [\phi ]_{N}\stackrel{\mathrm{def}}{=}(T_{N},N)\,\mid \! \sim \,\phi \\ \mathbb{O}_{N}\mathcal{A}_{x}\stackrel{\mathrm{def}}{=}(T_{N},N)\,\mid \! \sim \, O\mathcal{A}_{x}\end{array} }$$

For instance, let us consider the following example, where N 1 includes a simplified version of the three norms above, and circumstances C 1 are limited to the fact that John injured Tom:

Example 1.

$$\displaystyle{ \begin{array}{ll} C_{1} = &\{\mathsf{E}_{John}Injured(Tom)\} \\ N_{1} =&\{\mathsf{E}_{x}Injured(y) \Rightarrow \mathsf{E}_{x}Damaged(y) \\ &O\neg \mathsf{E}_{x}Damaged(y) \\ &\mathsf{E}_{x}Damaged(y) \Rightarrow O\mathsf{E}_{x}Compensated(y)\} \end{array} }$$

It is easy to see that the following inferences holds on the basis of example (1):

$$\displaystyle{ \begin{array}{c} (C_{1},N_{1})\,\mid \! \sim \,\mathsf{E}_{John}Damaged(Tom) \\ (C_{1},N_{1})\,\mid \! \sim \, O\mathsf{E}_{John}Compensated(Tom)\end{array} }$$

Therefore, we can say that John has damaged Tom and that it is obligatory that John compensates Tom, relatively to N 1 and C 1, i.e., that

$$\displaystyle{ \begin{array}{c} [\mathsf{E}_{John}Damaged(Tom)]_{N_{1},C_{1}} \\ \mathbb{O}_{N_{1},C_{1}}\mathsf{E}_{John}Compensated(Tom) \end{array} }$$

If John has really injured Tom (and no other relevant circumstances obtain, such as exceptions excluding the application of the norms at issue) we can simply say that, according to N 1, John has damaged Tom and it is obligatory that John compensates Tom i.e.:

$$\displaystyle{ \begin{array}{c} [\mathsf{E}_{John}Damaged(Tom)]_{N_{1}} \wedge \mathbb{O}_{N_{1}}\mathsf{E}_{John}Compensated(Tom) \end{array} }$$

On the basis of example (1) we can also say that it is obligatory that John refrains from damaging Tom

$$\displaystyle{ \mathbb{O}_{N_{1}}\neg \mathsf{E}_{John}Damaged(Tom) }$$

Given that it holds that \([\mathsf{E}_{John}Damaged(Tom)]_{N_{1}}\) we can conclude that the latter obligation has been violated, on the basis of the following definition.

Definition 8 (Violation).

An obligation O E x A of a normative system N is violated in circumstances C iff \((C,N)\,\mid \! \sim \, O\mathsf{E}_{x}A \wedge \neg \mathsf{E}_{x}A\), In other words the obligation is violated in C, iff both \(\mathbb{O}_{C,N}\mathsf{E}_{x}A\) and [¬E x A] C, N hold.

Here is another small example. The first norm in N 2 says that if one is in a public place then one is forbidden to smoke. The second says that places open to the public are (count as) public places.

Example 2.

$$\displaystyle{ \begin{array}{ll} C_{2} = &\{OpenToPublic(LectureRoom),in(John,LectureRoom)\} \\ N_{2} =&\{OpenToPublic(y) \Rightarrow PublicPlace(y) \\ &PublicPlace(y) \wedge in(x,y) \Rightarrow O\neg \mathsf{E}_{x}Smoke\} \end{array} }$$

We can say then say that according to N 2 in circumstances C 2 it is obligatory that John does not smoke (\(\mathbb{O}_{C_{2},N_{2}}\neg \mathsf{E}_{Tom}Smoke\)).

Clearly, the language of relativised obligations allows us to say that according to different normative systems different obligations hold. For instance, given that Canon law contains both a universal norm prohibiting the use of contraception and a constitutive rule saying any action meant to make a sex act unfruitful counts as artificial contraception, we can conclude that according to the Canon law a woman, say Ann, is forbidden to take the pill in order to have unfruitful sex acts. Similarly, given that Islamic law contains a norm that prohibits receiving interest on loans of money, we can say that according to Islamic law John is forbidden to receive interest on loans of money.

A notion of relativised permission can be provided that corresponds to the above analysis of obligation. While permissions can be modelled as the negation of prohibitions (\(P\mathsf{E}_{x}A\stackrel{\mathrm{def}}{=}\neg O\neg \mathsf{E}_{x}A\)), relativised permissions can be defined as follows.

Definition 9 (Relativised permission).

Let us say that it is permissible relatively to N and C that x does (or not does) an action, iff N and C entail that it is permissible to do (or not to do) that action:

$$\displaystyle{ \mathbb{P}_{C,N}\mathcal{A}_{x}\stackrel{\mathrm{def}}{=}(C,N)\,\mid \! \sim \, P\mathcal{A}_{x} }$$

Note that according to this definition, saying that an action E x ϕ is permissible relatively to normative system N and circumstances C (\(\mathbb{P}_{C,N}\mathsf{E}_{x}\phi\)) does not amount to saying that it is not the case that E x ϕ is forbidden relatively to the same system and circumstances (\(\neg \mathbb{O}_{C,N}\neg \mathsf{E}_{x}\phi\)). Proposition \(\mathbb{P}_{C,N}\mathsf{E}_{x}\phi\) is not equivalent to \(\neg \mathbb{O}_{C,N}\neg \mathsf{E}_{x}\phi\), since the former holds when (C, N) entails P E x ϕ, while the latter holds when (C, N) does not entail O¬E x ϕ (see Alchourrón 1969; Alchourrón and Bulygin 1971).

4 Reasoning with Normative Systems

Let us assume that Tom wants to know his position concerning the normative systems L (the law). In particular Tom is now wondering whether he should pay income tax on the capital gains he obtained by selling his house. Being committed to comply with the law, but not knowing what the law requires from him, Tom asks the tax expert Ann for advise. Assume that the Ann remembers that there is a rule in the tax code that establishes the requirement to pay income taxes on capital gains, but vaguely remembers that there are exceptions to it. This prompts Ann to look for exceptions, and she finds indeed one concerning houses. This exception says (in a simplified form) that capital gains from the sale of houses purchased more than 5 year before the sale and inhabited by the seller are exempted from income tax. Assume that Ann’s inquiry has led her to conclude that the legal system L she is considering, for instance Italian law, contains the following relevant norms:

$$\displaystyle{ \begin{array}{ll} L \supseteq &\{SellsHouse(x) \Rightarrow O\mathsf{E}_{x}PayIncomeTaxOnSale, \\ &BoughtMoreThan5Y earsBefore(x) \wedge HasInhabitedHouse(x) \\ & \Rightarrow (SellsHouse(x)\not\Rightarrow O\mathsf{E}_{x}PayIncomeTaxOnSale)\}\end{array} }$$
(1)

where the second norms in (1) says that under the indicated conditions the first one does not hold (is not applicable).

Ann then asks Tom whether, at the time of the sale, more that 5 years had elapsed from the Tom’s purchase, and whether he has been living in the house. Assume that Tom replies positively to the first question and negatively to the second one. Then Ann says: “Dear, Tom, unfortunately you are legally bound to pay income tax on your gains”. In fact, by combining the Italian law L with these factual circumstances (let us assume these circumstances are the only relevant ones), Ann can see that the following inference holds:

$$\displaystyle{ \begin{array}{l} (\{SellsHouse(Tom),\neg HasInhabitedHouse(Tom)\},L)\,\mid \! \sim \, \\ O\mathsf{E}_{Tom}PayIncomeTaxOnSale \end{array} }$$

so that, given that both factual premises are true, she can infer what she tells her client:

$$\displaystyle{ \mathbb{O}_{L}\mathsf{E}_{Tom}PayIncomeTaxOnSale }$$

If Tom asks for an explanation, Ann would probably answer by saying that whenever one has not lived in the house one sells, then according to the law one has the obligation to pay income tax:

$$\displaystyle{ SellsHouse(x)\wedge \neg HasInhabitedSoldHouse(x) \Rightarrow \mathbb{O}_{L}\mathsf{E}_{(}x)PayIncomeTaxOnSale }$$
(2)

Note that formula (2) does not express a norm of L (there is no norm in L which has exactly that content, see formula (1)). More generally (2) is no norm at all, but rather is a general conditional statement about L, namely the statement that in case that the seller has not inhabited the sold house, then L entails that the seller has to pay taxes on capital gains. Similarly, if Ann were contacted by Tom before making the sale, she would tell him: “Since you have not inhabited the house, if you seel it you will have to pay income tax on your capital gain”.

5 Dynamic Normative Systems

Let us now consider how an agent (a legislator) can have the ability to introduce new norms in N. For this purpose, we need to assume that N is a dynamic normative system (Kelsen 1967), including meta-norms which determine what new norms will be valid according to N.

In the framework we have described above, the idea of such a metanorm can be captured through an additional pattern for defeasible inference rules, which enables the production basic norm set to be expanded by further norms. Thus we obtain what we may call a dynamic extension of the normative argumentation system.

Definition 10 (Dynamic Normative Argumentation System).

A dynamic normative argumentation systems DS is obtained by adding to an argumentation systems S the following inference rules

  • Norm creation: all inference rules corresponding to the schema Valid(ϕ) ↦ ϕ, for any norm ϕ in \(\mathcal{L}\).

Note that I prefer to model this principle as a strict rule, but depending on how we understand the notion of validity, we could also model it as a defeasible rule (see Sartor 2008).

In a dynamic normative argumentation system, arguments may use rules that do not belong to an initial knowledge base, but that are qualified as valid by norms in that knowledge base. So, let us assume that the knowledge base K of a normative system includes a meta-norm saying that whatever norm ϕ is issued by the legislator Leg than ϕ is valid (for simplicity’s sake I do not consider the temporal dimension of validity, see Governatori and Rotolo 2010).

$$\displaystyle{ \mathsf{E}_{Leg}Issued(\phi ) \Rightarrow V alid(\phi ) }$$

Given this background, let us assume that legislator accomplished the action of issuing a new norm, for instance, a norm prohibiting any agent x to smoke:

$$\displaystyle{ \mathsf{E}_{Leg}Issued(O\neg \mathsf{E}_{x}Smoke) }$$

The accomplishment of the action described in this formula is a new fact, which is added to the true factual circumstances T N . Thus we can build the following sequence of arguments

  • A 1 = E Leg Issued(O¬E x Smoke), premise;

  • A 2 = E Leg Issued(ϕ) ⇒ Valid(ϕ), premise;

  • A 3 = A 2 ↦ (E Leg Issued(O¬E x Smoke) ⇒ Valid(O¬E x Smoke)), by specification;

  • \(A_{4} = A_{1},A_{3} \rightsquigarrow V alid(O\neg \mathsf{E}_{x}Smoke)\), by defeasible modus ponens;

  • A 5 = A4 ↦ O¬E x Smoke, by norm creation;

  • A 6 = A5 ↦ O¬E Tom Smoke, by specification.

Thus, on the basis of argument A 6 (which we assume to be unchallenged) we can conclude that smoking is forbidden to Tom according to N:

$$\displaystyle{ \mathbb{O}_{N}\neg \mathsf{E}_{Tom}Smoke }$$

6 Conclusion

In this paper I have shown how a reasoner may approach a normative system, namely a distinct set of norms, viewing it as an object that enables the derivation of normative conclusions that are relative to that system. For this purpose I have first considered how to model actions, obligations and norms. Then I have defined an argumentation system which takes as inputs knowledge bases of facts and norms, and produces appropriate arguments. On this basis I have considered how obligations and permission can be relative to a particular normative systems, and I have provided a meta-logical representation of this idea. Finally I have developed some considerations on how to model dynamic normative systems in this framework.

While this work is still very preliminary, I hope it can provide some clues on how to model metalevel reasoning with normative systems. Obvious extensions, to be considered in future work, concern integrating this idea with the decision-making process of the concerned agents (for a preliminary attempt, see Sartor 2012), and modelling reasoning with multiple distinct normative systems.