1 Introduction

We are interested in the problem of algorithmically learning a description for a formal language (a computably enumerable subset of the set of natural numbers) when presented successively all and only the elements of that language; this is called inductive inference, a branch of (algorithmic) learning theory. For example, a learner h might be presented more and more even numbers. After each new number, h outputs a description for a language as its conjecture. The learner h might decide to output a program for the set of all multiples of 4, as long as all numbers presented are divisible by 4. Later, when h sees an even number not divisible by 4, it might change this guess to a program for the set of all multiples of 2.

Many criteria for determining whether a learner h is successful on a language L have been proposed in the literature. Gold, in his seminal paper [10], gave a first, simple learning criterion, \(\mathbf {Txt}\mathbf {G}\mathbf {Ex}\)-learningFootnote 1, where a learner is successful if and only if, on every text for L (listing of all and only the elements of L) it eventually stops changing its conjectures, and its final conjecture is a correct description for the input language.

Trivially, each single, describable language L has a suitable constant function as a \(\mathbf {Txt}\mathbf {G}\mathbf {Ex}\)-learner (this learner constantly outputs a description for L). Thus, we are interested in analyzing for which classes of languages \(\mathcal {L}\) is there a single learner h learning each member of \(\mathcal {L}\). This framework is also known as language learning in the limit and has been studied extensively, using a wide range of learning criteria similar to \(\mathbf {Txt}\mathbf {G}\mathbf {Ex}\)-learning (see, for example, the textbook [11]).

In this paper, we put the focus on the possible descriptions for languages. Any computably enumerable language L has as possible descriptions any program enumerating all and only the elements of L, called a W-index (the language enumerated by program e is denoted by \(W_e\)). This system has various drawbacks; most importantly, the function which decides, given e and x, whether \(x \in W_e\) is not computable. We propose to use different descriptors for languages: programs for characteristic functions (where such programs e describe the language \(C_e\) which it decides). Of course, only decidable languages have such a description, but now, given a program e for a characteristic function, \(x \in C_e\) is decidable. Additionally to many questions that remain undecidable (for example, whether C-indices are for the same language or whether a C-index is for a finite language), it is not decidable whether a program e is indeed a program for a characteristic function. This leads to a new set of problems: Learners cannot be (algorithmically) checked whether their outputs are viable (in the sense of being programs for characteristic functions).

Based on this last observation, we study a range of different criteria which formalize what kind of behavior we expect from our learners. In the most relaxed setting, learners may output any number (for a program) they want, but in order to \(\mathbf {Ex}\)-learn, they need to converge to a correct C-index; we denote this restriction with \(\mathbf {Ex}_C\). Requiring additionally to only use C-indices in order to successfully learn, we denote by \(\mathbf {CInd}\mathbf {Ex}_C\); requiring C-indices on all inputs (not just for successful learning, but also when seeing input from no target language whatsoever) we denote by \(\tau (\mathbf {CInd})\mathbf {Ex}_C\). In particular, the last restriction requires the learner to be total; in order to distinguish whether the loss of learning power is due to the totality restriction or truly due to the additional requirement of outputting C-indices, we also study \(\mathcal {R}\mathbf {CInd}\mathbf {Ex}_C\), that is, the requirement \(\mathbf {CInd}\mathbf {Ex}_C\) where additionally the learner is required to be total.

We note that \(\tau (\mathbf {CInd})\mathbf {Ex}_C\) is similar to learning indexable families. Indexable families are classes of languages \(\mathcal {L}\) such that there is an enumeration \((L_i)_{i \in \mathbb {N}}\) of all and only the elements of \(\mathcal {L}\) for which the decision problem “\(x \in L_i\)” is decidable. Already for such classes of languages, we get a rich structure (see a survey of previous work [16]). For a learner h learning according to \(\tau (\mathbf {CInd})\mathbf {Ex}_C\), we have that \(L_x = C_{h(x)}\) gives an indexing of a family of languages, and h learns some subset thereof. We are specifically interested in the area between this setting and learning with W-indices (\(\mathbf {Ex}_W\)).

The criteria we analyze naturally interpolate between these two settings. We show that we have the following hierarchy: \(\tau (\mathbf {CInd})\mathbf {Ex}_C\) allows for learning strictly fewer classes of languages than \(\mathcal {R}\mathbf {CInd}\mathbf {Ex}_C\), which allow for learning the same classes as \(\mathbf {CInd}\mathbf {Ex}_C\), which again are fewer than learnable by \(\mathbf {Ex}_C\), which in turn renders fewer classes learnable than \(\mathbf {Ex}_W\).

All these results hold for learning with full information. In order to study the dependence on the mode of information presentation, we also consider partially set-driven learners (\(\mathbf {Psd}\), [2, 19]), which only get the set of data presented so far and the iteration number as input; set-driven learners (\(\mathbf {Sd}\), [20]), which get only the set of data presented so far; iterative learners (\(\mathbf {It}\), [8, 21]), which only get the new datum and their current hypothesis and, finally, transductive learners (\(\mathbf {Td}\), [4, 15]), which only get the current data. Note that transductive learners are mostly of interest as a proper restriction to all other modes of information presentation. In particular, we show that full-information learners can be turned into partially set-driven learners without loss of learning power and iterative learning is strictly less powerful than set-driven learning, in all settings.

Altogether we analyze 25 different criteria and show how each pair relates. All these results are summarized in Fig. 1(a) as one big map stating all pairwise relations of the learning criteria mentioned, giving 300 pairwise relations in one diagram, proven with 13 theorems in Sect. 3. Note that the results comparing learning criteria with W-indices were previously known, and some proofs could be extended to also cover learning with C-indices. For the proofs, please consider the full version of this paper [1].

Fig. 1.
figure 1

Relation of (a) various requirements when to output characteristic indices and (b) various learning criteria, both paired with various memory restrictions \(\beta \). Black solid respectively dashed lines imply trivial respectively non-trivial inclusions (bottom-to-top, left-to-right). Furthermore, greyly edged areas illustrate a collapse of the enclosed learning criteria and there are no further collapses.

In Sect. 4, we derive a similar map considering a possible relaxation on \(\mathbf {Ex}_C\)-learning: While \(\mathbf {Ex}_C\) requires syntactic convergence to one single correct C-index, we consider behaviorally correct learning (\(\mathbf {Bc}_C\), [6, 17]) where the learner only has to semantically converge to correct C-indices (but may use infinitely many different such indices). We again consider the different modes of data presentation and determine all pairwise relations in Fig. 1(b). The proofs are again deferred to the full version [1].

2 Preliminaries

2.1 Mathematical Notations and Learning Criteria

In this section, we discuss the used notation as well as the system for learning criteria [15] we follow. Unintroduced notation follows the textbook [18].

With \(\mathbb {N}\) we denote the set of all natural numbers, namely \(\{0,1,2,\ldots \} \). We denote the subset and proper subset relation between two sets with \(\subseteq \) and \(\subsetneq \), respectively. We use \(\emptyset \) and \(\varepsilon \) to denote the empty set and empty sequence, respectively. The set of all computable functions is denoted by \(\mathcal {P}\), the subset of all total computable functions by \(\mathcal {R}\). If a function \(f\) is (not) defined on some argument \(x\in \mathbb {N}\), we say that \(f\) converges (diverges) on \(x\), denoting this fact with \(f(x){\downarrow }\) (\(f(x){\uparrow }\)). We fix an effective numbering \(\{{\varphi _{e}}\}_{e\in \mathbb {N}} \) of \(\mathcal {P}\). For any \(e\in \mathbb {N}\), we let \(W_e\) denote the domain of \(\varphi _{e}\) and call \(e\) a \(W\)-index of \(W_e\). This set we call the e-th computably enumerable set. We call \(e\in \mathbb {N}\) a \(C\)-index (characteristic index) if and only if \(\varphi _{e}\) is a total function such that for all \(x\in \mathbb {N}\) we have \(\varphi _{e}(x)\in \{0,1\} \). Furthermore, we let \(C_e=\{x\in \mathbb {N}\}{\varphi _{e}(x)=1 }\). For a computably enumerable set L, if some \(e\in \mathbb {N}\) is a \(C\)-Index with \(C_e=L\), we write \(\varphi _{e}=\chi _{L}\). Note that, if a set has a \(C\)-index, it is recursive. The set of all recursive sets is denoted by \(\mathbf{REC} \). For a finite set \(D\subseteq \mathbb {N}\), we let \(\mathrm {ind}(D)\) be a \(C\)-index for \(D\). Note that \(\mathrm {ind}\in \mathcal {R}\). Furthermore, we fix a Blum complexity measure \(\varPhi \) associated with \(\varphi \), that is, for all \(e,x\in \mathbb {N}\), \(\varPhi _e(x)\) is the number of steps the function \(\varphi _e\) takes on input \(x\) to converge [3]. The padding function \(\mathrm {pad}\in \mathcal {R}\) is an injective function such that, for all \(e,n\in \mathbb {N}\), we have \( \varphi _{e}=\varphi _{_{\mathrm {pad}(e,n)}}\). We use \(\langle \cdot ,\cdot \rangle \) as a computable, bijective function that codes a pair of natural numbers into a single one. We use \(\pi _1\) and \(\pi _2\) as computable decoding functions for the first and section component, i.e., for all \(x,y\in \mathbb {N}\) we have \( \pi _1(\langle x,y \rangle )=x\) and \(\pi _2(\langle x,y\rangle )=y\).

We learn computably enumerable sets L, called languages. We fix a pause symbol \(\#\), and let, for any set \(S\), . Information about languages is given from text, that is, total functions \(T:\mathbb {N}\rightarrow \mathbb {N}\cup \{\#\} \). A text T is of a certain language L if its content is exactly L, that is, is exactly L. We denote the set of all texts as \(\mathbf {Txt}\) and the set of all texts of a language L as \(\mathbf {Txt}(L)\). For any \(n \in \mathbb {N}\), we denote with T[n] the initial sequence of the text T of length n, that is, and . Given a language \(L\) and \(t\in \mathbb {N}\), the set of sequences consisting of elements of \(L \cup \{ \# \}\) that are at most \(t\) long is denoted by \(L^{\le t}_\#\). Furthermore, we denote with \({\mathbb {S}\mathrm {eq}}\) all finite sequences over \(\mathbb {N}_\# \) and define the content of such sequences analogous to the content of texts. The concatenation of two sequences \(\sigma , \tau \in {\mathbb {S}\mathrm {eq}}\) is denoted by \(\sigma \tau \) or, more emphasizing, \(\sigma ^\frown \tau \). Furthermore, we write \(\subseteq \) for the extension relation on sequences and fix a order \(\le \) on \({\mathbb {S}\mathrm {eq}}\) interpreted as natural numbers.

Now, we formalize learning criteria using the following system [15]. A learner is a partial function \(h\in \mathcal {P}\). An interaction operator \(\beta \) is an operator that takes a learner \(h\in \mathcal {P}\) and a text \(T\in \mathbf {Txt}\) as input and outputs a (possibly partial) function \(p\). Intuitively, \(\beta \) defines which information is available to the learner for making its hypothesis. We consider Gold-style or full-information learning [10], denoted by \(\mathbf {G}\), partially set-driven learning (\(\mathbf {Psd}\), [2, 19]), set-driven learning (\(\mathbf {Sd}\), [20]), iterative learning (\(\mathbf {It}\), [8, 21]) and transductive learning (\(\mathbf {Td}\), [4, 15]). To define the latter formally, we introduce a symbol “?” for the learner to signalize that the information given is insufficient. Formally, for all learners \(h\in \mathcal {P}\), texts \(T\in \mathbf {Txt}\) and all \(i\in \mathbb {N}\), define

$$\begin{aligned} \mathbf {G}(h,T)(i)&= h(T[i]);\\ \mathbf {Psd}(h,T)(i)&= h(\mathrm {content}(T[i]),i);\\ \mathbf {Sd}(h,T)(i)&= h(\mathrm {content}(T[i]));\\ \mathbf {It}(h,T)(i)&= {\left\{ \begin{array}{ll} h(\varepsilon ), &{}\text {if }i=0;\\ h(\mathbf {It}(h,T)(i-1),T(i-1)), &{}\text {otherwise}; \end{array}\right. }\\ \mathbf {Td}(h,T)(i)&={\left\{ \begin{array}{ll} {?}, &{}\text {if }i=0;\\ \mathbf {Td}(h,T)(i-1), &{}\text {else, if }h(T(i-1))={?};\\ h(T(i-1)), &{}\text {otherwise}. \end{array}\right. } \end{aligned}$$

For any of the named interaction operators \(\beta \), given a \(\beta \)-learner \(h\), we let \(h^*\) (the starred learner) denote a \(\mathbf {G}\)-learner simulating \(h\), i.e., for all \(T\in \mathbf {Txt}\), we have \(\beta (h,T)=\mathbf {G}(h^*,T)\). For example, let \(h\) be a \(\mathbf {Sd}\)-learner. Then, intuitively, \(h^*\) ignores all information but the content of the input, simulating \(h\) with this information, i.e., for all finite sequences \(\sigma \), we have \(h^*(\sigma )=h(\mathrm {content}(\sigma ))\).

For a learner to successfully identify a language, we may oppose constraints on the hypotheses the learner makes. These are called learning restrictions. As a first, famous example, we required the learner to be explanatory [10], i.e., the learner must converge to a single, correct hypothesis for the target language. We hereby distinguish whether the final hypothesis is interpreted as a \(C\)-index (\(\mathbf {Ex}_C\)) or as a \(W\)-index (\(\mathbf {Ex}_W\)). Formally, for any sequence of hypotheses \(p\) and text \(T\in \mathbf {Txt}\), we have

$$\begin{aligned} \mathbf {Ex}_C(p,T)&\Leftrightarrow \exists n_0:\forall n\ge n_0:p(n)=p(n_0)\wedge \varphi _{p(n_0)}=\chi _{\mathrm {content}(T)};\\ \mathbf {Ex}_W(p,T)&\Leftrightarrow \exists n_0:\forall n\ge n_0:p(n)=p(n_0)\wedge W_{p(n_0)}=\mathrm {content}(T). \end{aligned}$$

We say that explanatory learning requires syntactic convergence. If there exists a \(C\)-index (or \(W\)-index) for a language, then there exist infinitely many. This motivates to not require syntactic but only semantic convergence, i.e., the learner may make mind changes, but it has to, eventually, only output correct hypotheses. This is called behaviorally correct learning (\(\mathbf {Bc}_C\) or \(\mathbf {Bc}_W\), [6, 17]). Formally, let \(p\) be a sequence of hypotheses and let \(T\in \mathbf {Txt}\), then

$$\begin{aligned} \mathbf {Bc}_C(p, T)&\Leftrightarrow \exists n_0:\forall n\ge n_0:\varphi _{p(n)}=\chi _{\mathrm {content}(T)};\\ \mathbf {Bc}_W(p, T)&\Leftrightarrow \exists n_0:\forall n\ge n_0:W_{p(n)}=\mathrm {content}(T). \end{aligned}$$

In this paper, we consider learning with \(C\)-indices. It is, thus, natural to require the hypotheses to consist solely of \(C\)-indices, called \(C\)-index learning, and denoted by \(\mathbf {CInd}\). Formally, for a sequence of hypotheses \(p\) and a text \(T\), we have

$$\begin{aligned} \mathbf {CInd}(p,T)&\Leftrightarrow \forall i,x:\varphi _{p(i)}(x)\in \{0,1\}. \end{aligned}$$

For two learning restrictions \(\delta \) and \(\delta '\), their combination is their intersection, denoted by their juxtaposition \(\delta \delta '\). We let \(\mathbf {T}\) denote the learning restriction that is always true, which is interpreted as the absence of a learning restriction.

A learning criterion is a tuple \((\alpha ,\mathcal {C},\beta ,\delta )\), where \(\mathcal {C}\) is the set of admissible learners, usually \(\mathcal {P}\) or \(\mathcal {R}\), \(\beta \) is an interaction operator and \(\alpha \) and \(\delta \) are learning restrictions. We denote this criterion with \(\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta \), omitting \(\mathcal {C}\) if \(\mathcal {C}=\mathcal {P}\), and a learning restriction if it equals \(\mathbf {T}\). We say that an admissible learner \(h\in \mathcal {C}\) \(\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta \)-learns a language \(L\) if and only if, for arbitrary texts \(T\in \mathbf {Txt}\), we have \(\alpha (\beta (h,T),T)\) and for all texts \(T\in \mathbf {Txt}(L)\) we have \(\delta (\beta (h,T),T)\). The set of languages \(\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta \)-learned by \(h\in \mathcal {C}\) is denoted by \(\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta (h)\). With \([\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta ]\) we denote the set of all classes \(\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta \)-learnable by some learner in \(\mathcal {C}\). Moreover, to compare learning with W- and C-indices, these classes may only contain recursive languages, which we denote as \({[\tau (\alpha )\mathcal {C}\mathbf {Txt}\beta \delta ]}_\mathbf {REC}\).

2.2 Normal Forms

When studying language learning in the limit, there are certain properties of learner that are useful, e.g., if we can assume a learner to be total. Cases where learners may be assumed total have been studied in the literature [13, 14]. Importantly, this is the case for explanatory Gold-style learners obeying delayable learning restrictions and for behaviorally correct learners obeying delayable restrictions. Intuitively, a learning restriction is delayable if it allows hypotheses to be arbitrarily, but not indefinitely postponed without violating the restriction. Formally, a learning restriction \(\delta \) is delayable, if and only if for all non-decreasing, unbounded functions \(r:\mathbb {N}\rightarrow \mathbb {N}\), texts \(T, T'\in \mathbf {Txt}\) and learning sequences \(p\) such that for all \(n\in \mathbb {N}\), \(\mathrm {content}(T[r(n)])\subseteq \mathrm {content}(T'[n])\) and \(\mathrm {content}(T)=\mathrm {content}(T')\), we have, if \(\delta (p,T)\), then also \(\delta (p\circ r, T')\). Note that \(\mathbf {Ex}_W\), \(\mathbf {Ex}_C\), \(\mathbf {Bc}_W\), \(\mathbf {Bc}_C\) and \(\mathbf {CInd}\) are delayable restrictions.

Another useful notion are locking sequences. Intuitively, these contain enough information such that a learner, after seeing this information, converges correctly and does not change its mind anymore whatever additional information from the target language it is given. Formally, let \(L\) be a language and let \(\sigma \in L_\#^*\). Given a \(\mathbf {G}\)-learner \(h\in \mathcal {P}\), \(\sigma \) is a locking sequence for \(h\) on \(L\) if and only if for all sequences \(\tau \in L_\#^*\) we have \(h(\sigma )=h(\sigma \tau )\) and \(h(\sigma )\) is a correct hypothesis for \(L\) [2]. This concept can immediately be transferred to other interaction operators. Exemplary, given a \(\mathbf {Sd}\)-learner h and a locking sequence \(\sigma \) of the starred learner \(h^*\), we call the set \(\mathrm {content}(\sigma )\) a locking set. Analogously, one transfers this definition to the other interaction operators. It shall not remain unmentioned that, when considering \(\mathbf {Psd}\)-learners, we speak of locking information. In the case of \(\mathbf {Bc}_W\)-learning we do not require the learner to syntactically converge. Therefore, we call a sequence \(\sigma \in L_\#^*\) a \(\mathbf {Bc}_W\)-locking sequence for a \(\mathbf {G}\)-learner \(h\) on \(L\) if, for all sequences \(\tau \in L_\#^*\), \(h(\sigma \tau )\) is a correct hypothesis for \(L\) [11]. We omit the transfer to other interaction operators as it is immediate. It is an important observation that for any learner h and any language L it learns, there exists a (\(\mathbf {Bc}_W\)-) locking sequence [2]. These notions and results directly transfer to \(\mathbf {Ex}_C\)- and \(\mathbf {Bc}_C\)-learning. When it is clear from the context, we omit the index.

3 Requiring C-Indices as Output

This section is dedicated to proving Fig. 1(a), giving all pairwise relations for the different settings of requiring C-indices for output in the various mentioned modes of data presentation. In general, we observe that the later we require C-indices, the more learning power the learner has. This holds except for transductive learners which converge to C-indices. We show that they are as powerful as \(\mathbf {CInd}\)-transductive learners.

Although we learn classes of recursive languages, the requirement to converge to characteristic indices does heavily limit a learners capabilities. In the next theorem we show that even transductive learners which converge to W-indices can learn classes of languages which no Gold-style \(\mathbf {Ex}_C\)-learner can learn. We exploit the fact that C-indices, even if only conjectured eventually, must contain both positive and negative information about the guess.

Theorem 1

We have that \({[\mathbf {Txt}\mathbf {Td}\mathbf {Ex}_W]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {G}\mathbf {Ex}_C]}_\mathbf {REC}\ne \emptyset \).

Proof

We show this by using the Operator Recursion Theorem (\(\mathbf{ORT} \)) to provide a separating class of languages. To this end, let h be the \(\mathbf {Td}\)-learner with \(h(\#)={?}\) and, for all \(x,y\in \mathbb {N}\), let \(h(\langle x,y \rangle )= x\). Let \(\mathcal {L} = \mathbf {Txt}\mathbf {Td}\mathbf {Ex}_W(h)\cap \mathbf{REC} \). Assume \(\mathcal {L}\) can be learned by a \(\mathbf {Txt}\mathbf {G}\mathbf {Ex}_C\)-learner \(h'\). We may assume \(h'\in \mathcal {R}\) [13]. Then, by ORT there exist indices \(e,p,q\in \mathbb {N}\) such that

Here, \(\varPhi \) is a Blum complexity measure [3]. Intuitively, to define the next \(\varphi _p(x)\), we add the same element to \(\mathrm {content}(\tilde{T})\) until we know whether \(\langle e, \tilde{T}[x]+1 \rangle \in C_{h'(\bar{\sigma })}\) holds or not. Then, we add the element contradicting this outcome.

We first show that \(L\in \mathcal {L}\) and afterwards that \(L\) cannot be learned by \(h'\). To show the former, note that either \(L\) is finite or \(\tilde{T}\) is a non-decreasing unbounded computable enumeration of \(L\). Therefore, we have \(L\in \mathbf{REC} \). We now prove that h learns L. Let \(T \in \mathbf {Txt}(L)\). For all \(n \in \mathbb {N}\) where T(n) is not the pause symbol, we have \(h(T(n)) = e\). With \(n_0\in \mathbb {N}\) being minimal such that \(T(n_0)\ne \#\), we get for all \(n\ge n_0\) that \(\mathbf {Td}(h,T)(n)=e\). As \(e\) is a correct hypothesis, \(h\) learns \(L\) from \(T\) and thus we have that \(L\in \mathbf {Txt}\mathbf {Td}\mathbf {Ex}_W(h)\). Altogether, we get that \(L\in \mathcal {L}\).

By assumption, \(h'\) learns \(L\) from the text \(\tilde{T} \in \mathbf {Txt}(L)\). Therefore, there exists \(n_0\in \mathbb {N}\) such that, for all \(n\ge n_0\),

$$ h'(\tilde{T}[n]) = h'(\tilde{T}[n_0]) \text { and } \chi _{L}= \varphi _{h'(\tilde{T}[n])}, $$

that is, \(h'(\tilde{T}[n])\) is a \(C\)-index for \(L\). Now, as \(h'\) outputs C-indices when converging, there are \(t, t' \ge n_0\) such that

$$ \varPhi _{h'(\tilde{T}[t'])}(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle ) \le t. $$

Let \(t'_0\) and \(t_0\) be the first such found. We show that \(h'(\tilde{T}[t'_0])\) is no correct hypothesis of L by distinguishing the following cases.

  1. 1.

    Case: \(\varphi _{h'(\tilde{T}[t'_0])}(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle ) = 0\). By definition of \(\varphi _q\) and by minimality of \(t_0'\), we have that \(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle \in L\), however, the hypothesis of \(h'(\tilde{T}[t'_0])\) says differently, a contradiction.

  2. 2.

    Case: \(\varphi _{h'(\tilde{T}[t'_0])}(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle ) = 1\). By definition of \(\varphi _q\) and by minimality of \(t_0'\), we have that \(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle \in L\), but \(\langle e,\varphi _q(\tilde{T}[n_0])+1\rangle \notin L\). However, the hypothesis of \(h'(\tilde{T}[t'_0])\) conjectures the latter to be in L, a contradiction.

   \(\square \)

Furthermore, the following known equalities from learning W-indices directly apply in the studied setting as well.

Theorem 2

([9, 12, 19]). We have that

$$\begin{aligned} {[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_W]}_\mathbf {REC}&\subseteq {[\mathbf {Txt}\mathbf {Sd}\mathbf {Ex}_W]}_\mathbf {REC}, \\ {[\mathbf {Txt}\mathbf {Psd}\mathbf {Ex}_W]}_\mathbf {REC}&= {[\mathbf {Txt}\mathbf {G}\mathbf {Ex}_W]}_\mathbf {REC}. \end{aligned}$$

The remaining separations we will show in a more general way, see Theorems 11 and 12. We generalize the latter result [9, 19], namely that Gold-style learners may be assumed partially set-driven, to all considered cases. The idea here is to, just as in the \(\mathbf {Ex}_W\)-case, mimic the given learner and to search for minimal locking sequences. Incorporating the result that unrestricted Gold-style learners may be assumed total [13], we even get a stronger result.

Theorem 3

For \(\delta ,\delta ' \in \{\mathbf {CInd}, \mathbf {T}\}\), we have that

$${[\tau (\delta )\mathbf {Txt}\mathbf {G}\delta '\mathbf {Ex}_C]}_\mathbf {REC}= {[\tau (\delta )\mathcal {R}\mathbf {Txt}\mathbf {Psd}\delta '\mathbf {Ex}_C]}_\mathbf {REC}.$$

We also generalize the former result of Theorem 2 to hold in all considered cases. The same simulating argument (where one mimics the iterative learner on ascending text with a pause symbol between two elements) suffices regardless the exact setting.

Theorem 4

Let \(\delta ,\delta '\in \{\mathbf {CInd},\mathbf {T}\}\) and \(\mathcal {C}\in \{\mathcal {R},\mathcal {P}\}\). Then, we have that

$$\begin{aligned}{[\tau (\delta ')\mathcal {C}\mathbf {Txt}\mathbf {It}\delta \mathbf {Ex}_C]}_\mathbf {REC}\subseteq {[\tau (\delta ')\mathcal {C}\mathbf {Txt}\mathbf {Sd}\delta \mathbf {Ex}_C]}_\mathbf {REC}.\end{aligned}$$

Interestingly, totality is not restrictive solely for Gold-style (and due to the equality also partially set-driven) learners. For the other considered learners with restricted memory, being total lessens the learning capabilities. This weakness results from the need to output some guess. A partial learner can await this guess and outperform it. This way, we obtain self-learning languages [5] to show the three following separations.

Theorem 5

We have that \({[\mathcal {R}\mathbf {Txt}\mathbf {Sd}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\subsetneq {[\mathbf {Txt}\mathbf {Sd}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\).

Theorem 6

We have that \({[\mathcal {R}\mathbf {Txt}\mathbf {It}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\subsetneq {[\mathbf {Txt}\mathbf {It}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\).

Theorem 7

We have that \({[\mathcal {R}\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\subsetneq {[\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\).

Next, we show the gradual decrease of learning power the more we require the learners to output characteristic indices. We have already seen in Theorem 1 that converging to C-indices lessens learning power. However, this allows for more learning power than outputting these indices during the whole learning process as shows the next theorem. The idea is that such learners have to be certain about their guesses as these are indices of characteristic functions. When constructing a separating class using self-learning languages [5], one forces the \(\mathbf {CInd}\)-learner to output C-indices on certain languages to, then, contradict its choice there. This way, the \(\mathbf {Ex}_C\)-learner learns languages the \(\mathbf {CInd}\)-learner cannot.

Theorem 8

We have that \({[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_C]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {G}\mathbf {CInd}\mathbf {Bc}_C]}_\mathbf {REC}\ne \emptyset \).

Since languages which can be learned by iterative learners can also be learned by set-driven ones (see Theorem 4), this result suffices. Note that the idea above requires some knowledge on previous elements. Thus, it is no coincidence that this separation does not include transductive learners. Since these learners base their guesses on single elements, they cannot see how far in the learning process they are. Thus, they are forced to always output C-indices.

Theorem 9

We have that \({[\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}= {[\mathbf {Txt}\mathbf {Td}\mathbf {Ex}_C]}_\mathbf {REC}\).

For the remainder of this section, we focus on learners which output characteristic indices on arbitrary input, that is, we focus on \(\tau (\mathbf {CInd})\)-learners. First, we show that the requirement of always outputting C-indices lessens a learners learning power, even when compared to total \(\mathbf {CInd}\)-learners. To provide the separating class of self-learning languages, one again awaits the \(\tau (\mathbf {CInd})\)-learner’s decision and then, based on these, learns languages this learner cannot.

Theorem 10

We have \({[\mathcal {R}\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C]}_\mathbf {REC}\setminus {[\tau (\mathbf {CInd})\mathbf {Txt}\mathbf {G}\mathbf {Bc}_C]}_\mathbf {REC}\ne \emptyset \).

Proof

We prove the result by providing a separating class of languages. Let h be the \(\mathbf {Td}\)-learner with \(h(\#)={?}\) and, for all \(x,y\in \mathbb {N}\), let \(h(\langle x,y \rangle )= x\). By construction, \(h\) is total and computable. Let \(\mathcal {L} = \mathcal {R}\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C(h) \cap \mathbf{REC} \). We show that there is no \(\tau (\mathbf {CInd})\mathbf {Txt}\mathbf {G}\mathbf {Bc}_C\)-learner learning \(\mathcal {L}\) by way of contradiction. Assume there is a \(\tau (\mathbf {CInd})\mathbf {Txt}\mathbf {G}\mathbf {Bc}_C\)-learner \(h'\) which learns \(\mathcal {L}\). With the Operator Recursion Theorem (\(\mathbf{ORT} \)), there are \(e,p \in \mathbb {N}\) such that for all \(x\in \mathbb {N}\)

Intuitively, for all \(x\) either \(\varphi _{p}(x)\) is an element of \(L\) if it is not in the hypothesis of \(h'\) after seeing \(\varphi _{p}[x]\), or there is an element in this hypothesis that is not in \(\mathrm {content}(\tilde{T})\). As any hypothesis of \(h'\) is a C-index, we have that \(\varphi _p \in \mathcal {R}\) and, as \( \varphi _{p}\) is strictly monotonically increasing, that L is decidable.

We now prove that \(L \in \mathcal {L}\) and afterwards that \(L\) cannot be learned by \(h'\). First, we need to prove that h learns L. Let \(T \in \mathbf {Txt}(L)\). For all \(n \in \mathbb {N}\) where T(n) is not the pause symbol, we have \(h(T(n)) = e\). Let \(n_0\in \mathbb {N}\) with \(T(n_0)\ne \#\). Then, we have, for all \(n\ge n_0\), that \(\mathbf {Td}(h,T)(n)=e\) and, since \(e\) is a hypothesis of L, \(h\) learns \(L\) from \(T\). Thus, we have that \(L\in \mathcal {R}\mathbf {Txt}\mathbf {Td}\mathbf {CInd}\mathbf {Ex}_C(h) \cap \mathbf{REC} \).

By assumption, \(h'\) learns \(\mathcal {L}\) and thus it also needs to learn L on text \(\tilde{T}\). Hence, there is \(x_0\) such that for all \(x \ge x_0\) the hypothesis \(h'(\tilde{T}[x])=h'(\varphi _{p}[x])\) is a \(C\)-index for \(L\). We now consider the following cases.

  1. 1.

    Case: \(\varphi _{h'(\varphi _p[x])}(\langle e, 2x \rangle ) = 0\). By construction, we have that \(\tilde{T}(x) = \langle e, 2x \rangle \). Therefore, \(\langle e, 2x \rangle \in L\), which contradicts \(h'(\varphi _p[x])\) being a correct hypothesis.

  2. 2.

    Case: \(\varphi _{h'(\varphi _p[x])}(\langle e, 2x \rangle ) = 1\). By construction, we have that \(\tilde{T}(x) \ne \langle e, 2x \rangle \) and thus, because \(\tilde{T}\) is strictly monotonically increasing, \(\langle e, 2x \rangle \notin L=\mathrm {content}(\tilde{T})\). This, again, contradicts \(h'(\varphi _p[x])\) being a correct hypothesis.

As in all cases \(h'(\varphi _{p}[x])\) is a wrong hypothesis, \(h'\) cannot learn \(\mathcal {L}\).   \(\square \)

It remains to be shown that memory restrictions are severe for such learners as well. First, we show that partially set-driven learners are more powerful than set-driven ones. Just as originally witnessed by for W-indices [9, 19], this is solely due to the lack of learning time. In the following theorem, we already separate from behaviorally correct learners, as we will need this stronger version later on.

Theorem 11

We have that \({[\tau (\mathbf {CInd})\mathbf {Txt}\mathbf {Psd}\mathbf {Ex}_C]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {Sd}\mathbf {Bc}_W]}_\mathbf {REC}\ne \emptyset \).

In turn, this lack of time is not as severe as lack of memory. The standard class (of recursive languages) to separate set-driven learners from iterative ones [11] can be transferred to the setting studied in this paper.

Theorem 12

We have that \({[\tau (\mathbf {CInd})\mathbf {Txt}\mathbf {Sd}\mathbf {Ex}_C]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_W]}_\mathbf {REC}\ne \emptyset \).

Lastly, we show that transductive learners, having basically no memory, do severely lack learning power. As they have to infer their conjectures from single elements they, in fact, cannot even learn basic classes such as \(\{ \{0 \}, \{1 \}, \{0,1\}\}\). The following result concludes the map shown in Fig. 1(a) and, therefore, also this section.

Theorem 13

For \(\beta \in \{\mathbf {It},\mathbf {Sd}\}\), we have that

$${[\tau (\mathbf {CInd})\mathbf {Txt}\beta \mathbf {Ex}_C]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {Td}\mathbf {Ex}_W]}_\mathbf {REC}\ne \emptyset .$$

4 Syntactic Versus Semantic Convergence to C-indices

In this section, we investigate the effects on learners when we require them to converge to characteristic indices. We study both syntactically converging learners as well as semantically converging ones. In particular, we compare learners imposed with different well-studied memory restrictions.

Surprisingly, we observe that, although C-indices incorporate and, thus, require the learner to obtain more information during the learning process than W-indices, the relative relations of the considered restrictions remain the same. We start by gathering results which directly follow from the previous section.

Corollary 1

We have that

$$\begin{aligned} {[\mathbf {Txt}\mathbf {Psd}\mathbf {Ex}_C]}_\mathbf {REC}&= {[\mathbf {Txt}\mathbf {G}\mathbf {Ex}_C]}_\mathbf {REC}, (\text {Theorem 3)}, \\ {[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_C]}_\mathbf {REC}&\subseteq {[\mathbf {Txt}\mathbf {Sd}\mathbf {Ex}_C]}_\mathbf {REC}, (\text {Theorem 4}), \\ {[\mathbf {Txt}\mathbf {G}\mathbf {Ex}_C]}_\mathbf {REC}&\setminus {[\mathbf {Txt}\mathbf {Sd}\mathbf {Bc}_C]}_\mathbf {REC}\ne \emptyset , (\text {Theorem 11}), \\ {[\mathbf {Txt}\mathbf {Sd}\mathbf {Ex}_C]}_\mathbf {REC}&\setminus {[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_C]}_\mathbf {REC}\ne \emptyset , (\text {Theorem 12}), \\ {[\mathbf {Txt}\mathbf {It}\mathbf {Ex}_C]}_\mathbf {REC}&\setminus {[\mathbf {Txt}\mathbf {Td}\mathbf {Ex}_C]}_\mathbf {REC}\ne \emptyset , (\text {Theorem 13}). \end{aligned}$$

We show the remaining results. First, we show that, just as for W-indices, behaviorally correct learners are more powerful than explanatory ones. We provide a separating class exploiting that explanatory learners must converge to a single, correct hypothesis. We collect elements on which mind changes are witnessed, while maintaining decidability of the obtained language.

Theorem 14

We have that \({[\mathbf {Txt}\mathbf {Sd}\mathbf {Bc}_C]}_\mathbf {REC}\setminus {[\mathbf {Txt}\mathbf {G}\mathbf {Ex}_C]}_\mathbf {REC}\ne \emptyset \).

Next, we show that, just as for W-indices, a padding argument makes iterative behaviorally correct learners as powerful as Gold-style ones.

Theorem 15

We have that \({[\mathbf {Txt}\mathbf {It}\mathbf {Bc}_C]}_\mathbf {REC}= {[\mathbf {Txt}\mathbf {G}\mathbf {Bc}_C]}_\mathbf {REC}\).

We show that the classes of languages learnable by some behaviorally correct Gold-style (or, equivalently, iterative) learner, can also be learned by partially set-driven ones. We follow the proof which is given in a private communication with Sanjay Jain [7]. The idea there is to search for minimal \(\mathbf {Bc}\)-locking sequences without directly mimicking the \(\mathbf {G}\)-learner. We transfer this idea to hold when converging to C-indices as well. We remark that, while doing the necessary enumerations, one needs to make sure these are characteristic. One obtains this as the original learner eventually outputs characteristic indices.

Theorem 16

We have that \({[\mathbf {Txt}\mathbf {Psd}\mathbf {Bc}_C]}_\mathbf {REC}= {[\mathbf {Txt}\mathbf {G}\mathbf {Bc}_C]}_\mathbf {REC}\).

Lastly, we investigate transductive learners. Such learners base their hypotheses on a single element. Thus, one would expect them to benefit from dropping the requirement to converge to a single hypothesis. Interestingly, this does not hold true. This surprising fact originates from C-indices encoding characteristic functions. Thus, one can simply search for the minimal element on which no “?” is conjectured. The next result finalizes the map shown in Fig. 1(a) and, thus, this section.

Theorem 17

We have that \({[\mathbf {Txt}\mathbf {Td}\mathbf {Ex}_C]}_\mathbf {REC}= {[\mathbf {Txt}\mathbf {Td}\mathbf {Bc}_C]}_\mathbf {REC}\).