Keywords

1 Introduction

Nowadays, the number of online non-English documents on the Internet is continuously increased, which requires the availability of high-performance cross-language information retrieval (CLIR) systems satisfying the Internet users’ needs. Query translation (QT) techniques are the main research task in the domain of CLIR [12]. Besides, the usefulness of the simple dictionary-based translation approaches in QT has been improved due to the availability of machine readable bilingual dictionaries for several languages. However, these approaches are lacked by many major challenges such as: (i) the translation disambiguation problem known as the difficulty to choose the correct translation corresponding to each source query term among all the possible translations existing in the dictionary; and (ii) the poor coverage of the available dictionaries suffering from the missing of many translations corresponding to new terminologies. Nevertheless, many research works in the literature [20, 22, 25] are dedicated to manually or automatically collect larger lexical resources in order to increase the coverage of these dictionaries. Moreover, translation ambiguity can decrease IR effectiveness. In order to overcome this limit, some approaches have used a phrase dictionary to firstly select noun phrases in the source query and secondly translate them as units.

We propose, assess and compare in this paper a new discriminative possibilistic QT approach dedicated to solve the problem of QT ambiguity using both a bilingual dictionary and a parallel text corpus in order to overcome some drawbacks of the dictionary-based QT techniques. In this approach, the relevance of a source query term translation is modeled by two measures: The possible relevance allows rejecting irrelevant translations, whereas the necessary relevance makes it possible to reinforce the translations not eliminated by the possibility. We compare this new approach to both the probabilistic and the probability-to-possibility transformation-based approaches [11], in which we start by identifying noun phrases (NPs) using the Stanford Parser Footnote 1 and translating them as units using translation patterns and a language model. Then, remaining source query terms are translated using a probabilistic word-by-word translation technique. Indeed, the suitable translation of each source query term or NP has a tendency to co-occur in the target language documents unlike unsuitable ones. Besides, additional words and their translations are automatically generated from a parallel bilingual corpus in order to increase the coverage of the bilingual dictionary. We assessed our approach using the French-English parallel text corpus Europarl Footnote 2 and the CLEF-2003 French-English CLIR test collection. Our results confirmed the performance of our new discriminative possibilistic approach compared to both the probabilistic and the probability-to-possibility transformation-based approaches [11], principally for short queries and using various evaluation scenarios and metrics.

This paper is organized as follows. We present in Sect. 2 related works in the field of QT and discuss the solutions aiming at solving the problem of translation ambiguity. In Sect. 3, we recall the necessary of possibility theory. The new discriminative possibilistic QT approach is given in Sect. 4. Section 5 details our experimentations, expose and discuss a comparative study between QT approaches. Section 6 concludes our work in this paper and suggests some perspectives for future research.

2 Related Work

Since the early 1990’s, many researchers (e.g. [19]) showed that the usefulness of a manually translating phrases have performed better results in CLIR effectiveness than a word-by-word dictionary-based translation. Moreover, CLIR performance was increased by [6] when they used their phrase dictionary generated from a set of parallel sentences in French and English. In the same way, Ballesteros and Croft [2] confirmed that translations of multi-word concepts as phrases are more perfect than word-by-word translations. Indeed, phrases translations were achieved using information existing in phrase and word usage available in the Collins machine readable dictionary. Unfortunately, it is hard to find or build an exhaustive phrase dictionary in CLIR since we have until now many missing phrases in these available lexicons. Therefore, it is not easy to select all missing phrases in the queries and suitably translate them. Actually, many unfamiliar phrases suffer from the problem of their identification and translation since they cannot be identified in any dictionaries. Hence, the problem of the lexicon coverage is one of the limits of this approach since we cannot know until now: How can one build an exhaustive phrase dictionary?

On the other hand, many authors have tackled the problem of translation ambiguity using word sense disambiguation (WSD) techniques. For example, Hull [18] used structured queries for disambiguation in QT task. Besides, co-occurrence statistics from corpora are used by Ballesteros and Croft [13] in order to decrease translation ambiguity in CLIR. Then, many disambiguation strategies are suggested and evaluated by Hiemstra and Jong [61] in CLIR tasks. Later, a technique-based statistical term similarity for word sense disambiguation was suggested and tested by [1] in order to enhance CLIR effectiveness. Even if the difficulty of translation ambiguity was significantly decreased, Xu and Weischedel [24] showed that it is not evident to accomplish perfect enhancement in CLIR performance. Recently, Lefever and Hoste [21] have argued the advantage of moving from traditional monolingual WSD task into a cross-lingual one. In fact, this new technique, namely “Cross-Lingual WSD” (CLWSD) has been involved in CLIR since SemEval-2010 and SemEval-2013 competitions. The participant in these exercises confirmed that CLIR effectiveness has been enhanced due to the new CLWSD technique which significantly resolves the problem of translation ambiguity.

QT techniques require training and matching models in order to compute a score of relevance (or similarities) between source query terms/phrases and their possible translations. Existing QT approaches in the literature are based on poor, uncertain and imprecise data, while possibility theory is naturally dedicated to this type of applications, since it takes into account of the imprecision and uncertainty at the same time and it makes it possible to express ignorance. However, the main challenge of our approach is that the context used in the translation disambiguation process of a given source query term can be also ambiguous. Consequently, we consider this phenomenon as a case of imprecision. That’s why we are inspired from the possibility theory which naturally applies to this kind of imperfection. By cons, the probability theory is not suitable to deal with such type of data. Besides, and given that the possibility theory is the best framework suitable for imprecision treatment, we have taken advantage of possibility distributions in order to solve the problem of translation ambiguity in CLIR task. Recently, we have proposed and tested in [11] a possibilistic QT approach derived from the probabilistic one using a probability-to-possibility transformation as a mean to introduce further tolerance in QT process. This approach has achieved a statistically significant improvement compared to the probabilistic one using both long and short queries.

3 Possibility Theory

We briefly present in the following sections the basic elements of possibility theory such as the possibility distribution (cf. Sect. 3.1), the possibility and necessity measures (cf. Sect. 3.2) and the possibilistic networks (cf. Sect. 3.3). More details and discussions about possibility theory are available in [7,8,9,10].

3.1 Possibility Distribution

The fundamental element of the possibility theory is the possibility distribution. Given the universe of discourse Ω = {ω 1 , ω 2 ,…, ω n }, we symbolised by π the basic concept corresponding to a function which associates to each element ω i ∈ Ω a value from a bounded and linearly ordered valuation set (L, <). Moreover, the possibility degree is defined as the value in which our knowledge on the real world is encoded. Indeed, this scale has two interpretations: (i) when the handled values reflect only an ordering between the different states of the world, it is the qualitative setting which can be applied using the min operator; and (ii) when the handled values have a real sense, it is the quantitative setting which can be applied using the product operator. Flexibility is modelled by allowing providing a possibility degree from the interval [0, 1]. We note that π(ω i ) = 1 means that it is fully possible that ω i is the real world, and π(ω i ) = 0 means that it is impossible that ω i is the real world. The possibility theory provides different significant exacting cases of knowledge as the following: (i) Complete knowledge (∃ ω i ∈ Ω, π(ω i ) = 1 andω j  ≠ ω i , π(ω j ) = 0); (ii) Partial ignorance (\( \forall \;\omega_{i} \in A \subseteq \Upomega \), π(ω i ) = 1, ∀ω i A, π(ω i ) = 0; when A is not a singleton); and (iii) Total ignorance: all values in Ω are possible (∀ ω i ∈ Ω, π(ω i ) = 1).

3.2 Possibility and Necessity Measures

The two dual measures in which a possibility distribution π on Ω enables events to be qualified in terms of their plausibility and their certainty are respectively known as the Possibility (Π) and the Necessity (N) [7]. Given a possibility distribution π on the universe of discourse Ω, the corresponding possibility and necessity measures of any event \( A \subseteq 2^{\Upomega } \) are respectively determined by the Eqs. (1) and (2):

$$ \prod (A) = { \hbox{max} }_{w \in A} \pi (w) $$
(1)
$$ N(A) = { \hbox{min} }_{w \notin A} (1 - \pi (w)) = 1 - \prod (\overline{A} ) $$
(2)

In fact, Π(A) provides an assessment similar to a degree of non-emptiness of the intersection of the fuzzy set having π as membership function with the classical subset A. Thus, Π(A) measures at which level A is consistent with our knowledge represented by π. While, N(A) assesses at which level A is certainly inferred by our knowledge represented by π; since it is a degree of inclusion of the fuzzy set corresponding to π into the subset A.

3.3 Possibilistic Networks (PN)

The numerical and graphical components are the main characteristic of a directed possibilistic network on a variable set V. These two components are defined as follow: (i) the distinct links in the graph are quantified via the numerical component. Indeed, it represents conditional possibility matrix of every node given the context of its parents. (ii) The graphical component is a directed acyclic graph (DAG). The DAG enables representing conditional dependency between dependent or independent variables. Every link denotes a dependency between two variables and every node in the graph denotes a domain variable. The graph structure encodes independence relation sets between nodes. Moreover, these possibility distributions should respect the normalization feature [4]. For each variable V:

  • If V is a root node and Dom(V) the domain of V, the prior possibility of V should satisfy:

    $$ max_{v\in Dom(V)} \varPi( v) = 1 $$
    (3)
  • If V is not a root node, the conditional distribution of V in the context of its parents denoted U V should satisfy:

    $$ max_{v\in Dom(V)} \varPi \left( {v|u_{V} } \right) = 1;\;u_{V}\in Dom\left( {U_{V} } \right) $$
    (4)

Where: Dom(V): domain of V; U V : value of parents of V; Dom(U V ): domain of parent set of V.

We suggest in this paper a new possibilistic approach for QT disambiguation based on possibilistic network (cf. Sect. 4). We link in this network the possible translations (T i ) to the terms of a source query SQ = (t 1 , t 2 ,…,t p ), which represents its context. In this case: v i  = t i ; u V  = T i ; Dom(V) = {t 1 , t 2 ,…,t p }; and Dom(U V ) = {T 1 , T 2,…, T N }.

We provide in Sect. 4.2 an illustrative example including detailed calculus. The possibilistic graph which associated conditional possibility distribution is based on the product operator. The product-based possibilistic graph (PPG), is generally comfortable for the numerical setting where possibility measures represent numerical values in [0, 1]. The possibility distribution of product-based possibilistic networks (π p ) obtained by the associated chain class is calculated via Eq. (5):

$$ \pi_{p} (V_{1} ,V_{2} , \cdots ,V_{N} ) = \prod\limits_{i = 1}^{N} {\varPi (\left. {V_{i} } \right|U_{{V_{i} }} )} $$
(5)

4 The Discriminative Possibilistic QT Approach

We present in Sect. 4.1 the formulae for calculating the Degree of Possibilistic Relevance (DPR) and an illustrative example with a detailed calculus in Sect. 4.2.

4.1 The Degree of Possibilistic Relevance (DPR)

Let us consider the source query (SQ) enclosing P terms and denoted as: SQ = (t 1 , t 2 ,…, t P ). We assume that SQ includes only one ambiguous term having several possible translations. We note the Degree of Possibilistic Relevance of a translation T j given SQ by DPR(T j |SQ). We have taken advantage of a possibilistic matching model of information retrieval (IR) used in [5, 12,13,14,15,16] in order to assess the relevance of a translation T j given a source query SQ. In the case of IR, the matching score is calculated between a query and a document. However, in case of QT disambiguation, we model the relevance of a translation given a source query using double measures: The possible relevance and the necessary relevance. The irrelevant translations are rejected due to the possible relevance, while the relevance of the remaining translations, which have not been rejected by the possibility, is reinforced due to the necessary relevance.

Figure 1 presents our possibilistic network which links the word translation T j to the terms of a given source query SQ = (t 1 , t 2 ,…, t P ). The output of the QT disambiguation process is the target query TQ = (T 1 , T 2 ,…, T P ). The later will be useful to retrieve a set of relevant documents on the target language.

Fig. 1.
figure 1

Possibilistic network of the QT disambiguation approach

Giving the source query SQ, the relevance of every word translation T j is computed as the following:

Analogically to the IR matching model proposed in [46–49], the possibility Π(T j |SQ) is proportional to:

$$ \varPi^{\prime}\left({T_{j}|SQ} \right) = \varPi\left({t_{1}|T_{j}}\right)* \ldots *{\varPi}\left({t_{P}|T_{j}}\right) \; = \; nft_{1j}\,* \ldots *nft_{Pj} $$
(6)

Where:

  • nft ij  = tf ij /max(tf kj ): the normalized frequency of the source term t i in the parallel text of the translation T j .

  • tf ij is the number of occurrence of the source term t i in the parallel text of the translation T j divided by the number of terms in the parallel text of the translation T j .

We compute the necessity to restore a relevant translation T j given the source query SQ, denoted N(T j |SQ), as the following:

$$ {\text{N}}\left( {T_{j} |SQ} \right) = 1- \Uppi \left( {\neg T_{j} |SQ} \right) $$
(7)

Where:

$$ \Uppi \left( {\neg T_{j} |SQ} \right) = (\Uppi \left( {SQ| \, \neg T_{j} } \right)*\Uppi \left( {\neg T_{j} } \right))/\Uppi \left( {SQ} \right) $$
(8)

At the same way Π(¬T j | SQ) is proportional to:

$$ \Uppi^{\prime}\left( {\neg T_{j} |SQ} \right) = \Uppi \left( {t_{1} | \, \neg T_{j} } \right)* \, \ldots *\Uppi \left( {t_{P} | \, \neg T_{j} } \right) $$
(9)

This numerator can be expressed as the following:

$$ \Uppi^{\prime}\left( {\neg T_{j} |SQ} \right) = ( 1- \phi T_{1j} )* \ldots * \, ( 1- \phi T_{Pj} ) $$
(10)

Where:

$$ \phi T_{ij} = Log_{10} \left( {nCT/nT_{j} } \right)\;*\;\left( {nft_{ij} } \right) $$
(11)

Where: nCT is the number of possible translations in the bilingual dictionary. But, nT j is the number of parallel texts of the translation T j containing the source term t i . This includes all possible translations existing in the bilingual dictionary.

We compute the Degree of Possibilistic Relevance (DPR) of each word translation T j giving a source query SQ via the following Eq. (12):

$$ DPR\left( {T_{j} |SQ} \right) = \Uppi \left( {T_{j} |SQ} \right) + {\text{N}}\left( {T_{j} |SQ} \right) $$
(12)

Finally, the suitable translations are those which have a high score of DPR(T j |SQ).

4.2 Illustrative Example

We provide here a numerical calculation example for reasons of brevity. But, we have already detailed in [11] some data/corpus-based examples in which we have showed the difference between the probabilistic and the possibilistic QT approaches.

Let us consider the source query SQ = (W, t 2, t 4, t 5, t 7), which contains only one polysemous source query term W in order to simplify the calculus in this example. We assume that W has two possible translations T 1 and T 2 in the bilingual dictionary. We suppose also that the parallel text of T 1 is indexed by the three terms {t 1, t 2, t 3, t 4} and the parallel text of T 2 is indexed by {t 1, t 4, t 5, t 6, t 7}. We have:

Π(T 1|SQ) = nf(W, T1)* nf(t2, T1)* nf(t4, T1)* nf(t5, T1) * nf(t7, T1) = 0*(1/4)*(1/4)*0*0 = 0. Where: nf(W, T1) is the normalized frequency of W in the parallel text of the first translation T 1.

Π(T 2| SQ) = nf(W, T2) * nf(t2, T2) * nf(t4, T2) * nf(t5, T2) * nf(t7, T2) = 0*0*(1/5)*(1/5)*(1/5) = 0. We have frequently Π(T j | SQ) = 0; except if all the words of the source query exist in the index of the parallel text of the translation.

On the other hand, we have not null values of N(T j | SQ):

N(T 1| SQ) = 1− ((1 − ϕ(T 1, W)) * (1 − ϕ(T 1, t 2)) * (1 − ϕ(T 1, t 4)) * (1 − ϕ(T 1, t 5) * (1 − ϕ(T 1, t 7))); nf(T1, W) = 0, so ϕ(T 1, W) = 0; ϕ(T 1, t 2) = log 10(2/1)*(1/4) = 0,075; ϕ(T 1, t 4) = log 10(2/2)*(1/4) = 0; ϕ(T 1, t 5) = 0; ϕ(T 1, t 7) = 0. So: N(T 1| SQ) = 1 − ((1−0) * (1−0.075) * (1−0) * (1−0) * (1−0)) = 1− (1* 0.925* 1* 1*1) = 0.075. Thus, DPR(T 1| SQ) = 0.075. N(T 2|SQ) = 1− ((1 − ϕ(T 2, W))* (1 − ϕ(T 2, t 2))* (1 − ϕ(T 2, t 4))* (1 − ϕ(T 2, t 5) * (1 − ϕ(T 2, t 7))). Where: ϕ(T 2, W) = 0 because nf(T2, W) = 0; ϕ(T 2, t 2) = 0; ϕ(T 2, t 4) = log 10(2/2)*(1/5) = 0; ϕ(T 2, t 5) = log 10(2/1)*(1/5) = 0.06; ϕ(T 2, t 7) = log 10(2/1)*(1/5) = 0.06. So: N(T 2|SQ) = 1 − ((1−0)* (1−0)* (1−0)*(1−0.06)* (1−0.06)) = 1 − (1* 1*1*0.94* 0.94) = 1 − 0.8836 = 0.1164. Thus, DPR(T 2|SQ) = 0.1164 > DPR(T 1|SQ) = 0.075.

We notice that the source query SQ is more relevant for T 2 than T 1; because it encloses three terms (t 4, t 5, t 7) of the index of the parallel text of T 2 and only two terms (t 2, t 4) of the index of the parallel text of T 1 .

5 Experiments and Discussion

We assess, compare and discuss in this section the discriminative possibilistic approach for QT disambiguation. Indeed, we suggest various evaluation scenarios and metrics using the CLEF-2003 standard CLIR test collection. We compare the performance and the efficiency of the discriminative approach to both the most known efficient probabilistic and the probability-to-possibility transformation-based ones [11]. Moreover, our evaluation is performed according to the TREC protocol. We used the IR matching model OKAPI-BM25 existing in Terrier Footnote 3 platform in order to retrieve English relevant document. We are focused on performance metrics such as Recall and Precision, mostly used in the evaluation of CLIR tools. The evaluation assumes that there is an ideal set the system is supposed to search. This ideal set is useful to define these two metrics as follows. The recall is the percentage of documents in the ideal set that were retrieved by the system, while the precision is the percentage of documents retrieved by the system, which are also in the ideal set. Moreover, we used the precision (Y-axis) over 11 points of recall in the X-axis (0.0, 0.1,…, 1.0) to draw all recall-precision curves.

Besides, we evaluated these approaches using both the Mean Average Precision (MAP) and the exact precision (R-Precision). The latter is defined as the precision at rank R; where R is the total number of relevant documents, while the MAP is the mean of the average precision scores for each query. The Equations of calculating the MAP and the R-Precision are given in [11]. We compute also the improvement percentage between two variations of the model variables. This percentage is obtained in generally for two variables A and B measuring the percentage of C as %C = [(B − A)/A]*100. Our 54 test queries enclose 717 French words having 2324 possible English translations in the bilingual dictionary. Indeed, we firstly generate our bilingual dictionary from the Europarl parallel corpus using all French words with their possible translations existing in this corpus in order to enlarge our lexicon coverage. Then, we have benefited from the online intelligent speller and grammar checker Reverso Footnote 4 in order to check this dictionary. Finally, the free online Google translateFootnote 5 has been used to enrich and check this bilingual dictionary.

Firstly, we provide in Fig. 2 the Recall-Precision curves comparing monolingual (English queries provided by CLEF-2003), discriminative, probabilistic and probability-to-possibility transformation-based (possibilistic) runs using the title part of the source queries as input data to our CLIR tool described in [11]. Secondly, we provide the precision values at different top documents P@5, P@10,…, P@1000, the MAP and the R-Precision. For example, the precision in point 5, namely P@5, is the ratio of relevant documents among the top 5 returned documents. The goal of the following experiments and discussion is to show and assess our contributions compared to these competitors QT disambiguation approaches mainly investigated and tested in [11].

Fig. 2.
figure 2

Recall-Precision curves comparing monolingual, discriminative, probabilistic and probability-to-possibility transformation-based (possibilistic) runs

If we focus only on the title part of the query, the context is limited to a few numbers of terms in which the identification of the NP is not frequent in this case. Therefore, the discriminative possibilistic approach significantly outperforms both the probabilistic and the probability-to-possibility transformation-based (possibilistic) approaches; except in some low-levels points of recall (0 and 0.1). Unfortunately, the monolingual run is still outperformed these approaches in all points of recall.

On the other hand, and in order to further confirm our conclusions made above, we provide a comparative study between the monolingual, the discriminative, the possibilistic and the probabilistic runs using the precision at different top documents, the MAP and the R-Precision metrics (cf. Fig. 3). It is trivial that the precision decreases when the number of returned documents increases. Moreover, short queries using only title are more efficient for the discriminative approach for all top returned documents; except for some rare cases such as P@100 and P@1000 where the number of returned documents is important, which increases the noise in the retrieved results. Unfortunately, the monolingual run is normally upper bound of CLIR performance in all our experiments, because we don’t involve in our tests any query expansion step before and/or after the translation process and no close phrases/words have enriched the source and/or the target queries before returning search results.

Fig. 3.
figure 3

Results using the precision at different top documents, MAP and R-Precision

We remark that the discriminative approach outperformed both of them in terms of the MAP and the R-Precision. In fact, these two metrics confirm again that short query using title is still suitable for the discriminative approach compared to its two competitors QT techniques. We present in Table 1 the improvement percentage of the discriminative approach compared to both the probabilistic and the probability-to-possibility transformation-based approaches using the precision at different top documents, the MAP and the R-Precision. Firstly, and compared to the probabilistic, the discriminative approach has performed a significant improvement in terms of precision of documents returned to the top of list using the title of the source query. For example, we have registered an improvement percentage more than 15% for P@15 and almost 14% for P@20, while the average improvement for all top documents is about 6%. Besides, if we focus on the MAP metric, the improvement of the MAP is about 14.4% and about 6% for the R-Precision metric. Secondly, and compared to the probability-to-possibility transformation-based approach, the discriminative achieved an improvement percentage more than 7.9% for P@10 and more than 11.5% for P@15, whereas the average improvement for all top documents is about 2.4%. Moreover, the MAP is about 8.5% and the R-Precision is about 5.3%. These results confirm our deductions concluded above about the efficiency of the discriminative approach in case of short queries. It is the case when the user provided to the CLIR tool a few number of terms in his/her source query due to his/her lack of language or his/her limit knowledge about the retrieved domain.

Table 1. The improvement percentage of the discriminative compared to both the probabilistic and the probability-to-possibility transformation-based approaches

On the other hand, we need to more investigate on the statistical significance of the improvement achieved by the discriminative possibilistic approach compared to its competitors in terms of precision at different top documents, the MAP and the R-Precision scores using short queries. To do this, we use the Wilcoxon Matched-Pairs Signed-Ranks Test as suggested by [17]. This statistical test is a non-parametric alternative to the paired t-test that enables us to decide whether the improvement by method 1 over method 2 is significant. Indeed, the t-test computes a p-value based on the performance data of both methods 1 and 2. The improvement is more significant for the smaller p-value. Generally, the improvement is statistically significant if the p-value is small enough (p-value < 0.05).

The improvement of the discriminative possibilistic approach compared to the probabilistic one is statistically significant (p-value = 0.010793 < 0.05), while it is not statistically significant (p-value = 0.085831 > 0.05) compared to the probability-to-possibility transformation-based approach. Globally, these tests confirms again the performance of our discriminative possibilistic approach in the disambiguation of short queries using title compared to both the known efficient probabilistic and to the probability-to-possibility transformation-based approaches using different assessment metrics. Finally, and in order to provide an objective evaluation of our approach, we are limited in our empirical comparative study to these two approaches (probabilistic and possibilistic); because the other state-of-the-art QT techniques detailed in Sect. 2 are assessed using both different linguistic resources (dictionary and parallel corpora) and different CLIR test collection for different pair of languages.

6 Conclusion and Future Work

The translations’ ambiguities in a QT process are considered as cases of imprecision since many possible translations are available for each ambiguous source query term. The disambiguation process consists of using the context of the source query which can be also ambiguous. Consequently, we propose in this paper a new discriminative possibilistic approach for QT disambiguation dedicated to improve the dictionary-based QT ones. This new technique has taken advantage of both a parallel text corpus and a bilingual dictionary in order to overcome some weaknesses of the-state-of-the-art QT approaches. Indeed, we have modelled the relevance of possible translations within two measures: the possible relevance allows eliminating irrelevant translations, whereas the necessary relevance makes it possible to reinforce the translations not rejected by the possibility. We assessed and compared the discriminative possibilistic approach using the French-English parallel text corpus Europarl and the CLEF-2003 French-English CLIR test collection. Our experiments highlighted the performance of our new discriminative possibilistic approach compared to both the known efficient probabilistic and the probability-to-possibility transformation-based approaches [11] using different assessment metrics. Indeed, the improvement of the discriminative approach compared to the probabilistic one is statistically significant in terms of precision at different top documents, the MAP and the R-Precision.

In spite of its significant efficiency in translating short queries, the discriminative possibilistic approach suffers from some weaknesses in case of specific domain queries. Consequently, the identification of the suitable translations requires a domain-specific translation [23] and a language model. Nevertheless, combining a language model with a domain-specific translation and their integration in the discriminative approach are not easy tasks. Moreover, the evaluation processes of our approach should be done in real contexts by allowing the users to contribute in its assessment. It is also relevant to assess the impact of QT disambiguation on CLIR efficiency before and/or after query expansion process using our recent techniques in [3, 13, 15].