1 Introduction

Searching for information has become a part of our daily lives and people often use native language [1]. However, with the rapid development of the Internet and the progress of globalization, the information resources provided by the Internet are no longer concentrated in a few languages such as English, and people can access information in various languages [2]. According to W3Techs’ statistical analysis, English accounted for 54.8% of the content languages used in the global Internet Web site as of April 2013, with less than 7% in other languages, of which Chinese accounted for 4.4% [3]. As for May 31, 2011, the number of Internet users worldwide released by Internet World Stats was 2.01 billion, of which 26.8% were English users (29.4% in 2008) and 24.2% were Chinese users in the second place (2008 18.9%) [4]. The growth of Internet users using Chinese is very significant, and other Internet users who use non-English also have different degrees of growth. Therefore, there is a growing demand for using native language to query information expressed in different languages [5].

2 State of the art

A key issue in information retrieval is the multi-semantic representation of an object, which is manifested as polysemy, parasynonyms and synonyms in linguistic terms [6]. Even if the words or phrases contained in the query appear in the document, they may represent another meaning depending on the context. This rich semantic representation of natural language increases the difficulty of retrieving and querying related documents in the IR system [7]. In cross-language information retrieval or multi-language information retrieval, queries and documents are expressed in different languages. In addition to the semantic combination of words or phrases in a single language, cross-language semantic combinations exist, which increases the difficulties of retrievaling related documents. To overcome the language barrier we have to answer four core questions: What should be translated, queries or documents, or translate queries and documents into some kind of internal intermediate representation? Which type of symbol form should be translated in stem, word or phrase? How to use translation? That is, one word in the source language L1 may correspond to multiple words in the target language L1. How many words shoule we choose, one, some or all of the words? For a translation of trust, we should give it a greater weight value. How to remove inappropriate translation? Improper translation can severely affect the performance of cross-language searches, we need different censoring methods to reduce their negative impact.

3 Methodology

3.1 Information retrieval model

The main difference between the various retrieval systems is what kind of information retrieval model the system uses. Information retrieval model is the core part of the retrieval system. The information retrieval model gives a representation of the documents and queries and defines their scoring functions. We define the information retrieval model formally using a quadruplet \([D,Q,F,R(q_{i} ,d_{j}) ]\), where: D is the logical view (or representation) of the document set; Q is the logical view (or representation) of the user’s information requirements, which are called queries; F is defined as model document representation, query and the framework of their relationship, \(R(q_{i} ,d_{j} )\) is defined as the ordering function of the relevance between query \(q_{i} \in Q\) and document \(d_{j} \in D\),which gives the document a sequence of queries about qi. Researchers have proposed many information retrieval models. This section describes only commonly used search models: boolean, vector space, probability, and language models. In the information retrieval model, terms are the basic units used to represent. Terms can be words (e.g., computer), stem (e.g., comput), or phrases (e.g., computer system). The form in which the term is taken depends on the indexing method used to represent the basic semantic unit of the information content. Terms using different methods can construct more complex representations. Boolean model is the first proposed classic information retrieval model. In a Boolean model, the document is represented as a collection of lexical items or as a Boolean combination, and the query is represented as a Boolean combination of lexical items. E.g:

$$D = t_{1} \wedge t_{2} \wedge t_{3}$$
(1)

Indicates that document D contains the terms t1, t2, and t3. This formula is also equivalently expressed by a set expression. The terms that are not in Boolean expressions are assumed not to exist in the document. Similarly, the query can also use Boolean expressions. E.g, \(Q = (t_{1} \wedge t_{2} ) \vee t_{3}\). If and only if \(D \to Q\), you can determine whether a document related to the query. The main drawback of the Boolean model is that it does not give different weights to terms. So, some extended boolean models incorporate term weights, which consist of terms of different weights. Therefore, it is also possible to use a weighted Boolean model \(D \to Q\). For example, fuzzy set expansion logic Boolean model, p-norm and so on.

Vector space models use vectors to represent documents or queries. The vector space consists of system-recognizable terms in the document. In the document vector and the query vector, each element value (\(d_{i}\) or \(q_{i} ,1 \le i \le n\)) indicates the weight of the corresponding term in the document or query. The concrete expression is as follows, the vector space is \(< t_{1} ,t_{2} , \ldots ,t_{n} >\).The document vector is \(< d_{1} ,d_{2} , \ldots ,d_{n} >\), the query vector is \(< q_{1} ,q_{2} , \ldots ,q_{n} >\), the weight of \(d_{i}\) or \(q_{i}\) can be a binary value. For example, if a term appears in a document or query, it is 1, otherwise 0. However, the current use of Tf-idf weight calculation mode. Tf (Term Frequency) is the term frequency in the document or query, Idf (Inverse Document Frequency) is the inverse document frequency. Idf weight is calculated as follows:

$$Idf = \log \frac{N}{{n(t_{i} )}}$$
(2)

where is the term in the vocabulary, N is the number of documents in the document set, and \(n(t_{i} )\) is the number of documents (also called document frequency) that include the term \(t_{i}\). The basic idea of the Tf-idf weighting model is that the more frequently a term appears in a document or query, the more important it is (Tf factor); the more a term appears in a document with more documents, the more important it is Low (Idf factor). Although the formal definition of Boolean model and vector space model can not solve the problem of uncertainty in retrieval. In the probability model, the relevance of document D to query Q is achieved by estimating the probability of P (rel| D, Q), where rel indicates the correlation. The probabilistic model is based on the principle of probability ordering: if a reference retrieval system responds to each requirement, it is sorted by descending order of relevance of the document and the requirement, where the correlation probability is based on all the data the system can get May be estimated accurately, then the system is based on known data and can obtain the overall effect of the optimal system. The simplest probability model is the Binary Independence Retrieval (BIR) model. The BIR model assumes that the terms are independent of each other. The document is sorted according to the optimal ratio of P (rel | D, Q) to P (irrel | D, Q):

$$score(Q,D) = \log \frac{P(rel|D,Q)}{P(irrel|D,Q)} = \log \frac{P(D|Q,rel)P(rel|Q)}{P(D|Q,irrel)P(irrel|D)} \propto \log \frac{P(D|Q,rel)}{P(D|Q,irrel)}$$
(3)

Among them, irrel in P (irrel| D, Q) is irrelevant. Documents D expressed as the collection \(\{ x_{1} ,x_{2} , \ldots ,x_{n} \} ,x_{i} = 1\) of independent binary events indicates that the term \(x_{i}\) appears in the document, \(x_{i} = 0\) indicating that the term \(x_{i}\) does not appear in the document. Therefore, the derived BIR model is transformed into the estimated conditional probabilities P (xi = 1| Q, rel) and P (xi = 1| Q, irrel). Ideally, we can get a sample document set that gives a correlation. Given such a sample set of documents, we can calculate the contingency table for each term. Suppose N is the total number of documents in the sample set, R is the number of related documents, \(r_{i}\) is the number of related documents contained \(t_{i}\), and \(n_{i}\) is the number of documents included \(t_{i}\).

3.2 Thematic dual space model algorithm

As described above, the overall cross-lingual information retrieval framework based on bilingual subject spaces and the cross-language retrieval process need to first establish thematic duality space. The process of establishing this space was achieved by training the TSD model on a bilingual corpus. In the implementation of PLS algorithm, the classical non-linear iterative partial least squares algorithm (NIPALS) is generally adopted. The Do while loop body in the concrete algorithm converges to \(u_{i}\) of the first correct singular vector for calculation \(X_{i - 1}^{T} t_{i}\), and calculates \(v_{i}\) at the same time. After TSD model training algorithm is used to establish thematic dual space, the query and document represented by each language can be mapped to this space to achieve the task of cross-language and single language retrieval.

In terms of both space and time complexity, we analyze the algorithm efficiency of the TDS model proposed in this chapter. As mentioned above, suppose m is the number of documents in a bilingual parallel document set, n is the number of terms in language L1, r is L2.

Number of terms in the language. Without loss of generality hypothesis n > r, k is the number of pairs of bilingual subjects pre-given by the model. In the training phase of TDS model, we first need to store the bilingual document matrix X and Y, the space complexity is m × n and m × r respectively. Second, the output of the algorithm needs to store a set of vectors that represent dual space \(< u_{i} ,v_{i} >\), namely the matrices U and V, whose spatial complexity is n × k and r × k, respectively. Together with the central vector of the stored two-document matrix, the spatial complexity of the TDS model during the training phase is:

$$mn + mr + (r + n)k + n + r$$
(4)

In the training phase of TDS model, its time complexity is klmn and reaches O (n4). However, when we actually calculated the experiment,because the vector group \(< u_{i} ,v_{i} >\) needs to be calculated by loop l times. we found that the convergence rate was faster and the number of cycles was less than 15 times, at about 10 times. In the retrieval stage algorithm of TDS model, the algorithm stores the matrices U and V, document vector d and query vector q, the central vectors AvgX and AvgY, Pi and Qi of the original document matrix as a numerical variable, and finally outputs the projection vector. So, when retrieving a document, its spatial complexity is 2n + k for language L1 and 2r + k for language L2. Because the projection functions ProjectToU (x, U, k) and ProjectToV (y, V, k) have only one cycle, the complexity of the algorithm is n × k or r × k (Table 1).

Table 1 Comparison of document pairing search performance between TDS and CL-LSI at a small number of topics

4 Result analysis and discussion

4.1 Document paired search performance comparison analysis

We run TDS and CL-LSI two models in WSJFT bilingual parallel corpus for document matching search experiment. The paired search results when bilingual subject numbers are 5, 10, 20, and 50 are listed in Table 2. “C → E” in Table 2 means that the Chinese document retrieves the English paired document for the query, and “E → C” means the English document retrieves the Chinese paired document for the query. Figures 2 and 3 show the matching results of the two models in the range of the number of topics [100, 1000] and the step size of 100, and Fig. 1 shows the Chinese documents for English language paired documents (C → E), and Fig. 3 Search for Chinese Paired Documents for English Documents (E → C).

Table 2 Comparison of document pairing search performance between TDS and CL-LSI at a small number of topics
Fig. 1
figure 1

Drools schematic diagram

As can be seen from Table 2, when the number of topics is only 10, the accuracy of the Chinese document pairing search for English documents (C → E) and the English document pairing search for Chinese (E → C) is 0.5546 and 0.6056, while the accuracy of the two paired search tasks for the CL-LSI model is only about 0.34. When the number of topics is 20, the accuracy of the TDS model is about 0.85 and the CL-LSI model is about 0.73. When the number of subjects is 50, the accuracy of the two models are more than 0.90, and the TDS model is about 0.95, the advantage is more obvious. Therefore, we can conclude that, compared with the CL-LSI model, the TDS model can obtain more abundant bilingual semantic information by extracting a smaller number of bilingual topics, and the document matching search achieves a higher accuracy and can be retrieved Most of the paired documents.

From Figs. 2 and  3, it can be observed that starting from the number of topics of 100 and the number of topics, the accuracy of document paired search increases greatly. When reaching about 500, the search performance tends to be stable. When the number of topics is around 800, the performance changes hardly change. We paired the English documents (C → E) and Chinese documents (E → C) of the two models with a paired t test with a 95% confidence level, and Chinese documents paired with English documents (C → E) The bilateral p value is 0.070, and the bilateral p value of the English document pairing search Chinese (E → C) is 0.027. This shows that the TDS model performs significantly better than the CL-LSI model when the English documents are paired search for Chinese (E → C), but the advantages of pairing search for English documents (C → E) are not obvious. The possible reason for this phenomenon is that training Chinese documents contains English special nouns and other content, while English documents do not have Chinese content. When pairing searches using Chinese documents, both models can retrieve relevant documents by using English content in Chinese documents, and these English words are translated into a single document with more translations, so there is not much difference between the two models. However, when searching for Chinese pairing using English documents as a query, one needs to search for the corresponding translation or related term in Chinese, and on the other hand, English terms in Chinese documents also affect the similarity of the documents. The TDS model because of the training of the two languages separately modeled, the ability to distinguish more powerful language, so paired search performance advantages more models. Overall, the paired search performance of the TDS model is generally superior to that of the CL-LSI model.

Fig. 2
figure 2

Chinese search English (C → E)

Fig. 3
figure 3

English search Chinese (E → C)

4.2 Model parameter sensitivity analysis

In the TDS model, two important parameters that affect performance are involved: the number of bilingual topics k and the TF threshold. Figures 4 and 5 show the effect on the MAP performance of the TDS model when the number of bilingual topics k varies for the three document sets WSJFT, TREC-5 & 6 and TREC -9. The value of k in the figure is in the range of [100, 2000] and the step size is 100.

Fig. 4
figure 4

TREC-5&6

Fig. 5
figure 5

TREC-9

In contrast to the document set WSJFT, the TDS model has a slightly different impact on performance over TREC-5 & 6 and TREC-9 document sets. The results also list the MAP curves for these two documents, with TREC-5 & 6 on the left and TREC-9 on the right. Each graph shows the changes of MAP values of five different length queries of T, D, N, TD and TDN. When the number of subjects is less than 1000, the MAP value of the model grows faster; after 1000, the MAP of TREC5 & 6 increases relatively slowly, but the MAP value of TREC-9 keeps increasing when the number of the subjects is small. The possible reason for this phenomenon is that the data distribution of the TREC document set and the training document set is quite different, which makes the model unable to extract the semantic information more relevant to the query. Some queries (such as CH79 in TREC-9) do not have sufficient semantic information in the bilingual subject space. Because the TDS model does not have enough capacity for bilingual semantic interpretation of these queries, the model can extract semantic information related to the topic as much as possible even if the number of topics is large, but the relevance is not high enough to establish an effective bilingual relationship.

When TDS models train document sets, different TF thresholds mean vocabulary of different sizes. Vocabularies of different lengths result in the missing of some words, thus affecting the semantic expression of bilingual topics. The effect of different length bilingual vocabularies on bilingual presentation ability of TDS model was analyzed after different TF thresholds were set for analysis. TF values were taken 3, 5, 10, the number of terms used in training more obvious impact. Each time the threshold value of TF increases, the total length of Chinese vocabulary is reduced by about 6000 words each time, and the English words are reduced by about 4000 words. The benchmark model is a single-language VSM model, which is represented by “ML” in the table. The MAP values of different query lengths are listed in the third row. The experimental results in the table are divided into three groups, respectively, take different TF values. “ % Of ML” is expressed as the ratio of MAP values in single language VSM. On the TREC-5 & 6 document set, the best performance of the model is obtained when the TF value is 10, the TF value is 3, and the TF value is 5, the performance is the worst. On the TREC-9 document set, the MAP values of the TDS model did not change much when the TF thresholds were 3 and 5, but when the TF value was 10, the performance was optimal and there was a significant improvement. The MAP values of the 4 length queries Approaching or exceeding the single-language VSM model. The main reason is that the TDS model can effectively construct the bilingual subject space. The change of the quantity of a certain term does not have a significant impact on the model performance. Too many bilingual words will bring noise and increase the computational complexity of the model. For the change of TF threshold, it has little effect on the performance of TDS model.

5 Conclusion

The Internet in the context of globalization has been characterized by multiple languages and the user’s information search needs are no longer confined to the use of native languages. How to help users retrieve the knowledge they need from multi-lingual information resources quickly and effectively is a frontier research field in information retrieval. Cross-Language Information Retrieval (CLIR) is one of the effective ways to solve this problem. Users are familiar with the differences between native and non-native languages, which inevitably bring about language barriers for users to utilize the Internet. This paper treats bilingual parallel documents as two views of language description objects and assumes that bilingual parallel documents share the same semantic information. These views are equivalent in essence. In order to modelling these views and extract their semantic abstract content, this paper focuses on the use of Partial Least Squares (PLS) statistical analysis theory to extract a series of bilingual subject pairs with the same semantic information from bilingual parallel corpora, and it constructs a thematic space that represents the correspondence between bilingual semantics. At last, this paper establishes a cross-language information retrieval framework based on bilingual subject space.