1 Introduction

Moments have been used successfully in image processing and pattern classification applications for many years, as discriminative descriptors, not only because of their simplicity but also for the extraordinary variety of subjects and applications where they are enlightened. Moments are used to characterize different types of 2D and 3D images in many applications such image analysis [19, 24, 32, 41], pattern recognition [1, 8, 11, 17, 27], face recognition [9], image reconstruction [6, 28, 35], image watermarking [36, 38, 39], image compression [22], edge detection [2], medical image analysis [5, 30, 31], stereo image analysis [29, 37, 40].

Among all types of moments, discrete orthogonal moments have considerable properties: they are characterized by low information redundancy and high discriminating power. In addition, they eliminate the need for numerical approximation and precisely satisfy the orthogonal property [41]. Discrete orthogonal moments are defined by projecting an object/image onto a set of discrete orthogonal polynomials. According to the used type of polynomials, various discrete orthogonal moments can be obtained, such as Tchebichef [19], Krawtchouk [41], Charlier [16], Meixner [25], Hahn [43], and Racah moments [44]. In the classification/recognition field of 3D objects, the 3D object must be identified regardless of its position (translated or rotated) and its size (large or small) [10]. Therefore, invariance to geometric transformations is an essential criterion in these areas. This fundamental requirement pushes researchers to extract the invariants of 3D moments. Several works have been proposed in this context. The first work on 3D invariants was published by Sadjadi and Hall [23]. Their results were later rediscovered (with some modifications) by Guo [12]. These authors extracted only the translation, rotation, and scaling invariants from the second order's geometric moments without any possibility of further extension. Many algorithms were later proposed to generalize the computation of 3D geometric invariant moments [26, 33, 34].

At the beginning of the century, several works have been proposed for computing 3D discrete invariant moments. These invariants are algebraically derived from the invariants of 3D geometric moments. This conventional method (also known as the indirect method) has been extended from the 2D case to cover the 3D case. However, the numerical approximations used when calculating the invariants of 3D geometric moments may cause numerical errors in the computed moments. To solve this problem, Hosny [13] calculated the exact values of 2D geometric moments using mathematical integration of the monomial terms over digital image pixels. He also offered the possibility to extend this method to cover the 3D case. This method can be used to reduce numerical approximation errors in the calculation of 3D geometric moment invariants. Even if the numerical approximation errors can be reduced by this method, it is limited by the high cost of calculation. This makes invariants of 3D moments derived by this method undesirable, especially in real-time systems for which compliance with time constraints is a fundamental criterion. As a solution, the authors in [3, 46], and [42] developed a method to construct the translation and scaling invariants of 2D Legendre, Tchebichef, and Krawtchouk moments, respectively, directly from their corresponding orthogonal polynomials. Later, the authors in [7, 14], and [1] extended this method to derive the translation and scaling invariants of 3D Legendre, Tchebichef, and Krawtchouk moments, respectively.

Among 3D discrete orthogonal moments, the 3D Charlier moments (3DCMs) are defined as the projection of the 3D image/object onto the Charlier polynomials (CPs). The CPs satisfy exactly the orthogonality property in the discrete domain, so the implementation of these polynomials does not require any numerical approximation. In addition, these polynomials are discrete, and 3D digital images/objects also, so there is no discretization error [18]. CPs are defined by the hypergeometric function \({}_{2}F_{0} (.)\) unlike the Tchebichef, Krawtchouk, and Hahn polynomials, which are defined by the hypergeometric functions \({}_{3}F_{2} (.)\), \({}_{2}F_{1} (.)\) and \({}_{3}F_{2} (.)\), respectively, which means that CPs are much simpler and contain fewer terms (Pochhammer symbols and factorials) than other types of discrete orthogonal polynomials.

Motivated by the interesting properties of CPs, this paper proposes a fast method for computing 3D Charlier moment invariants to translation and scaling (3DCMITS) directly from CPs. We first develop the translated and scaled CPs by writing them in the form of simple summations. Then, we develop the algebraic expressions existing between the transformed CPs and the normal CPs. Finally, we use these expressions to derive 3DCMITS. This method takes into account the requirements of accuracy and calculation time. It considerably reduces the computation time and does not require any numerical approximation. The experimental results have shown that the proposed calculation method reduces the calculation time with a rate of more than 99,67% compared to the conventional method.

The proposed descriptors can be applied in applications that require invariance to translation and scaling transformations such as 3D object classification, 3D image recognition, and protection of intellectual property of 3D meshes. For validation purposes and as an illustration of the interest of the proposed descriptors, this paper proposes a classification system for 3D objects based on the proposed 3DCMITS and Support Vector Machine (SVM) classifier. The results are verified with K-Nearest Neighbor (KNN) classifier and other existing works in the literature.

The rest of the paper is organized as follows. In Sect. 2, definitions of CPs, 3DCMs and conventional invariants of 3DCMs are presented. Section 3 presents the proposed method for calculating 3DCMITS. Algorithms for calculating translation and scale invariants using the proposed method are also given in this section. Experimental results for evaluating the performance of the proposed descriptors are given in Sect. 5. Section 6 concludes this paper.

2 3D Charlier Moments

In this section, we will present the mathematical framework of the theory of 3D Charlier moments, including the Charlier polynomials, the 3D Charlier moments and the conventional translation and scaling invariants of the 3D Charlier moments.

2.1 Charlier Polynomials

The \(n\) th Charlier polynomials (CPs) are defined by using the hypergeometric function as follows [21]:

$$ \left. \begin{gathered} C_{n}^{{a_{1} }} (x) = {}_{2}F_{0} ( - n, - x; - a_{1}^{ - 1} ) \hfill \\ { = }\sum\limits_{k = 0}^{n} {\frac{{( - n)_{k} ( - x)_{k} ( - a_{1}^{ - 1} )^{k} }}{k!} \, } \, \hfill \\ \end{gathered} \right| \, x,n = 0,1,2,.....\infty $$
(1)

where \(a_{1}\) is restricted to \(a_{1} > 0\), and \({}_{2}F_{0} (.)\) is the hypergeometric function, defined as:

$$ {}_{2}F_{0} (a,b;c) = \sum\limits_{k = 0}^{n} {\frac{{(a)_{k} (b)_{k} (c)^{k} }}{k!}} $$
(2)

and \((a)_{k}\) is the Pochhammer symbol, defined as:

$$ \begin{gathered} (a)_{0} = 1 \hfill \\ (a)_{k} = \frac{(a + k - 1)!}{{(a - 1)!}}{ ; }k \ge 1 \, \hfill \\ \end{gathered} $$
(3)

Equation (1) can be rewritten as follows:

$$ \begin{aligned} C_{n}^{{a_{1} }} (x) &= \sum\limits_{k = 0}^{n} {\frac{n!}{{k!(n - k)!( - a_{1} )^{k} }}\langle x\rangle_{k} } \, \hfill \\ &=\sum\limits_{k = 0}^{n} {\alpha_{n,k} } \langle x\rangle_{k} \hfill \\ \end{aligned} $$
(4)

where \(\alpha_{n,k}\) are the coefficients of the CPs given by:

$$ \alpha_{n,k} = \frac{n!}{{k!(n - k)!( - a_{1} )^{k} }} $$
(5)

and \(\langle x\rangle_{k}\) is the falling factorial defined as [4]:

$$ \langle x\rangle_{k} = \sum\limits_{i = 0}^{k} {s(k,i)x^{i} } $$
(6)

where \(s(k,i)\) are the Stirling numbers of the first kind.

The CPs satisfy the following orthogonal condition:

$$ \sum\limits_{x = 0}^{\infty } {\omega (x)C_{n}^{{a_{1} }} (x)C_{m}^{{a_{1} }} (x)} = \rho (n)\delta_{n,m} { ; }n,m \ge 0 $$
(7)

where \(\omega (x)\) and \(\rho (n)\) are the weight function and the squared norm of CPs, respectively, defined as:

$$ \omega (x) = \frac{{e^{{ - a_{1} }} a_{1}^{x} }}{x!} $$
(8)
$$ \rho (n) = \frac{n!}{{a_{1}^{n} }} $$
(9)

The normalized CPs with respect to the norm \(\tilde{C}_{n}^{{a_{1} }} (x)\) are defined as:

$$ \tilde{C}_{n}^{{a_{1} }} (x) = \frac{{C_{n}^{{a_{1} }} (x)}}{{\sqrt {\rho (n)} }} $$
(10)

To obtain numerical stability, the weighted CPs \(\overline{C}_{n}^{{a_{1} }} (x)\) are used. They are defined as follows:

$$ \overline{C}_{n}^{{a_{1} }} (x) = C_{n}^{{a_{1} }} (x)\sqrt {\frac{\omega (x)}{{\rho (n)}}} $$
(11)

The CPs satisfy the following recurrence relationship [45]:

$$ \overline{C}_{n}^{{a_{1} }} (x) = - \sqrt {\frac{{a_{1} }}{n}} \frac{{(x - n + 1 - a_{1} )}}{{a_{1} }}\overline{C}_{n - 1}^{{a_{1} }} (x) - \sqrt {\frac{{a_{1}^{2} }}{n(n - 1)}} \frac{(n - 1)}{{a_{1} }}\overline{C}_{n - 2}^{{a_{1} }} (x) $$
(12)

where

$$ \overline{C}_{0}^{{a_{1} }} (x) = \sqrt {\frac{\omega (x)}{{\rho (0)}}} {\text{ and }}\overline{C}_{1}^{{a_{1} }} (x) = \frac{{a_{1} - x}}{{a_{1} }}\sqrt {\frac{\omega (x)}{{\rho (1)}}} $$
(13)

2.2 3D Charlier Moments

Given a 3D image/object \(I(x,y,z)\) with size \(N \times M \times K\), the \((n + m + k)\) order 3D Charlier moments (3DCMs) of the image/object \(I(x,y,z)\) are defined as follows:

$$ CM_{n,m,k} = \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\overline{C}_{n}^{{a_{1} }} (x)\overline{C}_{m}^{{a_{1} }} (y)\overline{C}_{k}^{{a_{1} }} (z)} } } I(x,y,z) $$
(14)

Equation (7) leads to the following inverse transform

$$ I_{rec} (x,y,z) = \sum\limits_{n = 0}^{N - 1} {\sum\limits_{m = 0}^{M - 1} {\sum\limits_{k = 0}^{K - 1} {\overline{C}_{n}^{{a_{1} }} (x)\overline{C}_{m}^{{a_{1} }} (y)\overline{C}_{k}^{{a_{1} }} (z)} } } CM_{n,m,k} $$
(15)

where \(I_{rec} (x,y,z)\) is the reconstructed 3D image/object.

The difference between the original image/object \(I(x,y,z)\) and the reconstructed one \(I_{rec} (x,y,z)\) is measured by the Mean Square Error (MSE), which is defined as follows:

$$ {\text{MSE}} = \frac{1}{N \times M \times K}\sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\left( {I(x,y,z) - I_{rec} (x,y,z)} \right)^{2} } } } $$
(16)

The 3DCMs presented in this subsection are not invariant under the geometric transformations, such as translation and uniform/non-uniform scaling. In the next subsection, we present the conventional method for deriving the translation and scaling invariants of 3D Charlier moments.

2.3 Derivation of Translation and Scaling Invariants of 3D Charlier Moments Using 3D Geometric Moments

The conventional method to obtain the translation and scaling invariants of 3D Charlier moments is to derive them algebraically from the corresponding invariants of 3D geometric moments. This conventional method is called the indirect method. This subsection provides a brief introduction of this method.

The 3DCMs of a normalized image/object \(\hat{I}(x,y,z) = [\omega (x)\omega (y)\omega (z)]^{{{{ - 1} \mathord{\left/ {\vphantom {{ - 1} 2}} \right. \kern-\nulldelimiterspace} 2}}} I(x,y,z)\) can be developed as follows:

$$ \begin{aligned} CM_{n,m,k} &= \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\overline{C}_{n}^{{a_{1} }} (x)\overline{C}_{m}^{{a_{1} }} (y)\overline{C}_{k}^{{a_{1} }} (z)} } } \hat{I}(x,y,z) \hfill \\ &= \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\tilde{C}_{n}^{{a_{1} }} (x)\tilde{C}_{m}^{{a_{1} }} (y)\tilde{C}_{k}^{{a_{1} }} (z)} } } I(x,y,z) \hfill \\ &= [\rho {\text{(n)}}\rho {\text{(m)}}\rho {\text{(k)]}}^{{ - {1 \mathord{\left/ {\vphantom {1 2}} \right. \kern-\nulldelimiterspace} 2}}} \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {C_{n}^{{a_{1} }} (x)C_{m}^{{a_{1} }} (y)C_{k}^{{a_{1} }} (z)} } } I(x,y,z) \hfill \\ &= [\rho {\text{(n)}}\rho {\text{(m)}}\rho {\text{(k)]}}^{{ - {1 \mathord{\left/ {\vphantom {1 2}} \right. \kern-\nulldelimiterspace} 2}}} \sum\limits_{q = 0}^{n} {\sum\limits_{l = 0}^{m} {\sum\limits_{r = 0}^{k} {\chi_{n,q} \chi_{m,,l} \chi_{k,r} } } } m_{q,l,r} \hfill \\ \end{aligned} $$
(17)

where \(\chi_{n,i}\) are coefficients of CPs, defined as

$$ \chi_{n,i} = \sum\limits_{k = 0}^{n - i} {\alpha_{n,n - k} s(n - k,i)} $$
(18)

and \(m_{q,l,r}\) are the 3D geometric moments (3DGMs) of order \((q + l + r)\) of the 3D image/object \(I(x,y,z)\). The 3DGMs are defined using the discrete approximation as follows [10]:

$$ m_{q,l,r} = \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {x^{q} y^{l} z^{r} } } } I(x,y,z) $$
(19)

The 3D geometric moment invariants to translation (3DGMIT) \(\mu_{n,m,k}\) (also called the central geometric moments) are defined as follows [10]:

$$ \begin{gathered} \mu_{n,m,k} = \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {(x - \overline{x})^{n} (y - \overline{y})^{m} (z - \overline{z})^{k} } } } I(x,y,z) \hfill \\ { = }\sum\limits_{q = 0}^{n} {\sum\limits_{l = 0}^{m} {\sum\limits_{r = 0}^{k} {\left( \begin{gathered} n \hfill \\ q \hfill \\ \end{gathered} \right)\left( \begin{gathered} m \hfill \\ l \hfill \\ \end{gathered} \right)\left( \begin{gathered} k \hfill \\ r \hfill \\ \end{gathered} \right)( - 1)^{q + l + r} (\overline{x})^{q} (\overline{y})^{l} (\overline{z})^{r} } } } m_{n - q,m - l,k - r} \hfill \\ \end{gathered} $$
(20)

where \(\overline{x}\), \(\overline{y}\), and \(\overline{z}\) are the centroid coordinates of the 3D image/object \(I(x,y,z)\), defined as:

$$ \overline{x} = \frac{{m_{1,0,0} }}{{m_{0,0,0} }} \, , \, \overline{y} = \frac{{m_{0,1,0} }}{{m_{0,0,0} }} \, , \, \overline{z} = \frac{{m_{0,0,1} }}{{m_{0,0,0} }} $$
(21)

The 3DGMIT change in case of uniform/non-uniform scaling of the 3D image/object is expressed as follows:

$$ \begin{gathered} \mu^{\prime}_{n,m,k} = \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {a^{n} (x - \overline{x})^{n} b^{m} (y - \overline{y})^{m} c^{k} (z - \overline{z})^{k} } } } abcI(x,y,z) \hfill \\ { = }a^{n + 1} b^{m + 1} c^{k + 1} \mu_{n,m,k} \hfill \\ \end{gathered} $$
(22)

where \(a\), \(b\), and \(c\) are the scaling factors in the \(x\)-, \(y\)-, and \(z\)-direction of the 3D image/object, respectively.

The 3DGM invariants to translation and scaling (3DGMITS), denoted \(V_{n,m,k}\), can be obtained using the following normalization [10]:

$$ V_{n,m,k} = \frac{{\mu_{0,0,0}^{\xi + 2} }}{{\mu_{(n + \xi ),0,0} \mu_{0,(m + \xi ),0} \mu_{0,0,(k + \xi )} }}.\mu_{n,m,k} { ; }\xi = 0,1,2,..... $$
(23)

If we replace the 3DGMs \(m_{q,l,r}\) in Eq. (17) by 3DGMIT \(\mu_{q,l,r}\) and by 3DGMITS \(V_{q,l,r}\), we obtain the 3D Charlier moment invariants to translation (3DCMIT) and 3D Charlier moment invariants to translation and scaling (3DCMITS), respectively. However, this indirect method has two major drawbacks:

  • It is dependent on the numerical approximations used during the calculation of 3DGMIT in Eq. (20) and 3DGMITS in Eq. (23) which produces approximation errors.

  • It needs a lot of time to extract the 3DCMITS, because the coefficients of the CPs \(\alpha_{n,k}\) are time consuming.

These drawbacks limit the use of the indirect method: The first drawback can influence the accuracy of the results in all applications where 3DCMITS can be applied such as object classification, image recognition, and digital watermarking of 3D meshes. The second drawback makes the use of 3DCMITS impossible in real-time systems where compliance with time constraints in the execution of processing is very important, as is the accuracy of these processing results. This paper proposes a quick method to solve these drawbacks, without numerical approximation, to obtain 3DCMITS. This method is presented in the following section.

3 Proposed 3D Charlier Moment Invariants to Translation and Scaling

In this section, we develop a new method to derive the 3D Charlier moment invariants to translation and scaling (3DCMITS). This method uses the particular properties of the Charlier polynomials (CPs) to directly derive the 3DCMITS without going through the computation of 3D geometric moments (3DGMs). In the first subsection, we present the mathematical framework for algebraically deriving the 3D Charlier moment invariants to translation (3DCMIT), while in the second subsection, we derive the 3D Charlier moment invariants to scaling (3DCMIS). Finally, we offer in the third subsection the algorithms for calculating these descriptors.

3.1 3D Charlier Moment Invariants to Translation

The 3DCMIT of order (\(n + m + k\)) are noted '' \(\varphi_{n,m,k}\) '' for abbreviation.

For a normalized 3D image/object \(\hat{I}(x,y,z) = [\omega (x)\omega (y)\omega (z)]^{{{{ - 1} \mathord{\left/ {\vphantom {{ - 1} 2}} \right. \kern-\nulldelimiterspace} 2}}} I(x,y,z)\), the direct definition of 3DCMIT can be achieved simply by shifting the origin of coordinates at the image/objet centroid, as following:

$$ \varphi_{n,m,k} = \sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\tilde{C}_{n}^{{a_{1} }} (x - \overline{x})\tilde{C}_{m}^{{a_{1} }} (y - \overline{y})\tilde{C}_{k}^{{a_{1} }} (z - \overline{z})} } } I(x,y,z) $$
(24)

where \(\overline{x}\), \(\overline{y}\) and \(\overline{z}\) are given in Eq. (22).

In the following, we will establish the relations between the translated CPs {\(C_{n}^{{a_{1} }} (x - \overline{x})\), \(C_{m}^{{a_{1} }} (y - \overline{y})\), \(C_{k}^{{a_{1} }} (z - \overline{z})\)} and the normal CPs {\(C_{n}^{{a_{1} }} (x)\), \(C_{m}^{{a_{1} }} (y)\), \(C_{k}^{{a_{1} }} (z)\)}. In this development, we are only interested in CPs translated in the \(x\)-direction, where the results found can be obtained in the same way for CPs translated in the \(y\)- and \(z\)-direction, respectively.

Using Eq. (4), the CPs translated in the \(x\)-direction can be developed as follows:

$$ \begin{aligned} C_{n}^{{a_{1} }} (x - \overline{x}) &= \sum\limits_{k = 0}^{n} {\alpha_{n,k} \langle x - \overline{x}\rangle_{k} } \hfill \\ &= \alpha_{n,n} \langle x - \overline{x}\rangle_{n} + \alpha_{n,n - 1} \langle x - \overline{x}\rangle_{n - 1}+ \alpha_{n,n - 2} \langle x - \overline{x}\rangle_{n - 2} \hfill \\ &\quad + \,\alpha_{n,n - 3} \langle x - \overline{x}\rangle_{n - 3} ....................... + \alpha_{n,k} \langle x - \overline{x}\rangle_{k} \hfill \\ \end{aligned} $$
(25)

where

$$ \langle x - \overline{x}\rangle_{k} = \sum\limits_{l = 0}^{k} {\left( \begin{gathered} k \hfill \\ l \hfill \\ \end{gathered} \right)} \langle \overline{x}\rangle_{l} \langle x\rangle_{k - l} $$
(26)

Using Eq. (26), the CPs translated in the \(x\)-direction become:

$$ \begin{aligned} C_{n}^{{a_{1} }} (x - \overline{x}) &= \alpha_{n,n} \langle x\rangle_{n} \hfill \\ &\quad + \left[ {\alpha_{n,n - 1} + \alpha_{n,n} \left( \begin{gathered} n \hfill \\ 1 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{1} } \right]\langle x\rangle_{n - 1} \hfill \\& \quad + \left[ {\alpha_{n,n - 2} + \alpha_{n,n - 1} \left( \begin{gathered} n - 1 \hfill \\ \, 1 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{1} + \alpha_{n,n} \left( \begin{gathered} n \hfill \\ 2 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{2} } \right]\langle x\rangle_{n - 2} \hfill \\ &\quad+ \left[ {\alpha_{n,n - 3} + \alpha_{n,n - 2} \left( \begin{gathered} n - 2 \hfill \\ \, 1 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{1} + \alpha_{n,n - 1} \left( \begin{gathered} n - 1 \hfill \\ \, 2 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{2} + \alpha_{n,n} \left( \begin{gathered} n \hfill \\ 3 \hfill \\ \end{gathered} \right)\langle - \overline{x}\rangle_{3} } \right]\langle x\rangle_{n - 3} \hfill \\ \, + ........................ \hfill \\ \end{aligned} $$
(27)

Substituting the normal CPs into the series of decreasing falling factorials of \(x\), we find the following expression:

$$ C_{n}^{{a_{1} }} (x - \overline{x}) = \sum\limits_{k = 0}^{n} {v_{n,n - k} ( - \overline{x})} \, C_{n - k}^{{a_{1} }} (x) $$
(28)

where

$$ \begin{gathered} v_{n,n} ( - \overline{x}) = 1 \hfill \\ v_{n,n - k} ( - \overline{x}) = \frac{1}{{\alpha_{n - k,n - k} }}\left[ {\sum\limits_{l = 0}^{k} {\left( \begin{gathered} n - k + l \hfill \\ \, l \hfill \\ \end{gathered} \right)\alpha_{n,n - k + l} \langle - \overline{x}\rangle_{l} } \, - \sum\limits_{i = 0}^{k - 1} {\alpha_{n - i,n - k} v_{n,n - i} ( - \overline{x})} } \right]{ ; }k \le n \hfill \\ \end{gathered} $$
(29)

The following relationships can be taken from Eq. (29):

$$ v_{n,n} ( - \overline{x}) = 1 $$
(30)
$$ v_{n,n - 1} ( - \overline{x}) = \frac{n}{{( - a_{1} )}}\langle - \overline{x}\rangle_{1} $$
(31)
$$ v_{n,n - 2} ( - \overline{x}) = \frac{n(n - 1)}{{2(a_{1} )^{2} }}\langle - \overline{x}\rangle_{2} $$
(32)
$$ v_{n,n - 3} ( - \overline{x}) = \frac{n(n - 1)(n - 2)}{{3!( - a_{1} )^{3} }}\langle - \overline{x}\rangle_{3} $$
(33)
$$ v_{n,n - 4} ( - \overline{x}) = \frac{n(n - 1)(n - 2)(n - 3)}{{4!( - a_{1} )^{4} }}\langle - \overline{x}\rangle_{4} $$
(34)

By examining Eqs. (3034), we find that the expression \(v_{n,n - k} ( - \overline{x})\) can be written in the following form:

$$ \begin{gathered} v_{n,n - k} ( - \overline{x}) = \frac{n(n - 1)(n - 2)......(n - k + 1)}{{k!( - a_{1} )^{k} }}\langle - \overline{x}\rangle_{k} \hfill \\ \, = \frac{{\langle n\rangle_{k} }}{{k!( - a_{1} )^{k} }}\langle - \overline{x}\rangle_{k} = f(n,k)\langle - \overline{x}\rangle_{k} \hfill \\ \end{gathered} $$
(35)

where

$$ f(n,k) = \frac{{\langle n\rangle_{k} }}{{k!( - a_{1} )^{k} }} $$
(36)

Since we are interested in the normalized CPs defined by Eq. (10), we will generalize Eqs. (35) and (36).

By using Eqs. (4) and (10), the normalized normal CPs \(\tilde{C}_{n}^{{a_{1} }} (x)\) and the normalized translated CPs \(\tilde{C}_{n}^{{a_{1} }} (x - \overline{x})\) are expressed as follows:

$$ \tilde{C}_{n}^{{a_{1} }} (x) = \sum\limits_{k = 0}^{n} {\tilde{\alpha }_{n,n - k} } \langle x\rangle_{k} $$
(37)
$$ \tilde{C}_{n}^{{a_{1} }} (x - \overline{x}) = \sum\limits_{k = 0}^{n} {\tilde{v}_{n,n - k} ( - \overline{x})} \, \tilde{C}_{n - k}^{{a_{1} }} (x) $$
(38)

where

$$ \tilde{\alpha }_{n,n - k} = \frac{{\alpha_{n,k} }}{{\sqrt {\rho (n)} }} $$
(39)
$$ \tilde{v}_{n,n - k} ( - \overline{x}) = \frac{1}{{\tilde{\alpha }_{n - k,n - k} }}\left[ {\sum\limits_{l = 0}^{k} {\left( \begin{gathered} n - k + l \hfill \\ \, l \hfill \\ \end{gathered} \right)\tilde{\alpha }_{n,n - k + l} \langle - \overline{x}\rangle_{l} } \, - \sum\limits_{i = 0}^{k - 1} {\tilde{\alpha }_{n - i,n - k} \tilde{v}_{n,n - i} ( - \overline{x})} } \right] $$
(40)

Using Eqs. (38) and (40), the following expression can be derived using the same procedures as in Eq. (38):

$$ \tilde{v}_{n,n - k} ( - \overline{x}) = \tilde{f}(n,k)\langle - \overline{x}\rangle_{k} $$
(41)

where

$$ \tilde{f}(n,k) = \frac{{\langle n\rangle_{k} }}{{k!( - a_{1} )^{k} }}\sqrt {\frac{\rho (n - k)}{{\rho (n)}}} { ; }k \le n $$
(42)

Finally, the normalized translated CPs can be expressed in terms of normalized normal CPs as follows:

$$ \tilde{C}_{n}^{{a_{1} }} (x - \overline{x}) = \sum\limits_{q = 0}^{n} {\tilde{f}(n,k)\langle - \overline{x}\rangle_{q} } \tilde{C}_{n - q}^{{a_{1} }} (x) $$
(43)

The result of Eq. (43) can also be generalized for normalized translated CPs in the \(y\) and \(z\) directions as follows:

$$ \tilde{C}_{m}^{{a_{1} }} (y - \overline{y}) = \sum\limits_{l = 0}^{m} {\tilde{f}(m,l)\langle - \overline{y}\rangle_{l} } \tilde{C}_{m - l}^{{a_{1} }} (y) $$
(44)
$$ \tilde{C}_{k}^{{a_{1} }} (z - \overline{z}) = \sum\limits_{r = 0}^{k} {\tilde{f}(k,r)\langle \overline{z}\rangle_{r} } \tilde{C}_{k - r}^{{a_{1} }} (z) $$
(45)

By substituting Eqs. (43) - (45) in Eq. (24), the 3DCMIT become:

$$ \varphi_{n,m,k} = \sum\limits_{q = 0}^{n} {\sum\limits_{l = 0}^{m} {\sum\limits_{r = 0}^{k} {\tilde{f}(n,q)\tilde{f}(m,l)\tilde{f}(r,k)} \langle - \overline{x}\rangle_{q} \langle - \overline{y}\rangle_{l} \langle - \overline{z}\rangle_{r} } } CM_{n - q,m - l,k - r} $$
(46)

Eq. (46) shows that 3DCMIT are derived directly from 3DCMs.

It should be noted that the expression \(\tilde{f}(.)\) in Eq. (46) is independent of the 3D image/object. Therefore, this expression can be pre-computed, stored, recalled whenever it is needed to avoid repetitive computation.

3.2 3D Charlier Moment Invariants to Scaling

Either a 3D image/object is scaled non-uniformly with the scaling factors \(a\), \(b\) and \(c\) in the \(x\)-, \(y\)- and \(z\)-direction, respectively, the 3DCMs are changed as follows:

$$ CM^{\prime}_{n,m,k} = abc\sum\limits_{x = 0}^{N - 1} {\sum\limits_{y = 0}^{M - 1} {\sum\limits_{z = 0}^{K - 1} {\tilde{C}_{n}^{{a_{1} }} (ax)} \tilde{C}_{m}^{{a_{1} }} (by)} } \tilde{C}_{k}^{{a_{1} }} (cz)I(x,y,z) $$
(47)

In the following, we will establish the relations between the normalized scaled CPs {\(\tilde{C}_{n}^{{a_{1} }} (ax)\), \(\tilde{C}_{m}^{{a_{1} }} (by)\), \(\tilde{C}_{k}^{{a_{1} }} (cz)\)} and the normalized normal CPs {\(\tilde{C}_{n}^{{a_{1} }} (x)\), \(\tilde{C}_{m}^{{a_{1} }} (y)\), \(\tilde{C}_{k}^{{a_{1} }} (z)\)}. We are interested in this development only in the normalized CPs scaled in the \(x\)-direction, where the results found can be obtained in the same way for the normalized CPs scaled in the \(y\)- and \(z\)-direction, respectively.

The normalized CPs scaled in the \(x\)-direction can be expressed as a series of decreasing powers of \(x\), using Eqs. (7) and (5), as follows:

$$ \begin{gathered} \tilde{C}_{n}^{{a_{1} }} (ax) = \sum\limits_{k = 0}^{n} {\tilde{\alpha }_{n,k} } \langle ax\rangle_{k} = \sum\limits_{k = 0}^{n} {\tilde{\alpha }_{n,k} } \sum\limits_{i = 0}^{k} {s(k,i)a^{i} x^{i} } \hfill \\ \,\qquad = \sum\limits_{k = 0}^{n} {\sum\limits_{i = 0}^{k} {\tilde{\alpha }_{n,n - k} s(k,i)a^{i} x^{i} } } = \sum\limits_{i = 0}^{n} {\sum\limits_{k = 0}^{n - i} {\tilde{\alpha }_{n,n - k} s(n - k,i)a^{i} x^{i} } } \hfill \\ \,\qquad = \sum\limits_{i = 0}^{n} {\tilde{\chi }_{n,i} a^{i} x^{i} } \hfill \\ \end{gathered} $$
(48)

where

$$ \chi_{n,i} = \sum\limits_{k = 0}^{n - i} {\tilde{\alpha }_{n,n - k} s(n - k,i)} $$
(49)

It can be easily deduced from Eq. (48) the relationship between \(\tilde{C}_{n}^{{a_{1} }} (ax)\) and \(\tilde{C}_{n}^{{a_{1} }} (x)\) as follows:

$$ \sum\limits_{q = 0}^{n} {\gamma_{n,q} \tilde{C}_{q}^{{a_{1} }} (ax)} = a^{n} \sum\limits_{q = 0}^{n} {\gamma_{n,q} \tilde{C}_{q}^{{a_{1} }} (x)} $$
(50)

where

$$ \begin{gathered} \gamma_{n,n} = 1 \, \hfill \\ \gamma_{n,q} = \sum\limits_{i = 0}^{n - q - 1} { - \frac{{\chi_{n - i,q} \gamma_{n,n - i} }}{{\chi_{q,q} }}} \hfill \\ \end{gathered} $$
(51)

The normalized CPs scaled in the \(y\)- and \(z\)-direction can also be expressed as a function of normalized normal CPs as follows:

$$ \sum\limits_{l = 0}^{m} {\gamma_{m,l} \tilde{C}_{l}^{{a_{1} }} (by)} = b^{m} \sum\limits_{l = 0}^{m} {\gamma_{m,l} \tilde{C}_{l}^{{a_{1} }} (y)} $$
(52)
$$ \sum\limits_{r = 0}^{k} {\gamma_{k,r} \tilde{C}_{r}^{{a_{1} }} (cz)} = c^{k} \sum\limits_{r = 0}^{k} {\gamma_{k,r} \tilde{C}_{r}^{{a_{1} }} (z)} $$
(53)

From Eqs. (47), (50), (52), (53), we deduce the relationship between scaled 3DCMs and normal 3DCMs:

$$ \begin{gathered} \psi_{n,m,k} = \sum\limits_{q = 0}^{n} {\sum\limits_{l = 0}^{m} {\sum\limits_{r = 0}^{k} {\gamma_{n,q} \gamma_{m,l} \gamma_{k,r} } CM^{\prime}_{q,l,r} } } \hfill \\\qquad \, = a^{n + 1} b^{m + 1} c^{k + 1} \sum\limits_{q = 0}^{n} {\sum\limits_{l = 0}^{m} {\sum\limits_{r = 0}^{k} {\gamma_{n,q} \gamma_{m,l} \gamma_{k,r} } CM_{q,l,r} } } \hfill \\ \end{gathered} $$
(54)

The following relationships can be drawn from Eq. (54):

$$ \begin{gathered} 1) \, \psi_{0,0,0} = abcCM_{0,0,0} \hfill \\ 2) \, \psi_{n,0,0} = a^{n + 1} bc\sum\limits_{q = 0}^{n} {\gamma_{n,q} CM_{q,0,0} } \hfill \\ 3) \, \psi_{0,m,0} = ab^{m + 1} c\sum\limits_{l = 0}^{m} {\gamma_{m,l} CM_{0,l,0} } \hfill \\ 4) \, \psi_{0,0,k} = abc^{k + 1} \sum\limits_{r = 0}^{k} {\gamma_{k,r} CM_{0,0,k} } \hfill \\ \end{gathered} $$
(55)

The scaling factors (\(a\), \(b\), and \(c\)) in \(\psi_{n,m,k}\) can be canceled using the following normalization:

$$ \phi_{n,m,k} = \frac{{\psi_{n,m,k} \psi_{0,0,0}^{\xi + 2} }}{{\psi_{(n + \xi ),0,0} \psi_{0,(m + \xi ),0} \psi_{0,0,(k + \xi )} }}{ ; }\xi = 0,1,2,..... $$
(56)

The \(\phi_{n,m,k}\) are uniform/non-uniform scaling invariants of 3D Charlier moments.

Proof

Let \(\phi^{\prime}_{n,m,k}\) denotes the scaled version of \(\phi_{n,m,k}\), the relationship between them is:

$$\begin{aligned} \phi _{{n,m,k}}^{\prime } = & \frac{{\psi _{{n,m,k}}^{\prime } \psi _{{0,0,0}}^{{\prime \xi + 2}} }}{{\psi _{{(n + \xi ),0,0}}^{\prime } \psi _{{0,(m + \xi ),0}}^{\prime } {\text{ }}\psi _{{0,0,(k + \xi )}}^{\prime } }} \\ {\text{ }} = & \frac{{a^{{n + 1}} b^{{m + 1}} c^{{k + 1}} \psi _{{n,m,k}} {\text{ }}a^{{\xi + 2}} b^{{\xi + 2}} c^{{\xi + 2}} \psi _{{0,0,0}}^{{\xi + 2}} }}{{a^{{n + 1 + \xi }} b^{\xi } c^{\xi } \psi _{{(n + \xi ),0,0}} {\text{ }}a^{\xi } c^{\xi } b^{{m + 1 + \xi }} \psi _{{0,(m + \xi ),0}} {\text{ }}a^{\xi } b^{\xi } c^{{k + 1 + \xi }} \psi _{{0,0,(k + \xi )}} }} \\ {\text{ }} = & \frac{{\psi _{{0,0,0}}^{{\xi + 2}} }}{{\psi _{{(n + \xi ),0,0}} {\text{ }}\psi _{{0,(m + \xi ),0}} {\text{ }}\psi _{{0,0,(k + \xi )}} }} \cdot \psi _{{n,m,k}} \\ {\text{ }} = & \phi _{{n,m,k}} \\ \end{aligned}$$
(57)

The proof is completed.

Eq. (56) shows that 3DCMIS are directly derived from 3DCM.

3.3 Algorithms

After having calculated 3DCMIT and 3DCMIS by the proposed method, we summarize the steps for calculating these descriptors in the algorithms1 and 2.

Algorithm 1: Computation of 3D Charlier moment invariants to translation (3DCMIT).

figure a

Algorithm 2: Computation of 3D Charlier moment invariants to scaling (3DCMIS).

figure b

The algorithms 1 and 2 can be combined to obtain the 3D Charlier moment invariants to translation and scaling (3DCMITS). The 3DCMITS of order \((n + m + k)\) are noted ''\(\Phi_{n,m,k}\)'' for abbreviation.

4 Experimental Results

In this section, we give experimental results to validate the performance of the proposed descriptors. This section is divided into four subsections. In the first subsection, we test the ability of 3DCMs to reconstruct 3D images/objects. In the second subsection, we evaluate and compare the invariability of 3DCMIT and 3DCMIS derived by the proposed method and by the indirect method. In the third subsection, we compare the calculation time of 3DCMIT and 3DCMIS by the proposed method and the indirect method. In the last subsection, we present a classification system for 3D objects based on the proposed 3DCMITS and the SVM and KNN classifiers in both noise-free and noisy conditions.

4.1 3D object reconstruction using 3D Charlier Moments

The reconstruction of the original image/object from a set of its moments has been discussed in the literature quite often, because the reconstruction abilities of moments are connected with their classification power. Studying the image/object reconstruction problem provides an insight into moment properties that are important for classification without performing classification experiments. That's why so many authors have discussed image reconstruction from different moments in their papers [10]. In this subsection, we show the ability of 3DCMs to reconstruct 3D images/objects. In the first test, we use three objects "Ant", "Human", and "Teddy-bear" of size \(128 \times 128 \times 128\) (Fig. 1) extracted from the McGill 3D Shape Benchmark database [47]. The reconstruction tests are performed using 3DCMs with various values of moment order, ranging from 1 up to 125. Figure 2 shows the MSEs between the original objects and the reconstructed ones according to the reconstruction order ranging from 1 to 125. A set of reconstructed objects is shown in Fig. 3. Figure 2 show that MSEs decrease with increasing order of reconstruction. It is also shown that from order 60 (\(n = 20, \, m = 20, \, k = 20\)) the MSEs tend toward zero, which means that the reconstructed objects are very similar to the original ones. This is clearly shown in Fig. 3. These results show the ability of the 3DCMs to reconstruct 3D objects from a limited number of moments. Therefore, the lower order 3DCMs can replace a 3D object in all applications in order to reduce processing time.

Fig. 1
figure 1

Original 3D objects of size 128 × 128 × 128; a "Ant", b "Human", and c "Teddy-bear "

Fig. 2
figure 2

MSE of "Ant", "Human", and "Teddy-bear" 3D objects using 3DCMs

Fig. 3
figure 3

Reconstruction of "Ant", "Human", and "Teddy-bear " objects using 3DCMs for different classification order

In the second test, we compare the ability of 3DCMs to reconstruct 3D medical images. For this, we use a 3D Magnetic Resonance Image (3DMRI) of size \(100 \times 100 \times 100\) which is shown in Fig. 4. The capacity of 3DCMs is compared with that of other types of discrete orthogonal moments, such as 3D Krawtchouk moments (3DKMs), 3D Tchebichef moments (3DTMs), and 3D Hahn moments (3DHMs). Figure 5 shows the results of this comparison. This figure shows that the MSEs of all these moments tend towards zero with the increase in reconstruction order, which means that these moments are able to reconstruct the 3D medical images that have a high degree of complexity. It is also shown that the reconstruction performance of 3DCMs, 3DKMs, 3DTMs, and 3DHMs are very similar to each other, with reconstruction errors of the same order of magnitude \((10^{ - 3} )\), with a relative advantage of 3DCMs and 3DKMs when using low and high reconstruction orders, respectively. The results of this subsection clearly show that 3DCMs can successfully describe 3D images/objects with a limited number of orders, which means that these descriptors are desirable in applications that suffer from computational complexity.

Fig. 4
figure 4

3D Magnetic Resonance Image (3DMRI) of size 100 × 100 × 100

Fig. 5
figure 5

a Comparison of the reconstruction errors for 3DMRI by using 3DCMs, 3DKMs, 3DTMs, and 3DHMs. b Enlarge part of (a)

4.2 Invariability

In this subsection, we test the invariability of the proposed 3DCMIT and 3DCMIS under the geometric transformations of translation and uniform/non-uniform scaling. For this purpose, two 3D objects "Four" and "Airplane" (Fig. 6) of size \(128 \times 128 \times 128\) extracted from the McGill 3D Shape Benchmark database [47], are used as test objects. In order to measure the capacity of the proposed invariants to remain unchanged under the translation and uniform/non-uniform scaling, we use as objective criterion the deviation of the moments, which is defined as follows:

$$ X = \frac{\sigma }{|\mu |} \times (100 \, \% ) $$
(58)

where \(\sigma\) and \(\mu\) denotes the standard deviation and the mean of the 3D Charlier moment invariants, respectively. The small value of deviation \(X\) indicates that the invariants of the 3D Charlier moments are robust with respect to the translation and uniform/non-uniform scaling. The parameter of CPs \(a_{1}\) is fixed at the value \((N + M + K)/3\), where \(N \times M \times K\) is the size of 3D object, and the invariants of 3D Charlier moments are calculated up to order three in this subsection.

Fig. 6
figure 6

The 3D test object in subsection 4.2: a "Four", b "Airplane"

In the first test, we check the efficiency of the 3DCMIT descriptors constructed via Eq. (46). The original 3D object "Four" shown in Fig. 6a is translated with a set of translation vectors in the \(x\)-, \(y\)-, and \(z\)-direction. Fig. 7 shows a set of translated objects. The values of the 3DCMIT of some orders are presented in Table 1. It can be seen from this table that the deviation \(X\) of the values of the 3DCMIT is equal to zero, which means that the 3DCMIT remain unchanged whatever the translation effected to 3D object. Therefore, 3DCMIT are very robust against 3D object translation.

Fig. 7
figure 7

Set of translated 3D "Four" object

Table 1 Invariability of the proposed 3DCMIT

In the second test, we check the effectiveness of the proposed 3DCMIS descriptors constructed via Eq. (56). The original object "Four" is scaled non-uniformly in the \(x\)-, \(y\)-, and \(z\)-direction. Figure 8 shows a set of scaled "Four" object. Table 2 shows the values of 3DCMIS. This table clearly shows that the values of 3DCMIS remain unchanged under different non-uniform scaling transformations and that the deviation \(X\) is equal to zero, which proves the robustness of 3DCMIS with respect to the uniform/non-uniform scaling transformation.

Fig. 8
figure 8

Set of scaled 3D "Four" object

Table 2 Invariability of the proposed 3DCMIS

In the last test, we compare the invariability of the 3DCMITS by two methods: the indirect method based on Eq. (23), and the proposed method based on Eqs. (46) and (56). The 3D object "Airplane" presented in Fig. 6b is transformed by a set of mixed translation and scaling transformations. A set of transformed 3D "Airplane" objects is shown in Fig. 9. Table 3 shows the values of 3DCMITS obtained by the two methods. From Table 3 we can see that the deviation \(X\) of the proposed 3DCMITS are very small (\(10^{ - 10}\)) under the mixed transformations of translation and scaling, which shows that the 3DCMITS values of the transformed objects are very similar and almost identical to each other. In addition, the proposed 3DCMITS are very robust compared to the 3DCMITS derived from the indirect method which gives a deviation \(X\) of order \(10^{ - 4}\). Therefore, the proposed method offers better robustness against translation and scaling transformations than the indirect method. These satisfactory results are obtained thanks to the adopted strategy of extracting these invariants which does not require any numerical approximation, thus increasing the computational accuracy of 3DCMITS.

Fig. 9
figure 9

Set of 3D "Airplane" object transformed by a set of mixed translation and scaling transformations

Table 3 A comparison of the proposed method and the indirect one in terms of the invariability of 3DCMITS

4.3 The Computational Time of 3DCMITS

In this subsection, we evaluate the computation time of 3DCMIT and 3DCMIS. In the first test, we compare the computation time of 3D Charlier moment invariants using the proposed method and the indirect method. The Execution Time Improvement Ratio (ETIR) [15] is used as a criterion in this comparison. This ratio is defined as \({\text{ETIR (\% )}} = (1 - {{{\text{Time1}}} \mathord{\left/ {\vphantom {{{\text{Time1}}} {{\text{Time2}}}}} \right. \kern-\nulldelimiterspace} {{\text{Time2}}}}) \times 100\), where \({\text{Time1}}\) and \({\text{Time2}}\) are the execution time of the proposed method and the indirect method, respectively. Both methods are identical in terms of speed if \({\text{ETIR = 0\% }}\). The three objects of "Ant", "Human", and "Teddy-bear" (Fig. 1) are used as test objects. These objects are transformed with translations \(\Delta i = \Delta j = \Delta k\) going from − 1 to 2, and scaled in the \(x\), \(y\) and \(z\) directions with scaling factors \(a = b = c\) going from 0.5 to 2. The calculation process is performed 10 times for moment orders ranging from 1 to 12, from 1 to 24 and from 1 to 36 for each object. Tables 4 and 5 represent the average calculation time and the ETIR(%) of 3DCMIT and 3DCMIS, respectively. These tables show that the proposed method is faster than the indirect method with an improvement more than 99,67%. It should be noted that the ETIR is 99.67% for the test of translation invariance and 99,97% for the test of scaling invariance, which proves the effectiveness of the proposed method in terms of speed compared to the indirect method.

Table 4 The Average times (s) of the 3DCMIT for the 3D objects using the proposed method and the indirect one
Table 5 The Average times (s) of the 3DCMIS for the 3D objects using the proposed method and the indirect one

In the second test, we evaluate the speed of the proposed 3DCMITS for 3D medical images. For that, the 3DMRI in Fig. 4 is used as a test image. This image is translated with the translation vector (\(\Delta i = - 10\), \(\Delta j = + 6\), \(\Delta k = + 10\)) and scaled with the scaling factors (\(a = 1.1\), \(b = 0.8\), \(c = 1.4\)). The 3DCMITS based on the proposed method are compared with the 3D Tchebichef moment invariants (3DTMI) and the 3D Krawtchouk moment invariants (3DKMI) which are based on the methods presented in papers [7] and [1], respectively. The speed of these methods is also compared with that of indirect methods. The process of calculating these invariants is carried out 10 times for orders ranging from 0 to 12. The average times and the average ETIR are presented in Table 6. The results of Table 6 clearly indicate that the 3DCMITS based on the proposed method are very fast than those based on the indirect method with an improvement of 99,97%. These results also show that the computation time of 3DCMITS is less than the computation time of 3DTMI and 3DKMI based on the methods presented in [7] and [1], respectively. This may be due to the choice of Charlier polynomials as the basic function where they are defined by the hypergeometric function \({}_{2}F_{0} (.)\) which contain fewer Pochhammer terms and factorials, unlike the Tchebichef and Krawtchouk polynomials which are defined by the hypergeometric functions \({}_{3}F_{2} (.)\) and \({}_{2}F_{1} (.)\), respectively.

Table 6 Average times (s) and reduction percentage of invariant moments for transformed 3DMRI

4.4 3D Object Classification

For validation purposes and as an illustration of the interest of 3DCMITS, this subsection proposes a 3D object classification system to evaluate the performance of the proposed 3DCMITS. The proposed system is based on the proposed 3DCMITS and the Support Vector Machine (SVM) classifier. And to verify the found results, the K-Nearest Neighbor (KNN) classifier is also used in this study.

The performances of classification system based on the proposed 3DCMITS are compared with classification systems based on 3D Legendre moment invariants (3DLMI) [14], 3D Krawtchouk moment invariants (3DKMI) [1], and 3D Tchebichef moment Invariants (3DTMI) [7]. It should be noted that the 3DCMITS based on the indirect method are not considered in this subsection because of their high computation time.

The 3D object classification system consists of four main phases: the data preprocessing phase (training set and test set). Then, the construction phase of the feature vectors. Then, the data classification phase using the SVM classifier, and finally the validation phase and decision-making. The flowchart of the proposed classifier system is presented in Fig. 10 and described as follows:

Fig. 10
figure 10

3D Object recognition process

4.4.1 Phase 1: Data Preprocessing

The McGill 3D Shape Benchmark database [47] is used in this paper to validate the proposed 3DCMITS-based classification system. This database contains 19 different objects and each object has between 20 and 30 exemplars. For each object, the first three samples (Object_1, Object_2, Object_3) of size \(128 \times 128 \times 128\) are selected in order to create a base of 57 objects (\(19 \, classes \, \times \, 3\)). Three datasets were constructed for the purpose of this work: (1) Dataset No. 1 includes 1368 objects (\({57 } \times { 24 }instances\)) and is produced by the non-uniform scaling of the 57 objects with the scaling factors \(a, \, b, \, c \in \{ 0.5, \, 0.75, \, 1.25\}\), the case of \(a = b = c\) is not considered. (2) Dataset No. 2 comprises 1539 objects (\(190 \, \times { 27 }instances\)) and is produced by translating the 57 objects with the translation vectors \(TV(T_{x} , \, T_{y} , \, T_{z} )\) where \(T_{x} , \, T_{y} , \, T_{z} \in \{ - 6, \, + 2,{ + 10}\}\). (3) Dataset No. 3 comprises 3078 objects (\(57 \, \times { 54 }instances\)) and is generated with mixed translation and scale transformations of the 57 objects using translation vectors \(TV(T_{x} , \, T_{y} , \, T_{z} )\) where \(T_{x} , \, T_{y} , \, T_{z} \in \{ - 6, \, + 6\}\) and scaling factors \(a, \, b, \, c \in \{ 0.75, \, 1.5\}\), respectively. Finally, the salt-and-pepper noise is added in the three datasets mentioned above with the densities {1%, 2%, 3%, 4%}. In this paper, we choose 15% of the objects as a training set and the remaining 85% as a test set.

4.4.2 Phase 2: Construction of Feature Vector

The choice of an appropriate characteristic vector is a critical factor to have a high classification rate. For this, we only take into account the first 64 coefficients of the 3DCMITS. The lower order moments choice is justified by their ability to represent the significant characteristics of 3D objects and by their robustness against noise attacks [10]. For each object, the selected coefficients of the 3DCMITS are used to construct the following one-dimensional feature vector:

$$ \overrightarrow {{\text{V}}} = \{ V_{1} ,V_{2} ,V_{3} ,.......,V_{64} \} $$
(59)

Finally, each data object (the training set) is represented by a pair of {Class label, feature vector \(\overrightarrow {{\text{V}}}\)}.

4.4.3 Phase 3: Classifier

For high correct prediction of test data, the SVM classifier is adopted in this work because of their usefulness for data classification. Based on the training data, the SVM classifier produces a model that predicts the test data labels based on feature vectors of the test data. To avoid the numerical difficulties during computation, the feature vectors are first scaled to put their elements in the range [− 1, + 1]. Another reason for using this normalization is to avoid that feature vectors in larger numerical ranges dominate feature vectors in smaller numerical ranges. As the kernel of the model, the Radial Basis Function (RBF) is chosen in this classification test because of its ability to handle the case where the relationship between class labels and feature vector attributes is nonlinear. In addition, the RBF presents less numerical difficulties and it has only two parameters: penalty parameter \({\varvec{C}}\) and \({\varvec{\gamma}}\) parameter, which does not increase the complexity of the choice of the model unlike other kernels which have a high number of parameters. To identify the best parameters \({\varvec{C}}\) and \({\varvec{\gamma}}\), cross-validation is first used in our context. Then, the voucher (\({\varvec{C}}\),\({\varvec{\gamma}}\)) is used to form the training set.

4.4.4 Phase 4: Decision

The accuracy of 3D object recognition is measured by the following Correct Recognition Percentage (CRP):

$$ {\text{CRP}} = \frac{{\text{Number of correctly classified 3D objects}}}{{\text{The total of 3D objects used in the test }}} \times 100 $$
(60)

The classification performance of the proposed 3DCMITS is verified and compared with that of 3DLMI, 3DKMI, and 3DTMI presented in papers [1, 14], and [7], respectively. The classification accuracies based on SVM classifier of sets No. 1, No. 2 and No. 3 are presented in Tables 7, 8 and 9, respectively.

Table 7 Classification results of the 3D objects with translation transformations using SVM classifier
Table 8 Classification results of the 3D objects with scaling transformations using SVM classifier
Table 9 Classification results of the 3D objects with translation and scaling transformations using SVM classifier

Examining the results of these tables, Table 7 shows that for the case of 3D objects transformed by translation (Dataset No. 1), 3DCMITS, 3DTMI, and 3DKMI lead to high classification rates, with a relative preference for the 3DTMI which present the rate of 87,59%, while the proposed 3DCMITS and 3DKMI present the rates of 87,40% and 87,29%, respectively. In contrast, 3DLMI have the lowest rate with a rate of 79.96%. In Table 8, it is clearly shown that for the case of scaled 3D objects (Dataset No. 2), the proposed 3DCMITS result in the highest classification rates. The 3DCMITS have the highest rate of 86,64%, followed by 3DLMI with 79,12%, while 3DKMI and 3DTMI have the lowest rates of 75,28% and 69,92%, respectively. This is mainly due to the method of extracting the scale invariants of 3D moments of Krawtchouk and Tchebichef, respectively. The advantages of the proposed descriptors are clearly highlighted when dataset No. 3 is used: Table 9 shows that for the case of mixed translation and scaling transformations, 84,19% of 3D objects are classified correctly using the proposed 3DCMITS, while 77,08%, 73,19%, and 67,45% for the 3DLMI, 3DKMI, and 3DTMI, respectively. It should be noted that recognition accuracy decreases with increasing noise. However, the proposed 3DCMITS have recognition rates in excess of 84% even at high noise intensities. These results show that the proposed Charlier invariant moments are robust to translation and scaling transformations, whose recognition rate is higher than those of the 3D invariant moments of Legendre [14], Krawtchouk [1], and Tchebichef [7].

In order to confirm the above discussions and to validate the results found, the classification process is redone using the K-Nearest Neighbor (KNN) classifier. The Euclidean distance [20] is used for the KNN classifier and the value of the K parameter (number of nearest neighbors) is chosen by varying K until the minimum error is reached. The classification results of datasets No. 1, No. 2 and No. 3, based on the KNN classifier, are presented in Tables 10, 11, and 12, respectively. The results of these tables indicate that the proposed 3DCMITS lead to the highest classification rates and are superior to 3DLMI, 3DKMI, and 3DTMI, which result in the lowest classification rates. These results confirm the previous discussions in terms of the studied descriptors' effectiveness for the 3D object classification. It should also be noted that the results found based on the SVM classifier are higher than those based on the KNN classifier: with the proposed 3DCMITS, 87,40%, 86,64%, and 84,19% of objects are well classified using SVM classifier for dataset No. 1, dataset No. 2, and dataset No. 3, respectively. While 83,81%, 77,47%, and 75,73% are well classified using KNN classifier. This is due to the superiority of the SVM which offers satisfactory results in the field of 3D object classification/recognition.

Table 10 Classification results of the 3D objects with translation transformations using KNN classifier
Table 11 Classification results of the 3D objects with scaling transformations using KNN classifier
Table 12 Classification results of the 3D objects with translation and scaling transformations using KNN classifier

5 Conclusion

In this paper, we have proposed a fast method for calculating the translation and scaling invariants of 3D Charlier moment. This method is based on the extraction of these invariants directly via the Charlier polynomials without any numerical approximation. The calculation of these descriptors using this method allows, on the one hand, to considerably reduce the calculation time with an improvement of 99,97%, and on the other hand, to increase the robustness against translation and scale transformations. The interest of the proposed 3DCMITS is shown in the 3D object recognition application. The noise robustness and recognition accuracy of the proposed 3DCMITS are studied. Experimental results show the efficiency and superiority of the proposed 3DCMITS. These descriptors have a desirable capability for 3D image/object description and can be useful in the image analysis field.