1 Introduction

Accuracy is one of the most important key performance indicators (KPIs) of the five-axis motion platform. A command is given to this mechanism with three translation axes and two rotation axes, to make it reach the designated pose. The ideal situation is where the actual pose of the five-axis motion platform identifies with the commanded pose. However, due to various reasons, the five-axis motion platform cannot reach the designated pose, which leads to the spatial geometric error. According to the motion characteristics of geometric errors, they are usually divided into position-dependent geometric errors (PDGEs) and position-independent geometric errors (PIGEs). PDGEs are mainly caused by manufacturing defects of mechanical parts, such as the straightness error of the guide rail [1], while PIGEs are mainly caused during the assembly of mechanical parts [2]. The error, caused by mechanical parameters deviation, accounts for 65–95% of the total system error, in the kinematics equation. The geometric structure parameters are used to represent the geometric relationship of motion axes. For the five-axis motion platform, in ideal state, the three linear axes are perpendicular to each other, and the two rotary axes are vertical and intersecting. However, the platform will be affected by processing errors, installation errors, and other factors. For the PIGEs caused by installation, there are squareness errors in linear axis and angular errors and position errors in rotary axis [3]. Due to the cumulative effect of these errors, the ideal geometric structure relationships are difficult to achieve. There is a certain deviation between the actual geometry parameters and the design parameters of the five-axis motion platform. This deviation leads to inaccurate control modeling of multi-axis motion platform, which reduces the derived absolute accuracy of the construction [4,5,6].

This paper mainly studies the measurement, identification, analysis, and compensation of PIGEs of a five-axis motion platform. The different measurement methods of five-axis motion platform geometric errors can be divided into direct and indirect methods [7, 8]. According to the direct measurement method, a device is used to directly measure the error term of the platform. For example, a laser interferometer is usually used to measure the linear positioning error of a translational axis. The measurement accuracy of laser interferometer can reach submicron accuracy, but its measurement process is complex and time-consuming. Based on the indirect measurement method, the mathematical model of the relation between the pose error of the end effector and geometric error elements is first established, while the geometric error parameters are identified by measuring the pose error of the end effector. Due to its high efficiency, indirect measurement has become a research hotspot, in recent years. Various measuring instruments have been designed and used to calibrate geometric errors efficiently, based on the indirect approach. Abbaszadeh-Mir et al. [9] used the ball-bar sensor to measure the position and orientation deviation of the tool center point of the five-axis motion platform, to achieve the identification of geometric errors. Fan [10] established the identifying model based on homogeneous transform matrix, while double ball bar (DBB) is adopted to test error motions of rotary axes. Jiang [11] established a five-axis machine tool model with homogeneous transformation matrices and proposed a testing scheme to diagnose the PIGEs of linear and rotary axes by using double ball bar. Zhang et al. [12] designed two DBB-measuring paths in different horizontal planes to evaluate simultaneously five errors including one axial error, two radial errors, and two tilt errors caused by the rotary table. Ibaraki et al. [13] proposed an efficient scheme to calibrate error motions of rotary axes on a five-axis machining center by using the R-test, while they present an algorithm to identify location errors and such position-dependent geometric errors. Hong et al. [14] proposed a scheme to separate squareness errors of linear axes by performing a set of R-test measurement to observe error motions of rotary axes. A touch trigger probe [15, 16] is also used to calibrate geometric errors. Mchichi [17] designed a measurement strategy for the indirect calibration of a five-axis machine tool using the Scale and Master Balls Artefact method.

In order to reduce the measurement cost and operation complexity of measuring instruments, many scholars use vision to derive a high-precision pose of five-axis motion platform and a robot to calibrate the geometric errors. Ibaraki [18] used a monocular camera to measure the two-dimensional position error of a five-axis motion platform, whereas Liu et al. [19] used a binocular camera to calibrate the parameters of PIGEs of the five-axis motion platform. Li et al. [20] proposed a 3D high temporal-spatial measurement method based on a monocular camera and realized the 3D detection of contouring errors of arbitrary paths. However, Li et al. only measured the contouring error and did not calibrate the geometric error of five-axis machine tool. In addition to the traditional vision measurement, Wang et al. [21] proposed a method of robot position error estimation, based on vision and neural network. Stepanova et al. [22] used RGB-D stereo camera to measure the pose of a robot end effector. Wang et al. [23] calibrated the parameters of body structure, hand eye structure, and camera system, using binocular vision, for a 6-DOF industrial robot.

In view of the high cost of current high-precision measurement equipment and the complexity of the calibration operation of five-axis motion platform, a low-cost pose measurement scheme, based on monocular vision, is proposed. The path planning is carried out, so as to automatically collect the platform end effector pose, while the pose error is calculated to realize the calibration of PIGEs of the five-axis motion platform. Before and after calibration, the five-axis motion platform moves on a circular trajectory of the target coordinate system (TCS) origin, relative to the camera coordinate system (CCS), through a five-axis linkage. The comprehensive accuracy variations of the five-axis motion platform, before and after calibration, are tested by comparing the position and orientation errors of the theoretical circle track to the respective ones of the actual circle track. The comparison of experimental results verifies the applicability and effectiveness of the calibration scheme.

The rest of this paper is organized as follows: Camera calibration model is expressed in Section 2. Section 3 portrays the method of target pose measurement and detection of pose measurement accuracy. Section 4 describes the calibration method of kinematic geometric parameters of five-axis motion platform using monocular vision. Experimental results and verifications of Calibration method for kinematic geometric parameters of five-axis motion platform are described in Section 5. The paper is concluded in Section 6.

2 Camera calibration model

The unit of a monocular camera can be roughly divided into three parts: lens, photosensitive chip, and processing circuit. When the light passes through the lens, the object to be photographed will produce an “image” on the photosensitive chip, which will be processed by the processing circuit, to derive a picture. As shown in Fig. 1, the monocular camera, used in this paper, is installed on the z-axis of the five-axis motion platform, while a high-precision checkerboard target (referred to as the target) is pasted on the c-axis. The spatial position of the camera is adjusted based on the motion along the X, Y and Z axes of the five-axis motion platform, while the spatial position of the target is determined by the motion along the A and C axes.

Fig. 1
figure 1

Schematic structural diagram of five-axis motion platform

The imaging schematic diagram of the target, according to the ideal pinhole imaging model, is shown in Fig. 2, where five coordinate systems are defined. The world coordinate system Ow-XwYwZw is used as the datum coordinate system. In the camera CCS Oc-XcYcZc, the Zc axis coincides with the optical axis of the camera. Ob-XbYbZb is a three-dimensional rectangular coordinate system, referred to as the TCS, where the target is located on the XbYb plane. The image pixel coordinate system Op-uv represents the coordinates of the object point on the imaging plane. The image physical coordinate system o-xy, with pixels as the unit, represents the position of o point in the imaging plane coordinate system.

Fig. 2
figure 2

Schematic diagram of checkerboard target imaging

As shown in Fig. 2, the coordinates of any corner M, on the calibration board, are Xb, Yb, and Zb in the TCS and Xc, Yc, and Zc in the CCS. According to the pinhole imaging principle, the intersection point, between the line segment MOc and the imaging plane, is the image point m of the corner M. The coordinates of image point m, in the image pixel coordinate system, are u and v. The essence of calibration is to express the corner point coordinates in the form of pixel coordinates. Without considering lens distortion, the relationship [24, 25] between the coordinates (Xb, Yb, Zb) of corner point M, in the TCS, and the coordinates (u, v) of the corresponding image point m, in the image pixel coordinate system, is as follows:

$$ {\displaystyle \begin{array}{l}{Z}_e\left[\begin{array}{c}u\\ {}v\\ {}1\end{array}\right]=\left[\begin{array}{cccc}{f}_x& 0& {u}_0& 0\\ {}0& {f}_y& {v}_0& 0\\ {}0& 0& 1& 0\end{array}\right]\left[\begin{array}{cc}R& T\\ {}{0}^T& 1\end{array}\right]\left[\begin{array}{c}{X}_b\\ {}{Y}_b\\ {}{Z}_b\\ {}1\end{array}\right]=\\ {}\left[\begin{array}{cccc}{p}_1& {p}_2& {p}_3& {p}_4\\ {}{p}_5& {p}_6& {p}_7& {p}_8\\ {}{p}_9& {p}_{10}& {p}_{11}& {p}_{12}\end{array}\right]\left[\begin{array}{c}{X}_b\\ {}{Y}_b\\ {}{Z}_b\\ {}1\end{array}\right]={M}_1{M}_2\left[\begin{array}{c}{X}_b\\ {}{Y}_b\\ {}{Z}_b\\ {}1\end{array}\right]\end{array}} $$
(1)

where Zb=0, u0 and v0 represent the pixel coordinates of the center o point on the imaging plane, and fx and fy denote the normalized focal length of the camera lens along the x-axis and y-axis. M1 is the internal parameter matrix of the camera, which will remain constant, after the camera is installed and the focusing is concluded. M2 is the external parameter matrix of the camera, which changes according to the pose of the TCS. R is the rotation matrix and T is the translation matrix of conversion from the TCS to the CCS.

Due to the manufacturing accuracy and installation technology of camera lens, the original image will be distorted, which will make the image point, as defined by the three-dimensional point projection, deviate from the ideal position. Therefore, the radial distortion coefficients K1, K2, and K3 and tangential distortion coefficients P1 and P2 of the camera are introduced, to correct the image point position.

The internal parameters of the camera include camera internal parameter matrix and distortion coefficient vector. The camera internal parameters are calibrated based on the Zhang Zhengyou plane calibration method [26], so as to calculate more accurate target pose, according to the Perspective-N-Points (PNP) principle.

3 Target pose measurement and accuracy detection

PNP problem is a very important issue in computer vision. Using PNP principle to solve the target pose requires to accurately detect each corner on the target and to obtain the sub-pixel coordinates (ui, vi) of each corner point, in the image pixel coordinate system and the coordinates (Xbi, Ybi, Zbi) of each corner point in the TCS.

3.1 Corner detection

For the planar checkerboard patterns used in this paper, the common corner imaging image is shown in Fig. 3. In order to detect the corners, 8 convolution kernels A1, B1, C1, D1, A2, B2, C2, and D2 are constructed to convolute with the collected target image. As shown in Fig. 4, in order to construct convolution kernels A1, B1, C1, and D1, the convolution kernel coordinate system x1o1y1 is established. In Fig. 3, the position of corner a1 is defined as the origin of coordinate system x1o1y1, and the vertical boundary line and horizontal boundary lines of the a1 point are X-axis and Y-axis, respectively. Defining r is the radius of convolution kernel, while the position coordinates of each element of the convolution kernel are xi and yi in Fig. 4.

Fig. 3
figure 3

The common corner imaging image: a corner imaging image a1, b corner imaging image a2, c corner imaging image b1, and d corner imaging image b2

Fig. 4
figure 4

The coordinates diagram of convolution kernel

The Gauss fuzzy convolution kernel equation is as follows:

$$ G\left({x}_i,{y}_i\right)=\frac{1}{2\pi {r}^2}\cdot {e}^{-\frac{\left({x_i}^2+{y_i}^2\right)}{2{r}^2}} $$
(2)

According to Eq. (2), the weight value G(xi,yi) of each convolution kernels element is obtained, while the elements in each quadrant of the convolution kernel coordinate system x1o1y1 are normalized. Then the elements in quadrant I are proposed, while the rest of the elements in quadrant are assigned to 0, to obtain convolution kernel A1. The elements in quadrant III are proposed, while the other elements in quadrant are assigned to 0, to obtain convolution kernel B1. Also, the elements in quadrant II are proposed, while the rest elements in quadrant are assigned to 0, to obtain convolution kernel C1. Finally, the elements in quadrant IV are proposed, while the other elements in quadrant are assigned to 0, to obtain convolution kernel D1.

For the corner a1 in Fig.3, the response of convolution kernels C1 and D1 is higher than that of convolution kernels A1 and B1, while the mean value of the responses of convolution kernels A1, B1, C1, and D1 is larger. Similarly, convolution kernels A2, B2, C2, and D2 can be constructed to detect corners b1 and b2 in Fig. 3.

The experimental results showed that the 8 convolution kernels can detect all corner imaging image of the chessboard calibration board, not only the four corner imaging images shown in Fig. 3. The total of 8 convolution kernels are used to convolute the pixels in the target image. The corner likelihood of each pixel in the image is calculated, forming the set C. A non-maximum suppression algorithm is used to filter and suppress the values in set C, to obtain the maximum points in the set C. These maximum points constitute the point set Q of the candidate points of the corner, thus obtaining the initial coordinates of the corner points [27].

The accuracy of the initial coordinates of the corners is not enough; thus, further optimization is needed. In this paper, the gray value gradient method [28] is used to determine the sub-pixel coordinates of corners. The basic principle is shown in Fig. 5. If the true position of the corner is point C, the points near point C can be divided into two categories: (a) the point on the edge, such as P in Fig. 5, and (b) the non-edge point, such as Q in Fig. 5. For an ideal corner, the gradient vector of point P is perpendicular to the edge, while the gradient value at point Q is zero. Therefore, all the points near the corner have the following characteristics:

$$ \overrightarrow{\nabla {C}_i}\bullet \overrightarrow{CC_i}=0 $$
(3)
Fig. 5
figure 5

Gradient diagram of corners

where Ci is any point in a region centered on C and \( \overrightarrow{\nabla {C}_i} \) is the gradient vector at point Ci. This problem can be simplified into the following optimization problems:

$$ c=\arg \underset{c_i}{\min}\sum \limits_{c_i\in {N}_I(c)}{\left({g}_{c_i}^T\left({c}_i-c\right)\right)}^2 $$
(4)

The expression of the solution to the optimization problem is as follows:

$$ c={\left(\sum \limits_{c_i\in N}{g}_{c_i}{g}_{c_i}^T\right)}^{-1}\bullet \sum \limits_{c_i\in N}\left({g}_{c_i}{g}_{c_i}^T\right){c}_i $$
(5)

In Eqs. (4) and (5), the small letters c and ci represent the coordinates of point C and point Ci, respectively, gci represents the gradient vector at point Ci, while N represents the area near point c.

In order to use the detected corner pixel coordinates to solve the target pose, after detecting the sub-pixel coordinates of all corners, it is necessary to automatically sort the detected corner pixel coordinates. To this end, the energy method is used as a sorting algorithm, so as to derive the sub-pixel coordinates (ui ,vi) of each corner in a specific order.

3.2 Solution of target pose

The inner corner of the target, as used in this paper, is in a 14 × 14 format, while the distance between adjacent corner points is 2mm. Each corner point is accordingly arranged on the target plane. The coordinates of each corner point are considered in the TCS, shown in Fig. 2. The sub-pixel coordinates (ui, vi) of each corner are obtained, according to the method, presented in Section 2.1, while the PNP principle is used to solve the pose of the target. Considering Eq. (1), the Zc in the last line is eliminated, while the two constraints are obtained, as follows:

$$ \left\{\begin{array}{c}u=\frac{p_1{X}_b+{p}_2{Y}_b+{p}_3{Z}_b+{p}_4}{p_9{X}_b+{p}_{10}{Y}_b+{p}_{11}{Z}_b+{p}_{12}}\\ {}v=\frac{p_5{X}_b+{p}_6{Y}_b+{p}_7{Z}_b+{p}_8}{p_9{X}_b+{p}_{10}{Y}_b+{p}_{11}{Z}_b+{p}_{12}}\end{array}\right. $$
(6)

In case there are N corners, let Mi=(Xbi,Ybi,Zbi,1), the obtained coordinates (ui ,vi) and Mi are uniformly substituted into Eq. (6), and linear relations are obtained, as follows:

$$ \left[\begin{array}{ccc}{M}_1& 0& -{u}_1{M}_1\\ {}0& {M}_1& -{v}_1{M}_1\\ {}\vdots & \vdots & \vdots \\ {}{M}_N& 0& -{u}_N{M}_N\\ {}0& {M}_N& -{v}_N{M}_N\end{array}\right]\left[\begin{array}{c}{p}_1\\ {}\vdots \\ {}{p}_{12}\end{array}\right]=0 $$
(7)

The parameters (p1, p2p12) are obtained using the direct linear transformation method, while the external parameter matrices R and T of the camera are obtained by decomposition. Specifically, 196 sets of coordinates (ui, vi) are used to iterate obtain optimal external parameter matrices R and T in each target image. The objective function of obtaining external parameter matrices R and T is to minimize the reprojection error in the iteration process. Therefore, even if image noise, reflection, and uneven illumination cannot accurately detect the pixel coordinates of a single corner, the accuracy of the overall posture can also be guaranteed.

The pose relationship between the CCS and the TCS is obtained as follows:

$$ \left[\begin{array}{c}{X}_{\mathrm{c}}\\ {}{Y}_c\\ {}{Z}_c\\ {}1\end{array}\right]=\left[\begin{array}{cc}R& T\\ {}{0}^T& 1\end{array}\right]\left[\begin{array}{c}{X}_b\\ {}{Y}_b\\ {}{Z}_b\\ {}1\end{array}\right] $$
(8)

3.3 Verification of target pose measurement accuracy

In the case of the five-axis motion platform, used in this paper, the respective X, Y, and Z axes are moved in turn, while the corresponding motor encoder values and target image are collected at each axis movement. The actual traveled distance of motor axis, as calculated by the encoder measurements, is considered the actual distance da of camera optical center moving in space. The pose matrices R and T of the target, solved by processing the target image, provide the spatial distance variation value, used as the measurement distance dm of camera optical center moving in space. The deviation de between da and dm is calculated, to evaluate the accuracy of target pose measurement by monocular camera.

The experiment considers three groups. In each group of experiments, only one of translational axes of five-axis motion platform was moved, 1 mm each time, and the images were collected at the time of stopping, while it moved continuously for 10 times in one direction. The distance error value de, generated at the 10 points in each group motion, is calculated. The experimental results are shown in Fig. 6, where it is illustrated that the distance error caused by moving X and Y axes by single axis can remain below 5 μm. The distance between the camera and the target, in the z-axis direction, will affect the clarity of the recorded image, in the monocular camera measurement system, which will have a greater impact on the results of the target pose measurement. In Fig. 6, the initial moving position of z-axis shows the best focusing results in height. The focus height will gradually change, after continuously moving z-axis, thus affecting the clarity of the image, resulting in an upward trend of the distance error.

Fig. 6
figure 6

Verification experiment of monocular vision measurement accuracy

In practice, the camera will be adjusted to the best focus position, in the z-axis direction, so as to collect image data, while the measurement distance error of the camera is limited within a small error range, during the Z-axis motion. Comprehensive analysis shows that the target pose measurement error in this scheme can remain below 10 μm, when the distance between the camera and the target, in the z-axis direction, is well controlled to continuously provide clear recorded image with high measurement accuracy capability. The vision measurement system can meet the pose accuracy requirements, during the calibration process of the five-axis motion platform.

3.4 Comparison of pose estimation with different noise levels

Gaussian noise is added to the target image, and the mean value of Gaussian noise is 0 and the variance is σ. Then the corner detection algorithm used in this paper is compared with that in OpenCV. As shown in Fig. 7, when the value of σ is greater than 0.0003, the corner detection algorithm will detect the corner position by mistake in the target image. However, when the value of σ is greater than 0.0015, the corner position will be detected by mistake in the target image using the detection method in this paper. Let σ change from 0 to 0.0015 for experiment using the pose estimation method in this paper, and get the curve of the reprojection error of target image and the change of σ, as shown in Fig. 8. The experimental results show that the proposed method can accurately measure target pose, even in conditions of image shadow and noise, demonstrating certain robustness.

Fig. 7
figure 7

Display of corner detected by mistake

Fig. 8
figure 8

The curve of the reprojection error of target image and the change of σ

4 Calibration method for kinematic geometric parameters of five-axis motion platform

Before calibration, the relationship between the motor value Qd =[θx ,θy ,θz ,θa ,θc] of each axis of the five-axis motion platform and the theoretical pose Sd of the TCS, relative to the CCS, can be expressed by the forward kinematics solution F() and the inverse solution I() of the five-axis motion platform. The θx, θy, θz, θa, and θc in Qd represent the position data provided by the motor encoder of X, Y, Z, A, and C axis, respectively:

$$ {S}_d=F\left({Q}_d,{\Phi}_d\right) $$
(9)
$$ {Q}_{\mathrm{d}}=I\left({S}_d,{\Phi}_d\right) $$
(10)

where Φd is the nominal value of geometric structure parameters of five-axis motion platform. Due to the PIGEs of the five-axis motion platform, when this is required to move to the specified relative pose Sd, the actual relative pose Sa is as follows:

$$ {S}_a=F\left({Q}_d,\Phi \right) $$
(11)

where Φ=Φd+ΔΦ is the actual value of geometric structure parameters and ΔΦ is the error value of geometric structure parameters. Only when the accurate ΔΦ is identified, as far as possible, the accurate forward and inverse solution can be obtained and the final calibration can be realized. The schematic diagram of calibration process is shown in Fig. 9.

Fig. 9
figure 9

Schematic diagram of calibration process of five-axis motion platform

An important element of the calibration process is to automatically collect the required data. N images of target were recorded clearly in multiple positions, during the five-axis cooperative motion. According to Eq. (9), the coordinates expression of the TCS origin, relative to the CCS, is obtained, as follows:

$$ \left\{\begin{array}{l}x={f}_1\left({\theta}_x,{\theta}_y,{\theta}_z,{\theta}_a,{\theta}_c\right)\\ {}\mathrm{y}={f}_2\left({\theta}_x,{\theta}_y,{\theta}_z,{\theta}_a,{\theta}_c\right)\\ {}z={f}_3\left({\theta}_x,{\theta}_y,{\theta}_z,{\theta}_a,{\theta}_c\right)\end{array}\right. $$
(12)

The relationship between x, y, and z and θx, θy, and θz can be expressed linearly, after assigning values to θa and θc, as:

$$ \left[\begin{array}{c}x\\ {}y\\ {}z\end{array}\right]=M\left[\begin{array}{c}{\theta}_x\\ {}{\theta}_y\\ {}{\theta}_z\end{array}\right]+\left[\begin{array}{c}{u}_1\\ {}{u}_2\\ {}{u}_3\end{array}\right] $$
(13)

The coefficient matrix M represents the conversion factor, while u1, u2, and u3 represent the constant influence factors on x, y, and z, after the assignment of θa and θc. The relationship between θx, θy, and θz and x, y, and z can be obtained by matrix transformation, as follows:

$$ \left[\begin{array}{c}{\theta}_x\\ {}{\theta}_y\\ {}{\theta}_z\end{array}\right]={M}^{-1}\left[\begin{array}{c}x-{u}_1\\ {}y-{u}_2\\ {}z-{u}_3\end{array}\right] $$
(14)

Let the coordinates of the TCS origin, relative to the CCS, be x0, y0, and z0, when the position of the collected image is the optimal. Based on this condition, n theoretical poses Sdi are planned. According to Eq. (14), keeping the coordinates of x, y, and z as x0, y0, and z0, the values of θx, θy, and θz, under different values of θa and θc, are derived, so as to obtain the motor data Qdi of N groups of collected images. In addition, the path planning of the motion is conducted so as to avoid collision between the Z-axis and the C-axis of the five-axis motion platform. The motion command is sent by the controller to drive the motor to move to the motor position Qdi. Next, the image is collected by the monocular camera, while the target pose is obtained according to the method, presented in Section 2; that is, n actual poses Sai are obtained. Following, the pose error of the five-axis motion platform is determined as follows:

$$ \Delta {S}_i={S}_{ai}-{S}_{di} $$
(15)

When the error ΔΦ of geometric parameters is small enough, Eq. (15) can be linearized as follows:

$$ \Delta {S}_i={J}_i\Delta \Phi $$
(16)
$$ {J}_i=\frac{\partial F}{\mathrm{\partial \Phi }} $$
(17)

where Ji is the Jacobian matrix calculated when the platform is in the i-th pose. Combining Eqs. (16) and (17) results in the following expression:

$$ E=J\Delta \Phi $$
(18)

where J = [J1JiJn]T and E = [ΔS1⋯ΔSi⋯ΔSn]T. The optimum value of ΔΦ, as obtained by the least square method, is as follows:

$$ \Delta \Phi ={\left[{J}^TJ\right]}^{-1}{J}^TE $$
(19)

Equation (19) provides the first estimation of the error value ΔΦ of the five-axis motion platform. The value of ΔΦ is added to the nominal geometric structure parameters Φd, to obtain a new set of geometric structure parameters Φj.

This set of Φj values, instead of the nominal value Φd, is calculated again, within an iterative process, where the geometric structure parameters of the five-axis motion platform can be described as:

$$ \left\{\begin{array}{l}{\Phi}_0={\Phi}_d\\ {}{\Phi}_{j+1}={\Phi}_j+\Delta \Phi \end{array}\right. $$
(20)

The iteration process continues until ∆S is as small as the termination condition requires, resulting in an accurate geometric structure parameters Φ of the five-axis motion platform.

5 Test experiment of calibration accuracy

After the accurate geometric structure parameters Φ of the five-axis motion platform is identified, the motor axis is moved so that the target image in the camera is clearly at the center of the camera field of view, while the motor value Qd of the five-axis motion platform is at this time recorded. Substituting the value of Qd into Eq. (11), the pose matrix of the TCS, relative to the CCS, is obtained. This pose is considered the starting point to rotate around the Z axis of the CCS, in planning a five-axis linkage theoretical pose trajectory T1 of the TCS, relative to the CCS, where the TCS origin is a space circle in the CCS. Through the inverse solution of the kinematics model, the five-axis motor value Qi(0<i<=200) of the five-axis motion platform, corresponding to the theoretical circle trajectory T1, is obtained. The trajectory T1 is shown in Fig. 10.

Fig. 10
figure 10

The trajectory T1

The main computer sends the motor value Qi to the motion control card to move the five-axis motion platform along the set trajectory T1, executing five-axis cooperative motion. When the five-axis motor value moves to position of Qi, the camera is used to collect the target image and a total number of 200 images is collected. These images are processed based on the method in Section 2 and the actual pose matrix (Rai, Tai) of the five-axis motion platform is calculated. According to Eq. (11), the theoretical pose matrix (Rdi, Tdi), corresponding to the coordinates set Qi, is calculated. In order to quantify the calibration results, the formulas for solving the position error ∆Ti and orientation error ∆Ri between the actual pose and the theoretical pose are defined as:

$$ \Delta {T}_i=\left\Vert {T}_{ai}-{T}_{di}\right\Vert $$
(21)
$$ \Delta {R}_i=\left\Vert \log {\left({R}_{ai}^{-1}{R}_{di}\right)}^{\vee}\right\Vert $$
(22)

where v represents the transformation from Lie group to Lie algebra.

In the nominal geometric structure parameters Φd of five-axis motion platform, the position error ∆Ti and the orientation error ∆Ri, between the actual and theoretical positions, are calculated by the same method. The accuracy variations by the five-axis motion platform calibration are evaluated according to the differences in position error ∆Ti and orientation error ∆Ri, before and after model calibration. The experimental setup is shown in Fig. 11. The program running on the upper computer is used to collect data, to analyze the calibration results.

Fig. 11
figure 11

Experimental equipment

Figure 12 illustrates the position error of the five-axis motion platform, before and after calibration. Before calibration, the average value of the position error of the five-axis motion platform is 108.70 μm, while the maximum value is 125.30 μm. After calibration, the average value of the position error of the five-axis motion platform is reduced to 22.33 μm and the maximum value is reduced to 36.01 μm. Figure 13 shows the direction error, before and after calibration of the five-axis motion platform. Before calibration, the average value of the orientation error of the five-axis motion platform is 0.265° and the maximum value is 0.365°. After the calibration, the average value of the orientation error of the five-axis motion platform is reduced to 0.0357°, and the maximum value is reduced to 0.0789°.

Fig. 12
figure 12

Comparison of position error before and after calibration of five-axis motion platform

Fig. 13
figure 13

Comparison of orientation error before and after calibration of five-axis motion platform

By comparing the experimental data of position and direction errors, before and after calibration, one concludes that the position error and orientation error of the five-axis motion platform are considerably reduced, as a result of the calibration method presented in this paper. The comparison shows that more accurate geometric structure parameters can be obtained after the calibration of the five-axis motion platform, which results in the actual motion trajectory better approximating the theoretical motion trajectory. Furthermore, calibration significantly improves the comprehensive accuracy of the five-axis motion platform

6 Conclusions

In order to solve the problems of high measurement cost and complex operation of calibration of five-axis motion platform, this paper presents a designs of simple low-cost operation method, based on monocular vision pose measurement.

  1. (1)

    The pose measurement method, based on monocular vision, shows high accuracy, as the error of target position measurement remains below 10 μm, even in conditions with image shadow and noise.

  2. (2)

    The visual pose measurement method is combined with the kinematic loop method, to implement the kinematic model calibration of the five-axis motion platform. The geometric structure parameters of the error model are obtained by the method of equation iteration, while PIGEs are calibrated. After calibration, the average position error of the five-axis motion platform is reduced by 22.33 μm, from 108.70 μm and decreased by 79.46%, while the average value of orientation error is reduced from 0.265 to 0.0357° and decreased by 86.53%. The accuracy of the five-axis motion platform shows significant improvement, which verifies the applicability and effectiveness of the calibration scheme.