1 Introduction

A vehicle around-view monitoring (AVM) system will be a great help to enhance the safety of driving or parking by eliminating the blind spots around a vehicle. In order to eliminate the blind spots, four or six cameras are deployed on four sides of a vehicle. Figure 1 shows an example AVM image as presented in [1]. There has been increasing interest in AVM as an essential element of a smart vehicle in recent years.

In general, an AVM system consists of (1) acquisition of images surrounding a vehicle, (2) viewpoint transformation of the acquired images, and (3) stitching the images into a display image. The viewpoint transformation converts an original image from a camera into a bird’s-eye view. The image stitching process removes the overlapping area of each image and combines the images to be shown on the driver’s display. In order to secure a wide field-of-view (FOV), a fisheye lens is generally used for image acquisition.

Many results are available about study of the AVM. Yebes et al. developed image processing hardware for an AVM system with four cameras and applied it to a mobile robot [2]. The original image acquired through a fisheye lens requires rectification to compensate the radial distortion. Lo et al. proposed a rectification method to convert the fisheye image into an ideal pinhole image for their AVM system [3]. There are overlapping areas between the images from the cameras on the four sides of a vehicle. Some obstacle information in the overlapping area of each image may be lost in the stitching process of the images. A dynamic switching method was proposed to select one image from the AVM cameras to reduce the chance of missing obstacles in the overlapping area [4]. B. Zhang et al. proposed a photometric alignment method to blend the overlapping areas of the AVM camera images [1]. Sato et al. used a spatio-temporal bird’s-eye view image to construct a robust surrounding view of a vehicle or a mobile robot by compensating a broken camera or network disturbance [5]. They also showed that the surrounding bird’s eye view is more useful for achieving accurate remote control for a mobile robot than a conventional front view. Similarly, Lin et al. proposed a top-view image transformation involving the transformation of a perspective image into its corresponding bird’s-eye view [6]. They applied the top-view transformation to a vehicle parking assistant system with a camera at the rear end of a vehicle.

Fig. 1
figure 1

Vehicle AVM image sample [1]: a Original images from four cameras; b surrounding bird’s-eye view

On the other hand, the catadioptric approach is an image acquisition method that uses a reflector with a conventional camera. The catadioptric imaging method has been used for single-camera stereo image acquisition by optically dividing the image sensor plane [7, 8], or for wide FOV omnidirectional image acquisition by using a bowl-shaped convex/concave reflector [9,10,11,12]. Several types of these reflectors have been developed for omnidirectional imaging, including hyperbolic, parabolic, elliptic, and conic reflectors [13]. Wide FOV image acquisition is the most important part of the AVM for reducing the number of cameras and eliminating blind spots.

Based on the omnidirectional image of vehicle surroundings obtained by the catadioptric approach, it is possible to generate a bird’s-eye view of the ground [14] or a virtual perspective along the driver’s viewing direction [15]. Cao et al. proposed a special rectifying catadioptric omnidirectional reflector that preserves the actual distance on the ground surface in the image surface obtained by the reflector [16]. The rectifying reflector has the advantage of wide perspective image acquisition along the diameter without distortion on any plane perpendicular to the optical axis; it makes the image unwarping easy with reduced computation. A similar omnidirectional reflector was proposed based on a rectilinear projection scheme by Kweon et al. [17, 18]. The rectifying reflector was used along with a conventional hyperbolic omnidirectional reflector to obtain the surrounding view of a vehicle for a driver assistance system and autonomous driving [19]. As the imaging system uses the rectifying reflector and the hyperbolic reflector to provide two views in the overlapping area, it can be utilized to create a sparse 3D reconstruction. In [20], Yi et al. proposed an omnidirectional stereo vision method using a single camera with an additional concave lens for the 3D reconstruction. As ultrawide omnidirectional imaging sensors, Cheng et al. [21] and Sturzl et al. [22] presented similar designs consisting of an annularly combined catadioptric mirror and lens assembly, respectively. Their approaches overcame the central blind area occluded by the camera in the conventional catadioptric imaging system.

The main aim of this study is to provide a hybrid hyperbolic reflector for a vehicle catadioptric AVM. The hybrid reflector consists of the upper cylindrical section and the lower half-omnidirectional section with common focal point and surface continuity. The AVM with the hybrid reflector has the following potential advantages compared with the existing AVM systems that use a traditional fisheye lens or an omnidirectional reflector: (1) the horizontal side scene from the hybrid reflector has better image quality than that from the omnidirectional reflector or from the fisheye lens because the upper cylindrical section of the hybrid reflector has higher sensor utilization and (2) it is possible to observe the horizontal side scene as well as the vertical ground scene without any seamline between them. The conventional AVM systems present only the vertical ground scene in the driver’s display, as shown in Fig. 1 in general. Not only the vertical bird’s-eye view of the ground scene but also the horizontal panoramic view from the hybrid reflector are helpful for a driver to be aware in a driving or parking situation. Using geometrical optics, this study presents an image acquisition model of the hyperbolic reflector imaging system and the image reconstruction algorithm for the horizontal panoramic view and vertical bird’s-eye view of the ground scene.

The organization of this paper is as follows: In Sect. 2, the imaging system with the proposed hyperbolic reflector is briefly explained. The image acquisition model of the imaging system and the image reconstruction algorithm are described in Sects.  3 and 4, respectively. Experiments to verify the performance of the hyperbolic reflector imaging system and concluding remarks are presented in Sects. 5 and 6.

2 Imaging System with the Hybrid Hyperbolic Reflector

The hybrid hyperbolic reflector presented in this study has an upper section of a cylindrical hyperbolic reflector, and a lower section of a half-omnidirectional hyperbolic reflector. The hyperbolas in the upper and the lower sections share a common focal point with surface continuity. According to the property of the hyperbolic curve, it is possible to acquire a wide FOV image by using a conventional camera at the symmetric focal point of the hyperbolic reflector, as shown in Fig. 2. The horizontal FOV of the image is over \(180{^{\circ }}\) because of the upper cylindrical hyperbolic reflector, and the vertical downward FOV is over \(90{^{\circ }}\) from the lower omnidirectional hyperbolic reflector. Figure 3 shows the design of the hybrid hyperbolic reflector.

Fig. 2
figure 2

Image acquisition using the hybrid hyperbolic reflector

Fig. 3
figure 3

Design of the hybrid hyperbolic reflector

The hyperbolic function in the \(x-y\) plane is described as

$$\begin{aligned} \frac{x^{2}}{a^{2}}-\frac{y^{2}}{b^{2}}=-1 \end{aligned}$$
(1)

where a and b are the design parameters of the function. A pair of symmetric focal points of the hyperbolic function is represented herein as \(\left[ {0\;\;F} \right] ^{ t}=\left[ {0\;\;\sqrt{a^{2}+b^{2}}} \right] ^{ t}\) and \(\left[ {0\;\;{F}'} \right] ^{ t}=\left[ {0\;\;-\sqrt{a^{2}+b^{2}}} \right] ^{ t}\).

3 Image Acquisition Model

The image acquisition model is a mathematical relationship between an object point, \(P_o =\left[ {x_o \;y_o \;z_o } \right] ^{ t}\), in three-dimensional space and a corresponding image point, \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\), on the image plane. In the coordinate system shown in Fig. 2, it is assumed that the y coordinate value of the image plane is constant as \(y_i =F+\Lambda \), where \(\Lambda \) is the focal length of the camera. It is possible to obtain the image acquisition model of the catadioptric imaging system based on the geometrical optics. The object point in the celestial sphere is represented also by the longitude and the latitude angles as \(P_o =\left[ {\lambda \;\;{\varphi }} \right] ^{ t}\), where \(\lambda \) and \({\varphi }\) are given by

$$\begin{aligned} \begin{array}{l} \lambda =\tan ^{-1}\frac{x_o }{y_o -{F}'} , \\ {\varphi }=\tan ^{-1}\frac{z_o }{\sqrt{x_o ^{2}+(y_o -{F}')^{2}}} \\ \end{array} \end{aligned}$$
(2)

The longitude and the latitude angles of an object point are described with respect to \({F}'\) without loss of generality and the imaging camera is modeled as an ideal pinhole without lens distortion in this study.

3.1 Imaging Model for the Upper Cylindrical Hyperbolic Reflector

The upper section of the cylindrical reflector is the same as the hyperbolic reflector in the horizontal \(x-y\) plane and the planar reflector in the vertical z direction. It is convenient to describe the imaging model of the upper section separately in the horizontal plane and in the vertical direction.

Fig. 4
figure 4

Image acquisition model in the horizontal plane (\(x-y)\) for the upper section of the hyperbolic reflector (top view)

3.1.1 Horizontal Plane

Figure 4 shows the ray tracing in the horizontal plane for the upper section of the reflector. The horizontal cross-section is the hyperbolic curve represented as (1) above. A light ray from an object, \(P_o \) toward \({F}'\) changes its direction for F after reflection on the hyperbolic reflector surface. It has an image at \(P_i \) on the image plane through the pinhole of a camera at F. In Fig. 4, the light rays I and II are described as

$$\begin{aligned}&\hbox {Line I:} \ \frac{x}{\tan \lambda }=y-{F}' \end{aligned}$$
(3)
$$\begin{aligned}&\hbox {Line II:} \ \frac{x}{x_c }=\frac{y-F}{y_c -F} \end{aligned}$$
(4)

Because the reflection point, \(P_c =\left[ {x_c y_c } \right] ^{ t}\) on the horizontal plane is an intersection of (1) and (2), it is obtained as follows:

$$\begin{aligned} \begin{array}{l} x_c =n \tan \lambda , \\ y_c =n+{F}' \\ \end{array} \end{aligned}$$
(5)

where n is a solution of the following quadric equation

$$\begin{aligned} \left( {b^{2}\tan ^{2}\lambda -a^{2}} \right) n^{2}-2{F}'a^{2}n+a^{2}\left( {b^{2}-{F}'^{2}} \right) =0 \end{aligned}$$
(6)

By inserting (5) into (4) and evaluating x with \(y_i =F+\Lambda \) at the image plane, the relationship between an object point \(P_o \) and the corresponding image point \(x_i \) of \(P_i \) is obtained as follows:

$$\begin{aligned} x_i =\frac{\Lambda x_c }{y_c -F} \end{aligned}$$
(7)

3.1.2 Vertical Direction

At the reflection point, the upper section of the reflector is the same as a planar reflector in the vertical z direction. The imaging model of the planar reflector is described easily by using the effective viewpoint as depicted in Fig. 5a. The effective viewpoint is the symmetric point of the camera pinhole at F with respect to the tangential plane at the reflection point, \(P_c \) [23]. As shown in Fig. 5a, a light ray from the object point passes through the reflection point and the effective viewpoint, and has an image on the effective image plane. Those points are placed on a common vertical plane shown in Fig. 5b, where the image point \(z_i \) is obtained by

$$\begin{aligned} z_i =-\frac{t_i }{t_p +t_o }\cdot z_o \end{aligned}$$
(8)

where \(t_o \), \(t_p \), and \(t_i \) are given by

$$\begin{aligned} \begin{array}{l} t_o =\sqrt{(x_o -x_c )^{2}+(y_o -y_c )^{2}}, \\ t_p =\sqrt{x_c ^{2}+(y_c -F)^{2}}, \\ t_i =\sqrt{x_i ^{2}+\Lambda ^{2}} \\ \end{array} \end{aligned}$$
(9)

The value of \(x_c \), \(y_c \), and \(x_i \) for evaluating (9) is obtained in (5) and (6) as stated previously.

Fig. 5
figure 5

Image acquisition model in the vertical axis (\(x-z)\) for the upper part of the hybrid hyperbolic reflector. a Image acquisition model using effective viewpoint (top view) and b common vertical plane through four points of a light ray

As a summary of (6) and (7), the imaging model between an object point \(P_o =\left[ {x_o \;y_o \;z_o } \right] ^{ t}\) and a corresponding image point \(P_i =\left[ {x_i z_i } \right] ^{ t}\) is written as

$$\begin{aligned} \begin{array}{l} x_i = \frac{\Lambda x_c }{y_c -F}, \\ z_i =\frac{\Lambda }{y_c -F}\cdot \frac{1}{1+\frac{\sqrt{\left( {x_o -x_c } \right) ^{2}+\left( {y_o -y_c } \right) ^{2}}}{\sqrt{x_c^2 +\left( {y_c -F} \right) ^{2}}}}\cdot z_o , \\ \end{array} \end{aligned}$$
(10)

where \(x_c \) and \(y_c \) are given by (5). As the object point is described as \(P_o =\left[ {\lambda \;\;{\varphi }} \right] ^{ t}\), the image point in (9) is represented by

$$\begin{aligned} \begin{array}{l} x_i =\left. {f_x (\lambda , {\varphi })} \right| _{a,b,\Lambda } , \\ z_i =\left. {f_z (\lambda , {\varphi })} \right| _{a,b,\Lambda } , \\ \end{array} \end{aligned}$$
(11)

where a, b, and \(\Lambda \) are the design parameters of the imaging system.

3.2 Imaging Model for the Lower Omnidirectional Hyperbolic Reflector

In the coordinate system shown in Fig. 2, the lower section of the hyperbolic reflector is described as follows:

$$\begin{aligned} \frac{x^{2}+z^{2}}{a^{2}}-\frac{y^{2}}{b^{2}}=-1 \end{aligned}$$
(12)

Figure 6 shows the ray tracing in the vertical cross-sectional \(y-z\) plane of the lower section of the reflector.

Fig. 6
figure 6

Ray tracing for lower part of the hyperbolic reflector (side view)

The light rays I and II are represented in terms of \(\lambda \) and \({\varphi }\) in 3-dimensional space as (13) and (14), respectively:

$$\begin{aligned} \hbox {Line I:} \ \frac{x}{\tan \lambda }= & {} y-{F}'=\frac{z}{\tan {\varphi }\sqrt{1+\tan ^{2}\lambda }} \end{aligned}$$
(13)
$$\begin{aligned} \hbox {Line II:} \ \frac{x}{x_c }= & {} \frac{y-F}{y_c -F}=\frac{z}{z_c } \end{aligned}$$
(14)

It is possible to obtain the reflection point \(P_c =\left[ {x_c y_c z_c } \right] ^{ t}\) on the hyperbolic curve as the intersection of (12) and (13) as follows:

$$\begin{aligned} \begin{array}{l} x_c =k\tan \lambda , \\ y_c =k+{F}' , \\ z_c =k\tan {\varphi }\sqrt{1+\tan ^{2}\lambda } , \\ \end{array} \end{aligned}$$
(15)

where k is the solution of the following quadric equation

$$\begin{aligned} \begin{array}{l} \left[ {b^{2}\left\{ {\tan ^{2}\lambda +\tan ^{2}{\varphi }(1+\tan ^{2}\lambda )} \right\} -a^{2}} \right] k^{2} \\ \quad \quad \quad \quad \quad \quad \quad -2{F}'a^{2} k-a^{2}({F}'^{2}-b^{2})=0 . \\ \end{array} \end{aligned}$$
(16)

Inserting (15) into (14) and evaluating x and z at the image plane\(\left( {y_i =F+\Lambda } \right) \) gives the image points \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\) as follows:

$$\begin{aligned} \begin{array}{l} x_i =k\tan \lambda \frac{\Lambda }{k-2F} \\ z_i =k\tan {\varphi }\sqrt{1+\tan ^{2}\lambda }\frac{\Lambda }{k-2F} . \\ \end{array} \end{aligned}$$
(17)

The image acquisition model is summarized in terms of \(\lambda \) and \({\varphi }\) of an object point as

$$\begin{aligned} \begin{array}{l} x_i =\left. {g_x (\lambda , {\varphi }) } \right| _{a, b, \Lambda } \\ z_i =\left. {g_z (\lambda , {\varphi }) } \right| _{a, b, \Lambda } \\ \end{array} \end{aligned}$$
(18)

4 Image Reconstruction

It is possible to observe the side scene as well as the ground scene by the imaging system with the hyperbolic reflector in this study. Image reconstruction converts the side view from the upper section of the reflector into a horizontal panoramic view and the ground view from the lower section into a vertical bird’s-eye view.

4.1 Horizontal Panoramic View from the Upper Section of the Reflector

For each image point \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\), it is possible to obtain \(P_o =\left[ {\lambda \;\;{\varphi }} \right] ^{ t}\) for the corresponding object point on the basis of the image acquisition model in Sect. 3. The well-known Mercator projection is a conformal map that rearranges an object point on the surface of the Earth to a map point on a rectangular surface according to \(\lambda \) and \({\varphi }\) while preserving linear scale. Figure 7 shows image reconstruction using the Mercator projection for a part of the original image from the upper section of the reflector. In the figure, the coordinate frame is drawn at \({F}'\) to show that \(\lambda \) and \({\varphi }\) of an object point are described with respect to \({F}'\). According to the object point \(P_o =\left[ {\lambda \;\;{\varphi }} \right] ^{ t}\) from an original image point, the reconstructed image point, \(P_w =\left[ {x_w \; y_w } \right] ^{ t}\), is represented as follows [24]:

$$\begin{aligned} \begin{array}{l} x_w =l (\lambda -\lambda _o ), \\ y_w =l \ln \left[ {\tan \left( {\frac{\pi }{4}+\frac{{\varphi }}{2}} \right) } \right] \\ \end{array} \end{aligned}$$
(19)

where l is a scale factor and \(\lambda _o \) denotes the center of the longitudes, that is, the center of the vertical lines of the original image acquired. In this image reconstruction algorithm, object points at the same longitude are placed on the same vertical line in the reconstructed image, and object points at the same latitude are placed on the same horizontal line in the reconstructed image, resulting in a horizontal panoramic image.

Fig. 7
figure 7

Image reconstruction of the horizontal panoramic image

Fig. 8
figure 8

Image reconstruction for the bird’s-eye view

Fig. 9
figure 9

Imaging system with the hyperbolic reflector

4.2 Bird’s-Eye View from the Lower Section of the Reflector

Figure 8 illustrates the image reconstruction for the ground plane image from the lower section of the reflector, resulting in a bird’s-eye view image. In the coordinate system shown in Fig. 2, the ground plane equation is described as

$$\begin{aligned} z=-h \end{aligned}$$
(20)

As an intersection of the plane equation (20) and the line equation (13), it is possible to obtain the coordinates of the object points on the ground plane by

$$\begin{aligned} \begin{array}{l} \left. {x_o } \right| _{z=-h} =-\frac{h\tan \lambda }{\tan {\varphi }\sqrt{1+\tan ^{2}\lambda }}, \\ \left. {y_o } \right| _{z=-h} ={F}'-\frac{h}{\tan {\varphi }\sqrt{1+\tan ^{2}\lambda }} \\ \end{array} \end{aligned}$$
(21)

The bird’s-eye view of the ground plane is obtained as (22) by rearranging the coordinates of the object points (21) onto the reconstructed image plane:

$$\begin{aligned} \begin{array}{l} x_w =m \left. {x_o } \right| _{z=-h} \\ y_w =m \left. {y_o } \right| _{z=-h} \\ \end{array} \end{aligned}$$
(22)

where m is a scale factor.

5 Experiments

The imaging system with the hyperbolic reflector developed in this study is shown in Fig. 9. The camera has a resolution of \(1280\times 960\). Table 1 summarizes the parameter values of the reflector in Fig. 3.

Table 1 Parameters of the hyperbolic reflector (mm)
Fig. 10
figure 10

Camera placement for the central imaging. a Non-central image, b central image, c bird’s-eye view of the lower part of the central image

Fig. 11
figure 11

Image acquisition and reconstruction. a Original image, b reconstruction image of the upper part (side view panoramic), c reconstruction image of the lower part (bird’s-eye view)

The placement of the camera relative to the reflector is fundamental for developing a central catadioptric imaging system with a single viewpoint. The central imaging system has low computational cost while finding the reflection point on the reflector surface and the corresponding object point without causing distortion that would induce errors onto the image. Considerable research has been conducted on the calibration and placement of the catadioptric imaging system [25, 26]. This study adopts the following steps for imaging system calibration and camera placement: (1) employing the publicly available MATLAB calibration toolbox [27], which allows the ideal pinhole model of the camera and (2) manually adjusting the camera placement with respect to the reflector for the single viewpoint constraint as described in [28]. In the central imaging system with a hyperbolic reflector, a line connecting the vanishing points of a set of circles fitted on the images of the parallel three-dimensional lines intersects at the optical center of the image [28]. Figure 10 shows the calibration result using parallel three-dimensional lines in a grid pattern. In Fig. 10a, a set of circles fitted on the three-dimensional lines of the non-central image does not have a common vanishing point. Instead, the circles on the central image have common vanishing points and the line connecting the vanishing points intersects the camera optical center in Fig. 10b. Figure 10c is the bird’s-eye view of the lower part of the central image.

For the image reconstruction, an inverse mapping method is adopted to reduce the overall computational burden. According to (11) for the lower side of the image or (18) for the upper side of the image, it is possible to find a pixel \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\) from the original image corresponding to \(P_o =\left[ {\lambda \; {\varphi }} \right] ^{ t}\). Then, rearranging the pixel \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\) onto \(P_w =\left[ {x_w \; y_w } \right] ^{ t}\) in accordance with (19) or (22) results in the image reconstructed from the original image. The well-known bilinear interpolation method is adopted to interpolate the non-integer \(P_i =\left[ {x_i \; z_i } \right] ^{ t}\) [29].

Fig. 12
figure 12

Bird’s-eye view according to the height of the imaging system. Actual length of the grid pattern is 4846 mm. a Ground image in accordance with the height of the imaging system, b original images with different heights: Left: \(h_1 =1395 mm\), right: \(h_2 =1017 mm\) and c reconstruction of the bird’s-eye views using the same height, \(h=1395 mm\)

For a vehicle experiment, the imaging system is attached at the center of the left side of a vehicle. Figure 11 shows the results of the experiment. The horizontal panoramic view in Fig. 11b and the vertical bird’s-eye view in Fig. 11c are reconstructed from the original image in Fig. 11a. Those views are wide and natural to a driver. The images in Fig. 11b, c have low resolution around the left and the right sides because of the non-uniform resolution of the catadioptric imaging system with the hyperbolic reflector.

In the vehicle application, the height of the imaging system can change depending on the specific wheels and tires used or the ground condition, which may cause distortion in the image. Figure 12 shows the influence of the imaging system height on the reconstruction of the bird’s-eye view for the ground plane. Figure 12a explains that the central imaging system has only the object size change according to the height without further distortions in the image. The size change in the image in accordance with the height is shown in Fig.  12b. To achieve exact reconstruction, the bird’s eye view algorithm in (21) requires the actual height, h of the imaging system. The right side of Fig. 12c shows the size change of the same grid pattern reconstructed using the false height information.

6 Concluding Remarks

The AVM system can secure the safety of a vehicle in driving or parking. In order to reduce the number of cameras required to observe the surrounding area of a vehicle as widely as possible, a wide FOV imaging method is essential for the AVM system. The catadioptric imaging system using the hybrid hyperbolic reflector proposed in this study can observe the horizontal side view as well as the ground view surrounding a vehicle with a wide FOV. The reflector consists of an upper section of a cylindrical hyperbolic reflector and a lower section of an omnidirectional hyperbolic reflector. The upper and lower sections of the reflector share a common focal point with surface continuity. From the geometrical optics, the image acquisition model for the proposed imaging system is obtained in this study. The image acquisition model is used to reconstruct an image from the imaging system and to present the reconstructed natural image in the driver’s display. Experimental results showed a natural panoramic view from the horizontal side scene and a vertical bird’s-eye view from the ground scene surrounding a vehicle. Both reconstructed views are useful for the AVM to improve the safety of a vehicle.