Introduction

Since the very start of modern neurosurgery at the end of the nineteenth century, neurosurgical progress has been intimately related to improvements in intracranial localization. The knowledge of spatial relationships of lesions inside the cranium and the development of atraumatic approaches contributed essentially to reduced mortality and morbidity of neurosurgical interventions. Localization encompasses two questions: (1) where in the cranium the lesion or functional area must be assumed to be and (2) how to find this lesion or area during operation. This article will focus on the latter intraoperative aspect. The first question relates mainly to neuroradiologic developments and will be mentioned shortly first, because this imaging technique is also the basis and starting point for intraoperative localization methods.

At the beginning of our century, the topic of diagnostic localization of lesions could be understood only by analyzing the neurologic symptoms of the patient, without the possibility of referring to radiologic images. The first imaging technique, introduced by Dandy in 1918 [33], was visualization of the ventricles by direct injection of air and later of contrast medium into the ventricles or cisterna magna. From the shape and kind of ventricular displacement, lesions close to the midline structures could be detected. Later, in 1927, angiography was described by Egas Moniz [143]. This technique allowed localization of intracranial lesions either directly by visualization of the pathologic vessels or indirectly by associating the lesion with a lobe, depending on the characteristic frontal, temporal, or occipital displacement of the intracerebral arteries. Direct visualization of the cerebral tissue was not possible before CT was introduced by Hounsfield in 1973 [102].

Intraoperative localization technique, including positioning of the craniotomy and the direction of preparation, was based on knowledge of characteristic bony landmarks such as coronal suture, protuberantia occipitalis externa, etc., and the neurosurgeon's skill and 3D knowledge. As the operation proceeded, in particular at the skull base, identifying anatomic guiding structures, particularly cranial nerves, vessels, and characteristic bony landmarks, served as points of spatial orientation. This anatomic localization method was the gold standard not only before CT and MR imaging but also after their introduction. Microneurosurgeons used this detailed anatomic image information for better intraoperative identification of the anatomic structures and sophisticated planning of approaches [155, 156, 241, 242, 243, 244, 245, 246].

Parallel to the anatomic localization strategy, there was also the tendency from the beginning of neurosurgery to define the anatomic and pathologic structures in advance by using mechanical devices to define accurate approaches to targets and to provide objective information independent of the individual surgeon's skills. Frame-based stereotactic localization techniques were developed based on a rigid coordinate system in which the target and a straight trajectory were determined based on the image information. However, until recently microneurosurgeons were more comfortable with intraoperative anatomic identification than with stereotactic coordinates. The reason for this is not only that microsurgeons like to decide each step separately, depending on the changing anatomic situation in the course of operation but also because they trust more what they see rather than what computers calculate in abstract coordinates.

Frame-based and robotic technologies forced neurosurgeons to adapt their microsurgical techniques to the rigid stereotactic systems, which worked with high accuracy but reduced flexibility. Although frame-based stereotactic systems increased their intraoperative flexibility, it was computer-based navigation which, working interactively with images without visible coordinates and without instruments in the operating field, suited the visually and morphologically oriented neurosurgeons.

Definitions for interactive localizing techniques

At present, several terms are used for interactive localizing: computer-assisted surgery, computer-integrated surgery, computer-aided surgery, image-guided surgery, navigated surgery, and frameless stereotaxis.

Abdominal, head and neck, orthopedic, and ear, nose, and throat (ENT) surgeons prefer the term "computer-assisted surgery" (CAS), with "computer-integrated" or "computer-aided" surgery meant to cover a wide range of computer applications during surgery such as robotics, 3D rendering, surface simulations, and intraoperative localization. In this sense, the definition of CAS is too broad for neurosurgical purposes and does not characterize the neurosurgical applications. The aim is to define an operative technique which is basically microsurgical but makes use of advanced computer technology for better orientation during some parts of the surgical procedure. Therefore, frame-based stereotactic operations and robot-based interventions, which plan their complete approach in advance, will not be discussed in this article aside from their historical contributions. In neurosurgical literature, the terms image-guided surgery, navigated surgery, and frameless stereotaxy are more common for interactive computer-aided procedures.

The term image-guided surgery is too general, including by definition intraoperative imaging with ultrasound, CT, or MRI, which are not based on coordinate transformations and should be treated separately. On the other hand, frameless stereotaxy, in contrast to frame-based stereotaxy, indicates the frame as the main difference between classic stereotaxy and interactive computer-based localization. The essential difference between them is the different calculation of the position of the instruments in space and the interactivity with rapid real time visualization of the target point in the images and without use of a frame. There are even interactive localization devices that use frames to improve registration accuracy.

The imprecise metaphoric notation "navigated surgery" better describes the idea and aims behind this type of surgery. In Latin, "navigare" means to guide a ship on the sea. The interactive determination of the position during the journey was a presupposition for successful completion of a voyage. Apuzzo [6] pointed out the similarity of the basic concepts and the link, despite the shift in time, between nautical and cerebral navigation. In maritime localization, accurate cartography of islands and continents is just as necessary as atlases of the cerebral structures in neurosurgery. The interactive nautical localization was derived primarily from the positions of stars as fixed fiducials in the firmament. In Homer's Odyssey, we already find [98] a description of simple celestial navigation:

And he sat and guided his raft skillfully with the steering oar, nor did sleep fall upon his eyelids, as he watched the Pleiads, and late-setting Bootes and the Bear, which man also call the Wain, which ever circles where it is and watches Orion, and alone has no part in the bath of Ocean. For this star Calypso, the beautiful goddess, had bidden him to keep on the left hand as he sailed over the sea. —Homer, Odyssey V, pp 270–277 [98]

By means of sextants, the positions of the stars were later measured with greater accuracy using a polar coordinate system with angles, azimuth, and declination. More recently, maritime orientation is based on satellite technology which detects and calculates with high accuracy any transmitter source on land or sea. Except for dimensions, the present GPS localization methods have very much in common with armless cerebral navigation using LEDs as emitter sources and a camera system for localization of a pointer tip in space. In nautical navigation, the visualization of the pointer tip in the corresponding CT/MR images corresponds to the direct visualization of one's own position on a map. The main difference between maritime and cerebral navigation is the change and displacement of anatomic structures during operation. The hypothetical marine analogy would be that, during a voyage, the cartography changes, making constant actualization of the charts necessary. This is exactly the problem in contemporary cerebral navigation and can be overcome only by intraoperative actualization of the image data [74, 236].

History of frame-based localization

Devices for the mechanical localization of intracranial structures go back to the beginning of modern neurosurgery. In Moscow, Zernow described in 1890 [55, 64, 251] an apparatus called an "encephalometer" which was fixed to the patient's skull and served to localize intracerebral structures based on superficial landmarks. His pupil Altuchow investigated the position of the basal ganglia using this encephalometer [5]. Actual stereotactic calculation based on a coordinate system was invented by Horsley and Clarke in 1906 [27] and 1908 [100]. They described a rigid frame attached to the skull which served as an immovable coordinate system in relation to which each point in the brain could be referred. They developed an arc system as a stable holder for introducing instruments to the target points, which were derived from brain atlases. This method was used only for experimental studies.

Stereotactic methods in humans were applied by Kirschner in 1933 [108] to puncture the foramen ovale at the skull base to treat patients with idiopathic trigeminal neuralgia. Stereotactic calculation of targets in the deep brain areas was, however, hindered in humans as opposed to animals by individual variations in the intracranial structures and by the lack of appropriate imaging techniques to visualize these variations.

Dandy [33] invented ventriculography in 1918 by performing an X-ray in a patient with an open, penetrating head injury and seeing the configuration of the ventricles filled with air. Then he described a method to introduce air artificially into the ventricles to investigate deep brain structures. Ventriculography was also the fundamental technique for calculating targets in the thalamus and basal ganglia because they have a stable and definite relationship to the topography of the third ventricle.

Spiegel and Wycis [207, 208] performed the first stereotactic thalamotomy in humans in 1947 using the commissura posterior or pineal body as an internal individual reference system. Functional operations with similar frames and techniques were introduced by Talairach 1949 in Paris [209], by Riechert in 1952 [170] in Freiburg, Germany, and by Leksell 1949 [122] in Stockholm for the treatment of extrapyramidal movement disorders, intractable pain, epilepsy, and psychiatric disorders.

After the development of CT technology by Hounsfield in 1973 [102] and Cormack [28, 29] based on mathematical solutions published by the Viennese mathematician Radon in 1917 [164], stereotactic coordinate-based calculation was applicable in the whole intracranial space, enlarging the field of indications to biopsies, interstitial brachytherapy, endoscopy, and localization of tumors for open surgery [7, 93].

Till the end of the 1980s, frame-based stereotaxy was the standard method for accurately localizing small intracranial lesions by introducing catheters into the tumors [15, 48, 49, 123] or for determining the tumor volume in space [107]. Coordinate transformation of the selected target point between the image and the frame space was established using a localization frame. Some of these stereotactic systems, for instance, BRW [228] and Cosman-Roberts-Wells (CRW) frames, used the rods of the localization frame for calculating a transformation matrix identical to the paired-point transformation method of frameless stereotaxy. The three oblique rods of the localization frame served as fiducial markers, which additionally made use of an optimal spatial arrangement.

History of frameless localization and robotics

The idea of frameless, interactive, computer-aided surgery consisted in navigation systems able to show in real time the position of the tip of an instrument in the corresponding images and not requiring a stereotactic frame for calculation. Because the navigation devices were basically adaptations of robot technology to surgical applications in regard to all the mathematical and technical knowledge necessary for real-time navigation, it seems justified to start the historical review with robots.

The concept of robots preceded its technical realization by 2,700 years and was described by Homer in the 18th chapter of the Iliad, when Thetis was visiting the god Hephaistos in Olympus to ask for new armor for her son Achilles [97]. Thetis observed two types of robots in the smithy: one type was machines on wheels which moved intelligently and automatically. The second type was androids of gold resembling girls equipped with the cognitive capacity of living creatures who helped Hephaistos in the smithy:

Him she found sweating with toil as he moved to and fro about the bellows in eager haste; for he was fashioning tripods, twenty in all, to stand around the wall of his well-built hall, and golden wheels had he set beneath the base of each that of themselves they might enter the gathering of the gods at his wish and again return to his house, a wonder to behold. —Homer, Iliad XVIII, pp 372–378 [97]

And he put upon him a tunic, and grasped a stout staff, and went forth halting; but there moved swiftly to support their lord hand-maidens wrought of gold in the semblance of living maids. In them is understanding in their hearts, and in them speech and strength and they know cunning handiwork by gift of the immortal gods. —Homer, Iliad XVIII, pp 416–420 [97]

Because such machines moved from internal impetus without external impulse, they were called automataFootnote 1. The term automaton regained importance in the sixteenth and seventeenth centuries in Cartesian philosophy [38]. With the technical revolution in the eighteenth and nineteenth centuries, the ideas of automats were again very popular. People's fantasies were very similar to Homer's in wishing to build either machines or androids which would work instead of men. In the 1920s and 1930s, these androids were the starting point of a broad science fiction genre beginning with the Czech writer Karel Capek [23].

Capek also created the term robot. Robota means "to work" in Czech, and a robot (a machine working instead of a man by himself) was pictured by him in 1921 in a drama called "Rossum's Universal Robots." However, the first actual robots, called teleoperators, were developed in the Second World War to handle radioactive material and allow operators to perform tasks at a distance [70]. In 1949, the U.S. Air Force sponsored a program for a numerically controlled milling machine [185] which combined sophisticated servo-controlled systems with newly developing digital computer technology.

Before 1961, robots were not able to define their own position in space. At that time, tongs of a teleoperator slave arm were equipped with sensors to control the movements [185]. The first systems that could calculate their own position in space were developed in the late 1960s. Roberts [171, 172] demonstrated the possibility of defining the position of objects in space using a camera system and expressing their position and orientation by homogeneous coordinates.

The first robot driven by optical control was developed in 1967 by Wichmann [235]. Later, arm-based systems equipped with encoders to calculate angles in the joints were described by Pieper in 1968 [158]. He applied the theory of closed-link chains to obtain the mathematical solution for calculating the positions of device-related coordinates. These robots were able to calculate their own position in space and move to defined positions applying kinematic equations [37, 193].

Modified industrial robots with higher positioning accuracy were used in neurosurgery in the beginning of 1990s by Drake [43], Kwoh [117], Benabid [13], Fankhauser and Glauser [56, 67], Masamune [134], and Young [247]. In the middle of the 1980s, all the necessary mathematical and technical presuppositions existed for the realization of devices for navigated surgery:

  1. 1.

    Fast computers with appropriate data banks to handle image-based information in real time

  2. 2.

    Hardware and software for manipulators based on industrial robot technology with a high degree of accuracy

  3. 3.

    A high technical standard of image processing partly due to space research programs

This technology had to be assembled in such a way as to satisfy surgical requirements. Passive manipulators, digitizers, and sensory arms which were able to determine their own position in space and transfer this information into images were introduced into neurosurgery in 1986 by Roberts [180]. He adapted a microscope for navigated surgery. It was equipped with ultrasound-emitting sources and microphones arranged outside the operating field. The microphone data were transported to a computer, which calculated the position of the microscope in space. The target point chosen on the CT/MR images could be projected into the ocular of the microscope and used for orientation during the operation.

In 1987 Schlöndorf [197] developed an arm-based system for ENT applications. This navigation system was later applied by Mösges [144] and Adams [1] to neurosurgical purposes. In 1987, Watanabe independently introduced an arm-based navigation system for neurosurgical operations [113, 230, 231]. In Switzerland in 1988, Reinhardt [165, 166] was working on an armless navigation system which used a pointer emitting ultrasound sources. Magnetic sources were also described later by Kato [105], and infrared light-emitting diodes (LEDs) as emitting sources by Zamorano [248]. Additional robotic capabilities were integrated into navigated microscopes by Giorgi [65] and Luber [126].

Basic principles of navigated surgery

To see the tip of a pointer in an image space, a relationship between the device space and the image space has to be established. This operation is called registration or calibration of the navigation device. Basically, a transformation matrix (T) has to be calculated which maps the coordinates of any point between the image and the device spaces. Techniques with different mathematics have been proposed:

  1. 1.

    Fiducial-based paired-point transformation

  2. 2.

    Surface contour matching

  3. 3.

    Hybrid transformation

Paired-point transformation

Paired-point matching uses corresponding sets of points whose coordinates must be established in the image and in the device-related space [182]. Most common are markers stuck to the skin [74, 176, 213]. Because these skin markers can shift, other more invasive solutions have been proposed to improve accuracy. Reinhardt [166] returned to a frame equipped with an array of sonic emitters fixed to a stereotactic frame, and Rousseau [186] put markers on the stereotactic frame to have fixed fiducials. Another possibility is to implant screws in the skull [19, 138]. Instead of artificial markers, corresponding anatomic structures can be also used as fiducials [51].

Such calibration is valid only if the patient does not move during surgery. To make possible intraoperative changes in patient position (as is common in neurosurgery during operation by moving the operating table with the patient fixed to the table in a rigid clamp), recalibration of the system is necessary. This can be solved by a dynamic referential frame consisting of an array equipped with LEDs which is fixed to the operating table. This solution provides only for relative changes of the operating table. Ryan [188] proposed to fix an LED array with a single screw directly to the skull of the patient to compensate for any displacement of the head during surgery. This problem is particularly urgent in ENT surgery, because the patient's head is not fixed at all during this type of intervention. Under these conditions, continuous recalibration using an array of LEDs fixed to the head is mandatory. Instead of an array, Kai [104] proposed the implantation of single LEDs as markers.

From a mathematical point of view, at least three paired points are necessary to establish a transformation matrix (T) between both spaces. If, however, homogeneous coordinates are used, which is common in robot technology, the number of fiducials has to be at least four. The basic equation (equation 1) for coordinate transformation can be expressed in compact form in matrix notation:

$$ \matrix{ {\left( {\rm X} \right)} & {\left( {\rm T} \right)} & {\left( {{\rm X}^{\rm *} } \right)} \cr {\left( {\matrix{ {{\rm x}_{\rm 1} } & {{\rm y}_1 } & {{\rm z}_{\rm 1} } \cr {{\rm x}_{\rm 2} } & {{\rm y}_2 } & {{\rm z}_{\rm 2} } \cr {{\rm x}_{\rm 3} } & {{\rm y}_3 } & {{\rm z}_{\rm 3} } \cr } } \right)} & {\left( {\matrix{ {{\rm t}_{{\rm 11}} } & {{\rm t}_{{\rm 12}} } & {{\rm t}_{{\rm 13}} } \cr {{\rm t}_{{\rm 21}} } & {{\rm t}_{{\rm 22}} } & {{\rm t}_{{\rm 23}} } \cr {{\rm t}_{{\rm 31}} } & {{\rm t}_{{\rm 32}} } & {{\rm t}_{{\rm 33}} } \cr } } \right) = } & {\left( {\matrix{ {{\rm x}^* _1 } & {{\rm y}^* _1 } & {{\rm z}^* _1 } \cr {{\rm x}^* _2 } & {{\rm y}^* _2 } & {{\rm z}^* _2 } \cr {{\rm x}^* _3 } & {{\rm y}^* _3 } & {{\rm z}^* _3 } \cr } } \right)} \cr } $$
(1)

The matrix (X) on the left consists of the x, y, and z coordinates of the three markers in relation to the device space. The transformation matrix (T) refers to the nine variables t, which have to be determined. On the right side of the equation is the matrix (X*) of the x*, y*, and z* coordinates of the three markers in relation to the image space. The nine variables t of the transformation matrix can be calculated by multiplying matrices (X) and (T), leading to a system of linear equations which must be solved (equation 2):

$$ \matrix{ {{\rm x}_{\rm 1} {\rm t}_{{\rm 11}} + {\rm y}_{\rm 1} {\rm t}_{{\rm 21}} + {\rm z}_{\rm 1} {\rm t}_{{\rm 31}} } & {\rm = } & {{\rm x}^{\rm *} _{\rm 1} } \cr {{\rm x}_{\rm 2} {\rm t}_{{\rm 11}} + {\rm y}_{\rm 2} {\rm t}_{{\rm 21}} + {\rm z}_{\rm 2} {\rm t}_{{\rm 31}} } & {\rm = } & {{\rm x}^{\rm *} _{\rm 2} } \cr {{\rm x}_{\rm 3} {\rm t}_{{\rm 11}} + {\rm y}_{\rm 3} {\rm t}_{{\rm 21}} + {\rm z}_{\rm 3} {\rm t}_{{\rm 31}} } & {\rm = } & {{\rm x}^{\rm *} _{\rm 3} } \cr {} & {} & {} \cr {{\rm x}_{\rm 1} {\rm t}_{{\rm 12}} + {\rm y}_{\rm 1} {\rm t}_{{\rm 22}} + {\rm z}_{\rm 1} {\rm t}_{{\rm 32}} } & {\rm = } & {{\rm y}^{\rm *} _{\rm 1} } \cr {{\rm x}_{\rm 2} {\rm t}_{{\rm 12}} + {\rm y}_{\rm 2} {\rm t}_{{\rm 22}} + {\rm z}_{\rm 2} {\rm t}_{{\rm 32}} } & {\rm = } & {{\rm y}^{\rm *} _{\rm 2} } \cr {{\rm x}_{\rm 3} {\rm t}_{{\rm 12}} + {\rm y}_{\rm 3} {\rm t}_{{\rm 22}} + {\rm z}_{\rm 3} {\rm t}_{{\rm 32}} } & {\rm = } & {{\rm y}^{\rm *} _{\rm 3} } \cr {} & {} & {} \cr {{\rm x}_{\rm 1} {\rm t}_{{\rm 13}} + {\rm y}_{\rm 1} {\rm t}_{{\rm 23}} + {\rm z}_{\rm 1} {\rm t}_{{\rm 33}} } & {\rm = } & {{\rm z}^{\rm *} _{\rm 1} } \cr {{\rm x}_{\rm 2} {\rm t}_{{\rm 13}} + {\rm y}_{\rm 2} {\rm t}_{{\rm 23}} + {\rm z}_{\rm 2} {\rm t}_{{\rm 33}} } & {\rm = } & {{\rm z}^{\rm *} _{\rm 2} } \cr {{\rm x}_{\rm 3} {\rm t}_{{\rm 13}} + {\rm y}_{\rm 3} {\rm t}_{{\rm 23}} + {\rm z}_{\rm 3} {\rm t}_{{\rm 33}} } & {\rm = } & {{\rm z}^{\rm *} _{\rm 3} } \cr } $$
(2)

The matrix (T), applied to any point (x, y, z) in the device space, maps these coordinates of the pointer position in space into the corresponding coordinates (x*, y*, z*) in the image space (equation 3). The inverse matrix, which can easily be determined from the matrix (T), maps in the opposite direction from the image space into the device space. The inverse matrix (T)–1 is essential for microscope-based applications to visualize the target in the ocular of the microscope.

$$ \matrix{ {\matrix{ {\left( {{\rm x}\;{\rm y}\;{\rm z}} \right)} & {\left( {\matrix{ {{\rm t}_{{\rm 11}} } & {{\rm t}_{{\rm 12}} } & {{\rm t}_{{\rm 13}} } \cr {{\rm t}_{{\rm 21}} } & {{\rm t}_{{\rm 22}} } & {{\rm t}_{{\rm 23}} } \cr {{\rm t}_{{\rm 31}} } & {{\rm t}_{{\rm 32}} } & {{\rm t}_{{\rm 33}} } \cr } } \right)} \cr } } & {\rm = } & {\left( {{\rm x}^{\rm *} \;{\rm y}^{\rm *} \;{\rm z}^{\rm *} } \right)} \cr {} & {} & {} \cr {{\rm xt}_{{\rm 11}} + {\rm yt}_{{\rm 21}} + {\rm zt}_{{\rm 31}} } & {\rm = } & {{\rm x}^{\rm *} } \cr {{\rm xt}_{{\rm 12}} + {\rm yt}_{{\rm 22}} + {\rm zt}_{{\rm 32}} } & {\rm = } & {{\rm y}^{\rm *} } \cr {{\rm xt}_{{\rm 13}} + {\rm yt}_{{\rm 23}} + {\rm zt}_{{\rm 33}} } & {\rm = } & {{\rm z}^{\rm *} } \cr } $$
(3)

In case more than three fiducial markers are used for determination of the 3×3-dimensional transformation matrix (T), the calculation is overestimated and has in general no solution. Nevertheless, more fiducials are useful for selecting the best three fiducials (or four in case of homogeneity) which produce the smallest root mean square (RMS) error between both coordinate systems. The other possibility is to use all fiducials and reduce the amount again to three fiducials with new coordinates (x**, y**, z**) under the additional boundary condition (equation 4) so that the square error of the whole overestimated linear equation system is a minimum. This so-called method of linear least squares was originally developed 1821 by the famous German mathematician K.F. Gauss for astronomical measurements of the orbit of the planetoid Ceres. This problem is, however, common in all empirical sciences with the number of observations higher than that of variables.

$$ \matrix{ {\left\| {{\rm x}^{\rm *} - {\rm Xt}} \right\|} & = & {\min } \cr {} & {} & {} \cr {{\rm X}^{\rm t} {\rm Xt}} & {\rm = } & {{\rm X}^{\rm t} {\rm x}^{\rm *} } \cr } $$
(4)

Surface-based transformation

Surface-based transformations are iterative methods which, by continuous scaling, translation, and rotation, try to identify two rigid objects, applying optimizing criteria by minimizing the distance or a particular feature between both objects. They are understood as rigid body transformations (neglecting nonlinear inhomogeneities leading to image distortions in MRI) with six parameters, three of rotation and three of translation. The rotation is represented either by (3×3) Euler matrices (R1) (R2) (R3) using the Euler angles (equation 5) or by means of a (3×3) matrix (R*) of unit quaternions (equation 6) [119]. The translation in 3D space is represented by an additional vector V (a, b, c) with three variables.

$$ \matrix{ {{\rm R} = } & {\left( {\matrix{ {{\rm cos}\alpha _{\rm 3} } & {{\rm sin}\alpha _{\rm 3} } & 0 \cr { - {\rm sin}\alpha _{\rm 3} } & {{\rm cos}\alpha _{\rm 3} } & 0 \cr 0 & 0 & 1 \cr } } \right)} & {\left( {\matrix{ 1 & 0 & 0 \cr 0 & {{\rm cos}\alpha _{\rm 2} } & {{\rm sin}\alpha _{\rm 2} } \cr 0 & { - {\rm sin}\alpha _{\rm 2} } & {{\rm cos}\alpha _{\rm 2} } \cr } } \right)} & {\left( {\matrix{ {{\rm cos}\alpha _{\rm 1} } & {{\rm sin}\alpha _{\rm 1} } & {\rm 0} \cr { - {\rm sin}\alpha _{\rm 1} } & {{\rm cos}\alpha _{\rm 1} } & {\rm 0} \cr 0 & 0 & {\rm 1} \cr } } \right)} \cr } $$
(5)
$$ \matrix{ {{\rm q = }\left( {{\rm q}_{\rm 0} \;{\rm q}_{\rm 1} \;{\rm q}_{\rm 2} \;{\rm q}_{\rm 3} } \right)^{\rm t} } \cr \matrix{ \matrix{ {{\rm q}_{\rm 0} \ge 0} & {{\rm q}_{\rm 0}^{\rm 2} + {\rm q}_{\rm 1}^{\rm 2} + {\rm q}_{\rm 2}^{\rm 2} + {\rm q}_{\rm 3}^{\rm 2} = 1} \cr } \hfill \cr \hfill \cr} \cr {{\rm R} = \left( {\matrix{ {{\rm q}_{\rm 0}^{\rm 2} + {\rm q}_{\rm 1}^{\rm 2} - {\rm q}_{\rm 2}^{\rm 2} - {\rm q}_{\rm 3}^{\rm 2} } & {2\left( {{\rm q}_{\rm 1} {\rm q}_{\rm 2} - {\rm q}_{\rm 0} {\rm q}_{\rm 3} } \right)} & {2\left( {{\rm q}_{\rm 1} {\rm q}_{\rm 3} + {\rm q}_{\rm 0} {\rm q}_{\rm 2} } \right)} \cr {2\left( {{\rm q}_{\rm 1} {\rm q}_{\rm 2} + {\rm q}_{\rm 0} {\rm q}_{\rm 3} } \right)} & {{\rm q}_{\rm 0}^{\rm 2} + {\rm q}_{\rm 2}^{\rm 2} - {\rm q}_{\rm 1}^{\rm 2} - {\rm q}_{\rm 3}^{\rm 2} } & {2\left( {{\rm q}_{\rm 2} {\rm q}_{\rm 3} - {\rm q}_{\rm 0} {\rm q}_{\rm 1} } \right)} \cr {2\left( {{\rm q}_{\rm 1} {\rm q}_{\rm 3} - {\rm q}_{\rm 0} {\rm q}_{\rm 2} } \right)} & {2\left( {{\rm q}_{\rm 2} {\rm q}_{\rm 3} + {\rm q}_{\rm 0} {\rm q}_{\rm 1} } \right)} & {{\rm q}_{\rm 0}^{\rm 2} + {\rm q}_{\rm 3}^{\rm 2} - {\rm q}_{\rm 1}^{\rm 2} - {\rm q}_{\rm 2}^{\rm 2} } \cr } } \right)} \cr } $$
(6)

Surface-based matching was originally described by Pelizzari in 1987 [152] and is well known as the "head and hat problem." For applications in navigated surgery, the corresponding surface of the patient has to be matched with the appropriate 3D reconstruction from his images. The patient surface can be performed by selecting a sufficient number of points or lines with the pointer on the patient's skin or with a laser contouring device [22, 72, 96, 211] which scans two-dimensionally and is able to recombine for a 3D contour of the skin. All mathematical realizations of this procedure try to minimize an abstract distance function D between both objects. This can be the Euclidean surface distance, voxel similarity, information entropy, mutual information, or other statistic functions such as for chamfer matching [120, 217, 250].

One problem of these iterative methods using functions such as distance or features which have to be minimized is the error due to local relative minima. This means that the algorithm stops the procedure if it reaches a relative but not necessarily the best possible matching.

Closely related to the matching technique for registration is the integration of different imaging modalities into the basic preoperative CT/MRI data set such as fluoroscopy in cases of spinal surgery [120] or functional data such as magnetoencephalography (MEG), transmagnetic stimulation (TMS), functional MRI (fMRI) [103], or actualization of preoperative imaging data with intraoperative CT, MRI, or ultrasound [236].

Hybrid transformation

Hybrid transformation combines surface-based and paired-point-based methods such as skin or bone surfaces with few implanted fiducials. Maurer [136] proposed the hybrid method to increase the surface-based matching accuracy by implantation of a single marker into the skull. The combination of both methods increases accuracy and reduces the amount of invasively implanted fiducials. Schmerber [194] published a hybrid registration method for ENT surgery matching anatomic skin and bone landmarks with skin and bone surfaces obtained from CT images. To diminish the invasiveness of the registration of bone surface points, instead of a needle he used a specific 1D ultrasound sensor as a transcutaneous telemeter. To calculate the six parameters (p) representing the appropriate rotation and translation, the boundary condition was expressed minimizing the least square energy function (E)p (equation 7).

$$ {\rm E}\left( {\rm p} \right) = {1 \over {\rm M}}\sum\limits_{{\rm i = 1}}^{\rm M} {{1 \over {\sigma _i^2 }}\left[ {{\rm dist}\left( {{\rm S}_{{\rm skin}} {\rm T}\left( {\rm p} \right){\rm P}_{\rm s}^{\rm i} } \right) - r} \right]^2 + {1 \over {\rm N}}\sum\limits_{{\rm j = 1}}^{\rm N} {{1 \over {\sigma _{\rm j}^2 }}\left[ {{\rm dist}\left( {{\rm S}_{{\rm bone}} {\rm T}\left( {\rm p} \right){\rm P}_{\rm p}^{\rm i} } \right)} \right]^2 = \min } } $$
(7)

Technical realizations of devices for navigation

The basic equipment common to all navigation devices is a pointer system and a computer work station. The computer consists of a digitizer for evaluation of the pointer position in space, a central processor for image processing, matrix calculation, and planning of the approach, and a high-resolution computer screen [11]. The pointer system, the referential frame, and, in armless systems, the detector unit are connected to the digitizer. The spatial information is transferred from the digitizer to the central processor, which additionally obtains CT/MR images via optical disc or direct net connection. For microscope-based systems [21, 58, 109, 173, 181], the information flow is also centripetal from the computer back to the ocular of the microscope to view the target point as an overlay, contours of the tumor, and orientation of the microscope axis to the target. Microscopes equipped with robotic properties also obtain instructions from the central processor about their motion in space [126]. (Fig. 1, Fig. 2). For microscope-based navigation systems, the focus distance corresponds to the mechanical pointer [54].

Fig. 1.
figure 1

Basic principle of an arm-based navigated microscope. The encoders in each joint measure the angulation. Together with the known geometry of the arms, these data are sufficient to determine its own position in space. The length of the focus serves as pointer tip. The spatial information of the focus tip is visible in the appropriate images, and the target point can be introduced as an overlay in the ocular of the microscope

Fig. 2.
figure 2

MKM-navigated microscope with both navigational and robotic capabilities

From a technical point of view, we can apply three independent classification schemata for the navigation systems:

  1. 1.

    Microscopes or mechanical pointers. In microscope-based systems, the whole microscope is the navigation device, and the focus distance of the microscope plays the part of the mechanical pointer.

  2. 2.

    Active robots or passive devices. Characteristic of the robots is their active computer-controlled movement. The passive devices are however only able to determine their position in space and are otherwise moved by a person.

  3. 3.

    Arm-based or armless pointers. The distinction arm-based or armless refers to the method of determining the arm's own position in space. For the former, only one mechanical realization is possible.

For armless systems, further subdivision depending on the physical waves used as emitting sources into sonic, infrared, magnetic, and visible light waves is necessary. The sonic and infrared systems have active emitter sources fixed on the pointer, the magnetic-based systems use the pointer as the receiver, and the last method with visible light makes use of passive reflection of light on the objects of reference. Thus, together with the arm-based systems, there are five possible ways of technically determining the position in space. The precise classification of a navigation device includes the specification of the system regarding all three levels, which are independent of each other whether the device is a microscope or not, whether it an active robot or a passive device, and whether it is an arm-based or armless instrument.

For multiple coordinate manipulators (MKM), for instance, the correct technical specification would be: an arm-based, active microscope. The first level distinguishing between microscopes and other systems has two possibilities, the second level classifying robotic and passive arms also has two possibilities, and the last level as mentioned above has five realizations. This makes 2×2×5 or 20 technical combinations possible. All of them have been realized in modern industrial production.

Arm-based calculation of the device position

The arm-based systems consist of arms and joints with six degrees of freedom, three of translation (to arrive at each point in the space) and three of rotation (to reach the target from different angles) (Fig. 3, Fig. 4). The joints are equipped with sensory encoders to measure the angulation. This information is evaluated by the digitizer to calculate the position of the tip in space at a rate of 30 positions/s to 300 positions/s. The position of the pointer tip in relation to the global coordinate system is expressed by the kinematic equation, which reflects the relationship between the position of the pointer tip to the relative position and orientation of the links. In matrix notation, displacement of the pointer tip from point A to point B in space is described as a successive multiplication of the six link-related matrices (A), each referring to displacement in the link-related position (equation 8).

$${\rm (T6) = (A1)}\;{\rm (A2)}\;{\rm (A3)}\;{\rm (A4)}\;{\rm (A5)}\;{\rm (A6)}$$
(8)
Fig. 3.
figure 3

Arm-based pointer device. The position of the pointer in space is calculated by the computer from the extent of angulation in each joint and from the length of the arms. Using skin markers for registration, the position of the pointer tip is visualized in the appropriate CT/MR image

Fig. 4.
figure 4

Intraoperative calibration of the operating arm system (OAS), an arm-based navigation device

In the case of robots with active movements, an optimal trajectory to the target point has to be calculated, taking into consideration not only the position but also the velocity and acceleration of the arm.

Armless calculation of the device position

The armless navigation systems determine their position in space by sonic, magnetic, or optic emitter sources which are fixed to a flexible pointer (Fig. 5, Fig. 6). The emitted waves are detected by microphones, magnetic sensors, or by means of an infrared camera array.

Fig. 5.
figure 5

Armless pointer device. The pointer is equipped with two or three sources emitting ultrasound, infrared, or magnetic waves. A camera or specific detector system outside the operating field detects the waves, and the computer calculates the position of the pointer tip. After registration, the tip of the pointer is visible in CT/MR images

Fig. 6.
figure 6

Intraoperative calibration with the OTS. The patient is fixed in the Mayfield clamp. The pointer equipped with LEDs contacts the markers. The position of the LEDs is registered by the camera array

Ultrasound spark emitter sources send waves at constant frequencies of about 25 kHz. Travel time between the emitters and the receivers is measured using a synchronization technique. Provided the velocity of the ultrasound in the air is constant, the distance between the emitter sources and the receivers can be computed knowing the geometry of the receivers and calculating the emitters as intersections of spheres centered at the receivers. The rate of measurements is in the range of up to 100 points/s.

Optic systems use a similar triangulation technique which allows one to determine a point in space from two cameras whose orientation and intercamera distance are known. The triangulation technique is named thus because the two cameras and the point in space to be determined form the vertices of a triangle. Charged-coupled two-camera devices (CCD) are able to calculate a point with a frequency of 300 points/s. To diminish irrelevant data in the input information, the camera system is pulse-synchronized with the emitter source's LEDs. Special position-sensitive camera devices (PSD) using cameras with analog PSD sensors are also in use. A third technique which minimizes data input involves at least three linear cameras. The spot on each 1D signal corresponds to a plane in three dimensions. The three cameras provide three planes in the space, and their intersection defines in space the coordinates of the desired point.

Some optically based devices use not emitting light sources for position measurement but only passive light reflected from the fiducials. The triangulation technique is again applied for determining the fiducial in space. The first passive visual light registration, so-called machine vision, was introduced by Heilbrun [94].

The magnetic navigation system consists of direct-current magnetic field generator, transmitter, and receivers detecting the magnetic field. Kelly [106] described a device using a transmitter with three orthogonal coils which generate a low frequency, 144-Hz magnetic field centered on the x, y, and z axes. A high-frequency field is not recommended because of nonlinear behavior of the parameters, with a subsequent decrease in spatial accuracy. As opposed to the optical systems, here the pointer is constructed as the receiver containing three coils measuring magnetic field strength. This information enables determination of the position of the stylus in relation to the emitting source and further to the global coordinate system.

A different solution was proposed by Hunnerup and Nielsen [103]. They used a transmitter fixed on a pointer and calculated its position by measuring the magnetic flux density at three fixed points and comparing them to a simulation model of the magnetic density flux. Because with one measurement of a sensor at a particular point, the localization of the emitting source had more than one solution, measurements with three sensors at different points were necessary. The intersection of curves created by the three sensors provided unequivocal results regarding the spatial position of the emitting source.

Accuracy of the navigation systems

First we have to specify and define the terms "precision," "bias," and "accuracy." Engineers use the term "precise" for a number of measurements whose standard deviation is small, although the mean (m) of these measurements is not necessarily close to the true value. The measurements are however unbiased if their mean value corresponds to the true value. Their standard deviation in this second case is not necessarily small, as is required for precision. A number of measurements are accurate only if they are precise and unbiased (Fig. 7).

Fig. 7.
figure 7

Examples of precision, bias, and accuracy

In robot technology, two measurements of accuracy are common: repeatability and absolute position accuracy. Repeatability means the ability of the robot to perform the same motion and finish working at the same point. It reflects the precision of the system but not its accuracy, because the actual point does not necessarily coincide with the calculated point. Absolute position accuracy is a measure of coincidence of the pointer tip in space with the calculated coordinates. Robots usually have higher repeatability than absolute position accuracy. For navigated applications, only the absolute position accuracy is relevant.

A comparison of publications regarding the accuracy of the navigated devices is hindered by the different methods and parameters measured and the differing statistical evaluation, and even the unit in millimeters can differ, referring either to linear range error relative to the x/y/z coordinate axis or the euclidean distance d in space. It would be desirable to have a simple measure of accuracy for each device expressed as mean error in millimeters with a standard deviation which is independent of space and time during surgery. This simple reduction of the accuracy problem is however impossible for navigation systems.

We have to distinguish between three different types of accuracy, which are arranged here according to increasing complexity and sources of possible error:

  1. 1.

    Technical accuracy

  2. 2.

    Registration accuracy

    1. A.

      Phantom/cadaver measurements

    2. B.

      Intraoperative measurements

  3. 3.

    Application accuracy

    1. A.

      Phantom/cadaver measurements

Technical accuracy indicates how reliably the navigation device can define its own position in space. The measurements are performed under standard laboratory conditions, and the error is valid in reference to a particular volume in space and the distance of the sensors from the emitting sources.

Registration accuracy relates to coordinate transformation. It depends on the technical accuracy of determining the fiducials by the navigation device and in the image space. We must separate measurements performed under laboratory conditions and those during the operation. The error is space- but not time-dependent.

Application accuracy reflects the overall error during the whole procedure. It includes technical accuracy, registration accuracy, and changes in the anatomic structures during the procedure. This error is time- and space-dependent. Again, we have to distinguish between phantom/cadaver studies and the intraoperative situation.

Technical accuracy

Technical accuracy is an interplay of two factors: the development of more accurate instruments and cost effectiveness, which is a compromise between the installation of more accurate instruments and the demand not to exceed particular financial limits. Important for establishing the required mechanical accuracy is the comparison of registration accuracy and application accuracy of the navigated devices and also the mechanical accuracy of the frame-based stereotactic systems.

The mechanical accuracy of the navigated devices should be higher than their registration accuracy and also be in the range of the frame-based stereotactic systems. It is meaningless to construct extraordinarily accurate mechanical, navigated devices which are extremely expensive and work with 100 times the mechanical accuracy of their subsequent registration and application.

The first arm-based navigation device of Watanabe in 1987 [230] had an error in absolute positioning accuracy of up to 5 mm. Roberts [180], describing the first microscope-based system in 1986, reported a mechanical error with a mean of 2 mm and range of 0.7–6.0 mm.

Along with arm-based systems, the main limitation in accuracy is the resolution of the encoders. Different technical solutions are available, such as potentiometers, optoelectric encoders, and resolvers. In general with digital optoelectric encoders, an accuracy of 0.022–1.3' can be guaranteed over the full extension of 360° [145]. Defining a small working space with fewer angles, the accuracy can be further improved. Inside a 1-m working space, the technical accuracy of arm-based pointer systems is guaranteed by the manufacturers to be better than 0.6 mm. Maciunas [127] found a positioning accuracy within 0.1 mm in a space of 40×40×40 cm. In the clinical study of Zinreich [252], the technical accuracy was lower, with a mean of 1.12±0.61 mm.

From a technical point of view, the main advantage of arm-based systems is their robustness and independence from external influences such as temperature, air disturbances, surrounding objects, etc. They are however more difficult to use during operation and require more space in the immediate operating field than armless pointers.

Armless optic systems using the triangulation technique and charge-coupled two-camera devices (CCD), with 512×512-pixel image matrix resolution inside 1 m of working space and with a source camera distance under 2 m, are able to determine a point with an accuracy of 0.5 mm [145]. Increasing the distance between emitters and camera can significantly reduce the accuracy [145]. Position-sensitive devices (PSD) using two cameras with analog PSD sensors work in the same accuracy range. With three high-resolution cameras with a 4096×4096 image matrix calculating the intersection of three planes, the accuracy can be improved to 0.2 mm. The present passive optic systems working with high resolution cameras and without emitting diodes are also in the range of 0.3 mm.

With sonic instruments using a 24-kHz sound burst at a distance of 1 m, a mean accuracy of 0.89±0.63 mm (measuring the euclidean distance) can be achieved along with linear accuracy of better than 0.4 mm [101]. This accuracy can deteriorate under influences of temperature, moisture, and pressure. With appropriate measurement of these parameters, this error can be compensated [167]. Air disturbances under operating conditions additionally influence accuracy in the range of SD±1 mm [167].

Concerning the accuracy of navigation devices measuring the position in space from magnetic wavelet-emitting sources placed outside the cranial tissue, the tip of the pointer has an accuracy of 1.7 mm [105]. Because the magnetic flux density decreases in space by the power of three, leading to higher error in positioning accuracy due to high input data error, it is important to reduce the distance between the emitting source and the receiver. With a strong magnet at a distance of 20–30 cm from the receiver, an accuracy of 0.76 mm is possible [103]. Additionally, the accuracy is sensitive to magnetic objects between the emitters and receivers and can deteriorate to 3–4 mm [105].

Using microscopes as navigation instruments, we have to consider the effect of the focus on accuracy, instead of the mechanical pointer. Westermann [232] showed that, with the optical overlay error into the ocular being 0.017–0.039 mm and the error due to zooming and focus setting being 0.04–0.11 mm, functional error was negligible in comparison to that from the simultaneously used optical tracking system, with a mean error of 0.3±0.15 mm and maximum of 0.76 mm [232]. For arm-based microscopes, the mechanical accuracy is a critical factor because of their large working space. To guarantee stability, these instruments require a special, massive construction. In general, for this type of robots the repeatability inside 1 m is 0.1 mm and the absolute positioning accuracy is 10 mm. By additional improvement and optimization of the parameters for each device individually, the positioning accuracy can be guaranteed to be within 0.5 mm.

With the present navigation systems, the contribution of technical error in the range of 0.1–0.6 mm is of subordinate value in comparison to the registration and application errors.

Registration accuracy

Registration accuracy measures the error due to matrix-based coordinate transformation between the device and image spaces. It answers the question of whether the cursor representing the tip of the pointer is correctly placed in the image space. Possible sources of limitation of accuracy in the device space using the paired-point matching method are:

  1. 1.

    Technical accuracy of the device

  2. 2.

    Type of markers (bone, skin, anatomic)

  3. 3.

    Form of the markers

  4. 4.

    Displacement of the skin markers

  5. 5.

    Determination of the center of the markers with the pointer

  6. 6.

    Determination of the center as focused by microscope

In the image space, the most important limitations are:

  1. 1.

    Resolution of the image given by pixel size in the x and y planes and slice thickness in the z direction

  2. 2.

    An accurate rectangular image data set

  3. 3.

    Movement of the patient in the gantry during data acquisition

  4. 4.

    Determination of the center of the fiducials

  5. 5.

    Type of imaging system—CT, MRI, digital subtraction angiography (DSA)

All errors cited above constitute the input data error for calculating the transformation matrix (T). The influence of this input data error between the device and image space on the calculation of the transformation matrix (T) and, in the next step, on the coordinates of the target point depends on two additional factors: geometric arrangement of the fiducials and position of the target in space (Fig. 8, Fig. 9).

Fig. 8.
figure 8

On the left side, the arrangement is not correct in the z direction, leading to error amplification. On the right side, the fiducials are in correct placement, with no colinearity in the z direction

Fig. 9.
figure 9

As a rough estimate, the target area should be inside the polyeder, whose vertices are the fiducials

The error amplification due to input data error depends strongly on the structure of the matrices X and X* (equation 1), which reflect the geometric arrangement of the fiducials on the head of the patient. In case the fiducials are colinear in the x, y, or z direction, the input data error can increase in the range of centimeters. Therefore it is important to avoid colinearity in any coordinate by placing one fiducial apart from all other markers (Fig. 8). The dependency of input data error on the position of the target point in space and on the arrangement of the fiducials was proven by Darabi [34] in a computer simulation. As a rule of thumb, the target area should be inside a polyeder whose vertices are the fiducials (Fig. 9). Under this arrangement, the error amplification is negligible.

All navigation systems based on paired-point transformation have a check for estimating the registration error. It is the root mean square (RMS) of the distance differences of the fiducials between the image and the device space (equation 9):

$$ \sqrt {{{\sum {\rm (d}^{\rm *} {\rm N - dN)}^{\rm 2} } \over {\rm N}}} $$
(9)

dN is the distance between two markers in the image space, and d*N is the distance between the same markers in the navigation space

The use of skin markers is most common in clinical application. The mean intraoperative RMS using CT with 2–3-mm slice thickness is in the range of 1.5–4 mm [74, 106, 146, 175, 176, 177, 213, 238]. With screws drilled into the bone, the accuracy can be improved because displacement of the markers on the skin during head fixation can be avoided. Brinker [19] measured an RMS of 0.23±0.03 mm with 1-mm CT slice thickness and screws drilled into the skull under experimental laboratory conditions. However, fixation into the skull is an invasive method. The possible displacements of the skin markers can be overcome by performing the CT intraoperatively after fixating the patient with the clamp [135]. As an alternative to skin markers, Maurer [137] proposed a hybrid method combining one screw with a surface-based transformation. With head fixation, the registration error was 0.7 mm and, without fixation, it was up to 1.4 mm.

Registration using anatomic landmarks seems to have the greatest registration error probably because of the difficulty to touch accurately the same corresponding points in the image and on the patient [95]. Villalobos [220] found the registration error intraoperatively (measured as the difference in the CT image between the marker and the cursor in the image space during pointing with the tip on the real marker) to be twice the error using self-adhesive markers: 3.4±0.4 mm compared to 1.6±0.1 mm. In Brinker's cadaver study [19], the relative difference between RMS using implanted screws and using anatomic landmarks was even higher: 0.23±0.03 mm vs 0.67±0.2 mm.

As an alternative to adhesive skin markers, subcutaneously implantable fiducials were proposed with the possibility of repeated use and which can be detected by ultrasound [124, 198]. Under experimental conditions, they work with an accuracy of 0.5±0.1 mm.

Surface-based registration depends on the technical accuracy of the contour scanning device, the method by which the 3D surface is created, the imaging technique with automatic feature-detecting algorithms, and the shape and geometry of the objects to be matched. The more anatomic structures with complex geometry such as facial structures are included, the better the matching. An error check is more difficult to obtain for surface-based registration than for the paired-point method. Measures which characterize only the degree of correspondence between both objects are insufficient. If additional anatomic landmarks or fiducials are used, RMS can be established. Bucholz [22] estimated the accuracy using a laser contour device with an optical tracking system in the range of 1.7 mm. Under experimental conditions, the RMS in his measurements was 0.5 mm. Maurer [137] found better results with paired-point registration than with surface-based registration (0.76 mm and 1.19 mm, respectively). Matching 200 randomly taken skin points and 17 bone points with 3D surfaces, Schmerber [194] obtained from CT images under experimental conditions a reported mean position error of 1.35 mm and an RMS of 0.71 mm.

Taking into consideration different markers and starting with the greatest error, the order of succession of the registration error seems to be: anatomic superficial landmarks, adhesive skin fiducials, surface matching, hybrid method with one screw, and screws in the skull. Exactly the opposite direction of this sequence relates to the invasiveness and necessary expenditure.

The imaging systems and the accuracy of determining the center of the fiducials in the image space contribute equally to the accuracy of registration [203]. It is obvious that an exact 3D data set of the images is a prerequisite for accurate registration. This means in particular that the patient may not move during image data acquisition. Therefore, in small children or uncooperative patients, imaging must be performed under general anesthesia.

Registration accuracy also depends on the imaging system, whether CT, MRI, or DAS [30]. It is well known that MR images can have spatial distortion due to nonhomogeneity of the magnetic field. These nonlinear effects have consequences on the accuracy of the determination of fiducials [36]. Maciunas [129] developed a correction algorithm which, applied to MRI, significantly improved the euclidean error in target registration between CT and MRI from 3.833±0.992 mm to 1.986±0.605 mm. Maurer [137] showed that the registration error could be significantly improved for paired-point registration and surface registration if the nonlinearity in the MR images is corrected. Thus, paired-point registration improved from 1.15 mm to 0.76 mm and the mean surface-based registration improved from 2.2 mm to 1.19 mm.

Evaluating application accuracy, Golfinos [71] found slightly better results with MRI than with CT. This surprising result could however be explained by the different slice thicknesses he used (3 mm for CT and 2 mm for MRI). Image resolution also plays an important part in registration accuracy. The resolution is determined in the x and y axes by pixel size, and the z coordinate is determined by slice thickness. Modern CT and MRI devices use a 512×512-pixel matrix, which corresponds to a resolution of about 0.5 mm.

Slice thickness comprises the crucial limitation in resolution. Under best conditions, the slice thickness is 1 mm. Using this optimal imaging resolution and screw markers, Brinker [19] was able to demonstrate the high accuracy in two cadaver procedures to the skull base, with mean RMS of 0.23±0.03 mm and 0.19±0.27 mm. Consistent with Brinker's study, Doward [41] found a significant improvement in accuracy using 2-mm rather than 3-mm slice thickness. He also confirmed the higher accuracy of CT over MRI. At the beginning of navigation, Watanabe [230] described in his original paper a navigated procedure using 10-mm slice thickness in consideration of stability of the images. With higher slice thickness, the images jumped on the screen to and fro. Most authors using skin fiducial markers use 2–3-mm slice thickness for cranial applications. At the skull base and when higher accuracy is necessary, 1-mm is preferable.

The determination of the center of the fiducials influences the accuracy of the registration. Usually the procedure is performed by free-hand technique. Because of the geometric form of the fiducials, it is more difficult to define the center than in frame-based stereotaxy using rods. However, correcting the position of the center in coronal and sagittal images helps to improve accuracy. Another possibility for improving accuracy consists in an algorithm which detects the fiducials automatically [229].

Finally, the learning effect also has to be considered. For microscope-based navigation, Rössler [176] observed an improvement in accuracy after the first 25 cases from a mean of 4.8±3.36 mm to 2.2±0.86 mm. The influence of the navigation devices on registration accuracy is not significant as long as the error due to the registration is much higher than the mechanical accuracy of the apparatus. With adhesive skin markers, Wirtz [238] obtained similar RMS results with an arm-based system, optical pointer, and multiple coordinate manipulator (MKM) microscope of 2.9±1.2 mm, 3.3±0.9 mm, and 3.1±1.0 mm, respectively. Registration accuracy is crucial in the whole navigation procedure. Under optimal conditions with appropriate markers and image resolution, the RMS can be as low as 0.2 mm. Using skin markers and 2–3-mm slice thickness, a mean RMS in the range of 1.5±3.0 mm can be expected under operating conditions.

The RMS is a reliable measure of registration accuracy only if the geometric arrangement of the fiducials is not colinear. Amplification of the input data error is not expressed by the RMS. To estimate the spatial relationship of the error, the best method consists in using additional markers or anatomical structures distributed equally over the head of the patient and checking their positions with the pointer after registration. If the fiducials are not arranged colinearly, the effect of error amplification can be ignored.

Application accuracy

Application accuracy is a measure of the targeting error during surgery. It answers the question of how reliably the pointer tip in physical space corresponds to its anatomic position in the CT/MR image during the procedure. In addition, the specific application error reflects the loss of correspondence between the image-based anatomic determination before surgery and the intraoperative anatomy encountered in actual physical space. The error is consequently not only space- but also time-dependent. Limitations in application accuracy are due to:

  1. 1.

    Technical inaccuracy

  2. 2.

    Registration inaccuracy

  3. 3.

    CSF leakage

  4. 4.

    Removal of tumor

  5. 5.

    Opening of cysts and ventricles

  6. 6.

    Tissue distraction by spatula

  7. 7.

    Displacement of the head in the clamp

  8. 8.

    Displacement of the dynamic reference frame (DRF)

Application accuracy is the most unpredictable factor of the whole navigated procedure because it depends highly on the specific intraoperative situation, in particular CSF leakage, duration of the operation, position of the head, and type, size, and location of the lesion. Therefore, when analyzing application accuracy, it is important to distinguish between cadaver studies and intraoperative measurements [239]. Using the umbo, asterion, and tympanic membrane as anatomic markers, Vrioni [221] found an application errors at the skull base between 0.91 mm and 2.44 mm. The results of studies on intraoperative application accuracy are also difficult to compare not only because they refer to different statistical parameters but especially because of their different measures for accuracy estimation.

Using skin markers for error estimation, Barnett [8] and Guthrie [81] found a mean error during surgery of 1.5±0.7 mm with a sonic pointer system. With the same method, Guthrie obtained a mean error of 3.3 mm and a range of 0.6–7.6 mm with an arm-based pointer system. However, skin markers reflect registration accuracy more than intraoperative application accuracy.

Golfinos [71] used bony landmarks intraoperatively as reference points to check the application accuracy. He used four measurements during the procedure. Immediately after calibration, the accuracy was better than 2 mm in 92% of MRI cases and 82% of CT cases. The error degraded during operation, and after the third evaluation it was less than 2 mm in 77% of the MRI cases and 62% of the CT cases. However, using bone landmarks checks the accuracy of whole head movements during surgery but does not relate to additional relative displacement of the brain structures in relation to the skull. Therefore, the application error of the intracerebral structures could be expected to be even higher. It would be more conclusive to use intracranial structures such as vessels, but the exact assignment between these structures and their correspondence in the image is difficult.

Thomas [213] measured the offset of the localized bone surface, cortex surface, superficial lesion margin, and tumor bed from the appropriate image positions, his mean results being 2.2±1.5 mm, 5.8±4.7 mm, 0.9±1.4 mm, and 2.9 ±4.5 mm, respectively. The high deviation at the cortex level is due to brain shift. However, more important is the finding that the error was very low at the level of the tumor margin. The same group evaluated application error in cases of stereotactic biopsy using postoperative MR images and image fusion. Dorward [41] found a mean linear error of 2.3±1.9 mm and a mean euclidean error of 4.8±2.0 mm.

For small lesions, the displacement of anatomic structures during operation is mainly due to CSF leakage. Immediately after opening the dura, application accuracy is in the range of registration accuracy and decreases with the course of time. How much CSF leaks depends on the position of the patient's head and the location of the craniotomy.

Dorward [40] measured a mean shift of the cortex surface of 4.6 mm after opening the dura, 5.1 mm after reaching the tumor margin, and 6.7 mm of the cortex at the end of the resection. In 28 investigated cases, Roberts [180] found an average displacement of 1 cm, with a dominant direction component associated with gravity. In a preliminary study, Mauerer [139] observed only little deformation of the midline and the hemisphere contralateral to the operation but shifting within 1 cm of the tumor and structures gravitationally above the surgical side. In our cases, displacement depended mainly on CSF leakage and the volume of the removed tumor.

Miga [142] developed a model update based on gravity-induced brain deformation. The error from brain shift could be improved from an average of 5.7 mm to 1.2 mm. The same group also published a mathematical method to update the CT/MR images by sparse data from digital cameras or ultrasound [66, 90, 174]. An alternative possibility is to perform intraoperative CT or MR images and replace or match them into the previous data set [114, 166]. Using intraoperative MR images, Wirtz [236, 237] actualized the image data and used them for navigation during the entire surgery including whole tumor resection. Another possibility is to use the open MRI or CT intraoperatively instead of navigation [4, 110, 214].

Summary regarding navigated accuracy

In conclusion, the error of technical device accuracy is low, in the range of 0.1–0.6 mm. The registration error is higher, depending on image resolution and the type of fiducials, being typically in the range of 0.2–3.0 mm. The application accuracy is the most unpredictable, depending on time, CSF leakage, and localization of the operation, being in the range of 0.6–10.0 mm.

Clinical applications

Tumor localization

Localization of small intracranial tumors is presently the most frequent application of navigated technology in neurosurgery for adults [8, 9, 35, 42, 44, 46, 52, 71, 74, 79, 81, 82, 83, 84, 88, 121, 127, 128, 140, 146, 148, 167, 168, 175, 176, 177, 178, 179, 190, 195, 199, 201, 205, 206, 212, 213, 218, 222, 223, 224, 225, 237, 238] and children [45, 85, 226]. It is used to decide the type of craniotomy and find the tumor atraumatically after dura opening. Typically on the day of the operation or 1 day in advance, CT/MR images with markers are performed without gantry tilt to obtain a Cartesian data set. The images are transferred to the navigation work station, and the patient is intubated and fixed with the head clamp. Then the registration is performed, taking around 10–15 min (Fig. 10). After calibration, the pointer is moved slowly over the patient's scalp and the corresponding cursor of the tip is observed moving in real time in the appropriate axial, coronal, and sagittal images. The best position of the craniotomy is decided and, if necessary, additional superficial anatomic structures such as sagittal and transverse sinus can be defined and drawn on the scalp. With an extension function or by sliding the LED holder more closely to the pointer tip, the trajectory to the lesion can be visualized. After craniotomy, the pointer is used to decide the corticotomy and the direction to reach the lesion. For very small or deep-seated tumors with long access to the tumor through the gray and white matter, a catheter can also be implanted through the pointer, which itself is fixed in space by two Laila arms or a special instrument holder [41]. The most frequent clinical indications for navigated localizations are small meningeomas [10], gliomas [130], metastases [83] (Fig. 11), and cavernous hemangiomas [216] (Fig. 12).

Fig. 10.
figure 10

Intraoperative registration with the OAS. Five skin markers are visible in the frontotemporal area

Fig. 11. Fig. 12.
figure 11

Computed tomographic image with a metastasis in the right occipital region which was localized by navigation

Computed tomography of a hemangious cavernoma in the right frontal lobe close to the motor strip

The delineation of tumor borders is in general no more accurate if brain shift must be assumed. Nevertheless, the navigation rules out overlooking big tumor remnants. To improve intraoperative accuracy, additional intraoperative image actualization is necessary. The navigation can also be helpful in delineating superficial low-grade astrocytomas at the beginning of the operation while investigating the cortical surface (Fig. 13) and also in subtemporal transtentorial approaches for resecting the tentorium at the proper place. Using a navigated microscope, the spatial information of the tumor and its borders are directly overlaid into the ocular.

Fig. 13.
figure 13

Computed tomography of a low-grade astrocytoma in the right frontal region. Circle shows the actual position of the pointer tip

Skull-base surgery

From a technical point of view, skull-base surgery is an ideal application of navigation because no intraoperative shifting is to be expected. Ear, nose, and throat surgeons therefore use navigation as an orientation aid for nasal, paranasal, and otologic interventions [26, 89, 91, 92, 204, 232, 233, 234]. In cases of tumor surgery, navigation is usually not necessary for finding lesions, because there are enough anatomic landmarks for orientation, but it helps estimate the position of important anatomic structures such as the carotid artery or cranial nerves, particularly if they are inside the tumor, as in medial peritonal wing meningeomas or transnasal pituitary surgery [50, 191]. The navigation can also help make approaches through the petrosal bone more safe, sparing structures of the inner ear in orofacial approaches to the C2 vertebra [47], and in tumor recurrences at the skull base with a changed anatomic situation. Navigation can be equally applied to drill osseous meningeomas more safely and radically from the skull base [200].

Biopsies

Biopsies can be performed with arm-based and, more easily, with armless pointer systems [20]. For adaptation, the pointer has to be stabilized by a self-retaining retractor attached to the reference arc placed close to the operating field. For this purpose, we use two Laila arms which fix the LED pointer in space (Fig. 14). Using the pointer from the optical tracking system (OTS), the LED holder can be adapted for biopsies by removing the metallic part of the pointer. After registration, the LED holder of the pointer alone is moved in space under optical control on the triplanar reformatted images. Then the pointer is fixed and the cannula introduced through the reducing tubing (the length being identical to that of the removed, metallic pointer). Another adaptation is to use not the pointer but a special probe holder equipped for this purpose with LEDs. Germano [62] reported on 34 biopsies performed successfully with this technique on tumors with diameters of 3.5 cm. Dorward [41] also adapted an optical navigation pointer for biopsies. One advantage over frame-based calculation is the fast finding of a trajectory under visual control and the possibility of radiologic data acquisition before general anesthesia.

Fig. 14.
figure 14

Navigated stereotactic biopsy with the OTS. The probe holder with the LEDs is fixed by two Leila arms. The Nashold biopsy cannula is introduced through the pointer to the target point

Endoscopy

Endoscopic procedures are mostly performed by free-hand technique [57]. Navigation is however useful in endoscopy if the trajectory has to be planned carefully, for instance in cases of narrow access through the foramen of Monro into the third ventricle to prevent damage to the fornix [77] (Fig. 15, Fig. 16). Either the navigation is performed with the pointer to mark the burr hole and decide the trajectory, or the endoscope itself is used as a pointer navigation system (Fig. 15). In the latter case, an LED array is attached to the endoscope, which is calibrated according to the length of the shaft. Another possibility is to replace the pointer tip by the endoscope and fix it in the LED holder at the appropriate length [39, 99, 169]. Stable fixation by retractors is crucial, because the rigid endoscope with cameras attached is top-heavy and may move, depending on the pivot of the fixation. The application accuracy is low only as long as the ventricles or the cysts are not open. This is also the case when planning the trajectory at the beginning of the procedure.

Fig. 15.
figure 15

Axial coronal and sagittal CT images of a patient with hydrocephalus during an endoscopically navigated operation. The trajectory through the lateral ventricle and foramen of Monro is visible in all three planes

Fig. 16.
figure 16

Endoscopic image of the foramen of Monro with plexus chorioideus and the fornix as the border of the foramen. The view into the third ventricle makes it possible to identify the corpora mammilaria and the adhaesio interthalamica

Inside the third ventricle, Muacevic used navigation in ventriculostomies to fit optimally through the narrow passage of the foramen of Monro [147]. Schröder and Gaab [199] also used navigation in aqueductoplasty to define the best approach to the aqueduct. Navigation is also helpful in cases of biopsies or cysts in the wall of the third ventricle because it gives information about the structures behind the surface of the ventricle wall which are visible by the endoscope. In this way, the best place can be selected for removal of the spacements or perforation point [73].

Above the third ventricle, fenestration of a cavum vergae can be supported by stereotactic methods to plan a trajectory sparing eloquent cortical areas. Also, in cases of very large cysts around the lateral ventricles, the perforation into the lateral ventricle can be cumbersome if the membrane is thick and not translucent. Orientation inside the cyst is difficult if anatomic landmarks are missing. In these cases, navigation helps to find the point of perforation into the lateral ventricle [183]. The endoscopic image itself can be further digitized and mapped into the image space of the navigation system [112]. The spatial assignment allows performance of distance measurements and, theoretically, also coagulation in the presence of bleeding. Zamorano used endoscopic navigation in over 150 tumor cases including biopsies, colloid cyst removals, and tumor extirpations [249].

Procedures close to functionally important cortical areas

The idea of this application is to identify and delineate special functional cortical areas and map this information into the CT/MR images used for navigation [87]. This seems particularly useful for extirpation of lesions close to eloquent cortical or subcortical areas [46].

These cortical areas can be established from conventional CT or MRI using neuroradiologic criteria such as the topography of the sulci, the hump in the gyrus precentralis which represents the hand, the volume of the gyrus precentralis as compared to the thinner gyrus postcentralis, and others. However, in cases of brain edema or cortical displacement by the tumor, localization can be difficult. The classic method of identifying these regions intraoperatively was cortical stimulation. Cushing [32] described in 1909 "a faradic stimulation of the postcentral gyrus in conscious patients." Penfield later used this method in epilepsy surgery [153, 154]. Ojemann [160] applied the stimulation in particular close to speech-relevant areas during ablative epilepsy surgery. Cortical stimulation in the motor strip can be performed under general anesthesia; in speech-relevant areas, the patient has to be awake [215].

Recently, alternative methods were introduced by matching functional information into CT/MR images used for navigation. For the functional mapping, transcranial magnetic stimulation [53, 78], magnetic source imaging [60], MEG [192], and fMRI [18, 31, 59, 78, 102, 115, 116, 132, 187, 202] were used.

Gugino [78] described a transcranial magnetic stimulation device equipped with LEDs which maps the functional information after surface-based registration into appropriate MR images. Using several activation points, the broca region and the gyrus precentralis can be identified. A different approach consists in matching the fMRI images into the CT/MR images performed for navigation. The capacity of MRI to visualize functional activity of the brain was observed first by Belliveau in 1991 [12]. He demonstrated signal changes in the human visual cortex immediately after bolus injection of gadolinium. In 1992, Ogawa [159] described a second imaging technique without contrast medium which detects changes of deoxyhemoglobin as an endogenous contrast medium during cortical activation. Blood flow and the concentration of deoxyhemoglobin in the cerebral capillaries influence the intensity of the MRI signal and reflect indirectly the activity of particular brain areas (the BOLD effect). The main advantage of fMRI over transcranial magnetic stimulation is its capacity to investigate very different brain areas and cognitive functions. Technical problems of fMRI in combination with navigation are:

  1. 1.

    MRI echo planar imaging (EPI) sequences are much more spatially distorted than the common T1 and T2 sequences

  2. 2.

    Movement artefacts must be excluded

  3. 3.

    Only very cooperative patients can be investigated

  4. 4.

    Evaluation of the images is very time-consuming

  5. 5.

    Regarding specificity, very sophisticated psychologic paradigms must be developed

Functional neurosurgery

Intracranial neurosurgical interventions in the deep brain structures regarding pain, extrapyramidal movement disorders, and particularly epilepsy are the classic indications for applying frame-based technique. Nevertheless, navigation can be successfully used in epilepsy surgery for localization and introduction of subdural strip and grid electrodes or for implanting deep-brain electrodes in the hippocampus [150]. Equally interesting is the navigated orientation during ablative surgery in cases of epilepsy, such as with hippocampectomy, to control the resection size [161, 162, 210, 219, 240].

Insertion of catheters

Laborde [118] reported the drainage of abscesses guided by navigation. It is also possible to use these catheters for local antibiotic therapy. Rohde [184] introduced a catheter by navigation into intracranial bleedings for evacuation and lysis therapy. Other authors used the navigated placement of catheters in connection with interstitial radiation therapy [80].

Spinal surgery

Compared to cranial applications, navigation in spinal surgery has some peculiarities regarding anatomy. The spinal cord is much more flexible and therefore dependent on the position of the patient. Skin markers are thus not applicable. Registration must be performed after preparation of the vertebrae on their characteristic anatomic landmarks either with a paired-point technique or in combination with surface matching (Fig. 17, Fig. 18). The dynamic reference frame is fixed on a processus spinosus inside the operating field to register any displacement close to the working space. Tools such as screwdrivers have to be equipped with LED sources. Recently, fluoroscopy could also be integrated into the navigation space [86].

Fig. 17.
figure 17

Intraoperative registration in a spinal procedure using several anatomic bony landmarks

Fig. 18.
figure 18

Intraoperative navigated implantation of screws during a spinal procedure. The screwdriver is equipped with LEDs

Accuracy in spinal applications improved significantly using special spinal navigation technique. In cadaver studies, Glossop [69] found an application accuracy of 2.11±1.49 mm with an angular deviation of 5.11±4.39° using paired-point registration and 2.57±1.34 mm with angular deviation of 4.23±1.84° in combination with surface registration. Merloz [141] measured application accuracy in vivo within 1 mm.

The main clinical indication for computer-aided navigation in spinal surgery is the transpedicular insertion of screws in the thoracolumbar region [16, 17, 25, 68, 141, 149]. Carl [24] showed in cadaver studies that insertion of screws by navigation is more accurate than by fluoroscopy alone. This result was confirmed in vivo by Merloz [141], who observed less than optimal placement of the pedicle screws with navigation in 9% of cases and without navigation in 44%. Navigation was also used in anterior cervical corpectomy [2] to provide better orientation laterally towards the vertebral artery.

Social and economic aspects

Navigation devices are expensive acquisitions for hospitals, with prices in the range of 75,000–250,000 Euros. Therefore it is important to demonstrate their favorable effect on patient outcome and lower operation and hospitalization time. Economic parameters have been investigated concerning duration of operation, hospitalization time, and cost of treatment. Regarding operation time, Alberti [3] found no benefit in navigation for tumor extirpation but significant time savings for stereotactic biopsies. Bingman [14] and Germano [63] showed significantly shorter hospitalization time: for 170 intracranially operated patients with tumor removal, he reported a decrease in stay from 10.8±1.3 days to 7.5±1 days using navigation surgery. For operation of meningeomas, Paleologos [151] found that navigation use reduced patient stay in the hospital from 13.5 days to 8.5 days. Severe complications dropped from 14% to 6% with navigation, and costs per patient were approximately 20% less in this group. Warnick [227] investigated patients with metastasis, including 19 operated on with OTS navigation and 11 in the conventional way. He found no cost reduction by the application of navigation, only reductions in operation time and hospital stay.

Discussion

Stereotactic and microsurgical techniques are accepted methods in contemporary neurosurgery. By means of stereotaxy, straight trajectories and target points are calculated with high accuracy in advance to reach any intracranial point or volume [75, 76]. Microsurgical technique, on the other hand, takes advantage of the imaging modalities to select carefully the approach [154, 155, 156, 157, 189, 241, 242, 243, 244, 245, 246] and reach and extirpate lesions by atraumatic treatment of the tissue with appropriate instruments under microscopic control. Intraoperative orientation in the latter method is based on the surgeon's knowledge of topographic anatomy.

The indications for stereotactic and microsurgical technique have been clearly delineated. When precise determination of a target point was necessary, the stereotactic method was applied, for instance in cases of functional surgery in the basal ganglia and thalamus, in biopsy cases, in interstitial radiation therapy, and for the localization of small tumors. On the other hand, whenever orientation using anatomic structures was sufficient, the microsurgical method was applied exclusively.

With the development of computer-based, navigated surgery, the equilibrium of both techniques seemed to change by the appearance of a third technique somewhere between stereotaxy and microsurgery, claiming to be an accurate stereotactic method and simultaneously a microsurgical technique. This evoked reactions on both sides, fearing loss of competence. The objections of stereotactic neurosurgeons against navigation could be summarized as: Is the expensive navigated method necessary in the presence of much more accurate and consequently better, classic, frame-based stereotaxy?

Microneurosurgeons formulated their hesitations about navigated surgery claiming a future much too dependent on technology. "What happens if the navigation device does not work? Neurosurgeons will lose their skills in finding lesions if they use navigation exclusively for localization." Didactic argument pointed in the same direction: "Young neurosurgeons will not learn classic anatomic localization and then will be unable to operate without the navigation device." Others claim that navigation is only a plaything for neurosurgeons charmed by the magic of high tech. These neurosurgeons argue that intracranial lesions can be found equally successfully without navigation or in difficult cases by means of ultrasound or frame-based stereotaxy.

We would like to deal first with the argument of stereotactically oriented neurosurgeons. It is an error to presume that accuracy is the only criterion of whether a method suits the neurosurgical concept. This assumption is true for stereotactic methodology but not necessarily for microsurgical needs. Microsurgeons require an orientation aid to find lesions atraumatically, but this does not necessarily require subpixel accuracy. It also suits visually oriented microsurgeons to see a target interactively directly in the image space or in the microscope for decisions on craniotomy or direction of the approach and for spatial orientation of his instruments in relation to anatomic structures. They can use the navigation device repeatedly without time-consuming calculation of the coordinates. Furthermore, they are concerned that stereotactic instruments inside the operating field do not restrict their space and view. These qualities are fulfilled by navigated devices and compensate for accuracy loss, which is indeed greater than with frame-based systems.

The tolerable error of the navigation devices depends on the particular application and must be decided during the intervention. An error in the white matter by the navigation device even in the range of 3 mm or 4 mm is still lower than when relying only on neurosurgical knowledge (Fig. 19). Too, if the neurosurgeon is able to calculate the localization and the approach to a small lesion accurately, he feels more confident. The corticotomy is associated with less stress, particularly in functionally important areas such as the central region, if after opening the dura without superficial visible pathology he can confidently approach the right sulcus or make a corticotomy in the correct place just above the lesion by using a navigation device.

Fig. 19.
figure 19

Computer simulations with different arrangement of the fiducials. The closed lines are places with identical standard deviation. Top row The fiducials were arranged optimally around the whole head. The error is very low and does not increase inside the head. Lower row The fiducials were placed only in the frontal area. The amplification of the input data error is steep and, particularly in the occipital region, the error is not tolerable

We think Ostertag [163] is right in his reservation toward unreflected enthusiasm for high-tech instruments. The new techniques and instruments have to be at least as successful in mastering practical neurosurgical problems as already established methods. His main argument against navigated surgery deserves additional attention. He aims to show different accuracy for each method and states:

"The navigation method is prone to significant errors as to the intraoperative localization based upon preoperative 3D images. The maximum error can be up to 2.6 cm depending on the extent of the so-called brain shift. In comparison, classic frame-based stereotaxy has a mean error of ±1 mm and remains a gold standard for the exact three-dimensional localization of a given lesion."

In his formulation, Ostertag makes a logical error by comparing the accuracy of both systems under different application modalities. He compares the accuracy of a burr hole procedure with an open craniotomy instead of the accuracy of two stereotactic techniques. It would be more correct to compare the accuracy of both systems for the same type of operation, as in the introduction of a catheter to a target point, for example. Further, if the brain shift, as he states, is the reason for the intraoperative error of navigation, then this error is independent of the stereotactic instrument used and, for frame-based applications during open craniotomy, would also be in the range of 2.6 cm.

Nevertheless, comparing the registration accuracy of frame-based and navigational systems, frame-based calculations are in the range of the image resolution [61, 75, 76, 131], and frameless calculation depends on the method of registration. Using skin markers, the accuracy is in general no better than 2 mm. Under best possible conditions, applying the hybrid method or screws in the skull, the accuracy is also in the range of that with frame-based systems.

The doubts of microneurosurgeons about navigated instruments are in my opinion groundless. The argument of excessive dependence on the navigation and therefore being unable to operate in case the navigation fails is invalid. Navigation instruments are merely tools to improve and simplify some parts of the surgical procedure, not to replace them. Neurosurgeons always have to be able to perform operations without navigation devices. From the didactic point of view, navigation does not interfere negatively with neurosurgical training. On the contrary, we believe that the immediate feedback of spatial information between the pointer tip and the corresponding images much improves the learning effect. Additionally, the stability and robustness of the navigation software is so good that technical failures are extremely rare and usually occur only at the beginning, during the training period.

Another question concerns whether navigation instruments are necessary. Of course neurosurgery was also performed successfully before navigation techniques were developed. This is however not the point. The question is: Can microsurgical technique be improved by navigational methods? Someone using a standard big frontal craniotomy for a small meningeoma does not require navigation because the tumor will be inside his operating field anyway. The big standard approaches go back to the time of angiographic localization, where lesions could sometimes be related only to the frontal, temporal, or occipital lobe, depending on the characteristic displacement of the arteries in angiography. However, by improved CT/MR imaging, much more precise anatomic localization is possible, and navigation techniques help in the decision on atraumatic openings and approaches.

If we concede that working with navigation is better than working without it, the question regarding necessity remains. A possible alternative is to use frame-based localization in difficult situations instead of expensive navigated devices. Independently of the above arguments about the better ergonomy of navigated instruments, frame-based procedures put restraints on possible approaches. From the mechanical point of view, it is nearly impossible to use subfrontal, subtemporal, suboccipital, or interhemispheric approaches with the frame. Nevertheless, navigation does not completely displace frame-based stereotaxy. The main indications for classic stereotaxy remain very small targets approached along straight trajectories.

It is sometimes difficult to convince people who do not work with navigation instruments about the benefit of this method. They need their own experience during intraoperative applications. However, the facts that discussions on this topic have nearly ceased (similar to the application of microscopy 30 years ago) and most centers use navigation speak in favor of this technology.

The last objection regarding excessive technology in the operating field making it necessary to pay more attention to the instruments than to the surgery is no longer valid. In the meantime, the software of most navigated instruments is simple, the registration itself is not time-consuming, and at least in our department there is no longer the feeling that using navigation comprises a special type of operation. In fact, it is commonly and routinely used during neurosurgical procedures.

It is difficult to give any prognosis for the development and role of navigated surgery in the future. This is because technology in the computer branch is changing so quickly and new developments not based on the standard coordinate transformation can alter the present situation completely. Unconventional navigation application such as holistic superpositions [111] have been proposed. Manwaring [133] reported on a magnetic emitting source fixed directly to the patient's skull, producing a nonlinear magnetic field through the brain. Flexible catheters with a magnetic tip could be introduced along these nonlinear magnetic trajectories to the target point.

Navigational instruments are presently undergoing a process of evolution and natural selection, with many species arising due to different technical realizations. The best will survive. The criteria for selection are simplicity, accuracy of the neurosurgical procedure, and cost effectiveness. Although we cannot predict future developments, we can already state that navigation reduces risks and hazards during operation and hinders neurosurgical interventions from resembling Odysseus' adventurous wandering on the long journey home from Troy to Ithaca.