Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

Imaging techniques have been extensively used for diagnosis in many medical specialties, whose main purpose is to create visual representations of the interior of human body. Internal anatomic structures of the body can be revealed, providing physicians with valuable clinical information about patient’s conditions. Technological advances in medical imaging have aided the comprehension of structures and functions of organs, as well as the detection of abnormalities.

Several medical imaging modalities (Macovski 1983) are currently available for medical diagnosis, such as computed tomography, magnetic resonance imaging, ultrasonography, mammography, photon emission tomography, among others.

Among the various tools available for medical applications, InVesalius (Amorim et al. 2015) is an open-source framework with several functionalities for processing, analysis, visualization, and 3D printing support for medical images.

This chapter focuses on relevant image processing and computer graphics techniques for the generation of three-dimensional (3D) models for visualization and bioprinting.

In the following sections, we will describe some relevant aspects of medical imaging, preprocessing, segmentation, registration, volume rendering, 3D printing, and biofabrication.

2 Medical Image Acquisition

This section briefly describes (i) some medical imaging modalities used for diagnostic purposes and (ii) DICOM standard.

2.1 Imaging Techniques

Medical imaging is one of the most common noninvasive techniques for acquiring and interpreting images for medical diagnosis purpose. Images of internal structures of the body can be generated and analyzed to provide a visual representation of organs and tissues.

Three-dimensional visualization of anatomic structures allow for more accurate surgical planning, reducing interventions and patient’s risk.

Several imaging modalities have been developed to assist clinicians in diagnosis and treatment of patients, making beneficial impacts on healthcare. Imaging technologies currently available include computed tomography, magnetic resonance imaging, ultrasound, positron emission tomography, among others.

Computed tomography (CT) uses X-ray images acquired from different orientation angles to generate cross-sectional images around specific areas of a patient. The attenuation coefficients measured by the CT equipment are organized into a two-dimensional sinogram (Leondes 2005). Image processing and computer graphics techniques are used to produce a three-dimensional image from a sequence of two-dimensional slices (radiographic images), known as reconstruction process (Fig. 1).

Fig. 1
figure 1

Example of sinogram 2D and reconstructed image

Each pixel of the reconstructed image is represented in the Hounsfield unit (HU) scale. The HU value for a voxel with average linear attenuation coefficient μ is given by

$$ \mathrm{HU}=1000\times \frac{\mu -{\mu}_{\mathrm{water}}}{\mu_{\mathrm{water}}-{\mu}_{\mathrm{air}}} $$
(1)

where μair and μwater are the linear attenuation coefficients of air and water, respectively. In this scale, water is represented by 0, air by −1000, and more dense bones are represented by 3000 HU (Hsieh 2003).

Typical applications of CT include the detection of brain tumors, visualization of blood vessel defects, and differentiation of soft tissues, among others. There are some variations of computed tomography such as positron emission tomography (PET-CT) used to diagnose certain tumors (Wahl and Wagner 2009) and cone beam tomography (CBCT) used in dentistry (Kapila 2014).

Magnetic resonance imaging (MRI) is another medical imaging technique used in radiology for diagnosis and treatment purposes.

MRI scanners employ strong magnetic fields to align the nuclei of hydrogen atoms (proton) and apply radio frequency pulses in the region under analysis to create a resonance in the proton atoms (Kelsey et al. 2013). When the radio frequency pulse is turned off, the protons align again to the magnetic field, releasing energy that is recorded by the MRI equipment. Signals at different frequencies are used to reconstruct an image.

Advantages of the MRI include its absence of ionizing radiation and capacity of providing detailed images of soft tissues (for instance, in brain and spine examination). Disadvantages include the cost of equipments compared to other modalities (Sistrom and McKay 2005), as well as counter-indication in patients with metallic devices (pacemakers, orthopedic implants) or who are claustrophobic. Figure 2 shows some examples of MRI images.

Fig. 2
figure 2

Examples of MRI images

Ultrasonography is a diagnostic imaging technique that employs sound waves to form images. This technique allows the visualization of internal structures of the body such as muscles, joints, tendons, blood vessels, liver, and heart. Ultrasonography is also widely used in examination of pregnant women to evaluate the development of the fetus.

As shown in Fig. 3, a transducer converts electrical current into sound waves with frequencies between 3 and 30 MHz (Lutz et al. 2011), which are sent into the body tissues. These frequencies are higher than those audible to humans (>20 KHz). The waves are echoed off the internal structures and reflected back to the transducer, which converts the waves into electrical signals. Then, a computer converts the electrical signal patterns into an image (Dowsett et al. 2006).

Fig. 3
figure 3

Main components of an ultrasound system

Ultrasonography has several advantages compared to other medical imaging techniques. It can provide real-time images, the equipments are portable and lower in cost, and it does not use ionizing radiation. Disadvantages include certain limitations in terms of its field of view, noise, and difficulties in visualizing structures behind air and bones. Figure 4 shows a 2D ultrasound image.

Fig. 4
figure 4

Example of a 2D ultrasound image

2.2 DICOM Format

Until the early 1980s, the examinations based on medical imaging were printed on X-ray film or paper in the case of ultrasonography. Many practices and specifications were generated by a variety of manufacturers, causing serious problems in terms of data management and connectivity incompatibilities (Oosterwijk and Gihring 2002). With the development of advanced imaging techniques and availability of computers, manufacturers started storing medical images in digital format, however, in their own proprietary format.

Digital Imaging and Communications in Medicine (DICOM) (Pianykh 2009) is a standard for storing, printing, and transmitting information in medical imaging. The first version of the standard was released by American College of Radiology (ACR) and National Electrical Manufacturers Association (NEMA) in 1985. The second and third versions were released in 1988 and 1993, respectively.

DICOM provides a file format definition and a network communication protocol based on the Transmission Control Protocol/Internet Protocol (TCP/IP) (Stallings 2005). DICOM enables the integration of different equipments, such as scanners, computers, printers, and servers from different manufacturers, such that files can be exchanged among heterogeneous devices capable of recognizing patient images and data in DICOM format. DICOM standard has been widely adopted by hospitals, medical offices, and dental clinics.

3 Preprocessing

The preprocessing step concerns the techniques for enhancing characteristics of the image, before its use in subsequent tasks such as segmentation and recognition. The preprocessing includes noise filtering, contrast enhancement, and edge detection.

3.1 Noise Filtering

Noise is a random variation in color or brightness of the image, but not present in the physical objects of the scene. Noise usually occurs during the image acquisition process, and it is an undesirable artifact and may decrease the efficiency of segmentation, registration, and classification stages. As desirable characteristic, a smoothing filter, should reduce the occurrence of noise while preserving the edges of an image, since they are important visual features.

Several techniques have been proposed to filter images with the purpose of reducing noise effect. Pixel or voxel values are modified by taking into account intensity values over a neighborhood.

Mean filtering (Gonzalez and Woods 2002; Parker 2010; Russ 2015) is a simple method for smoothing images, where each pixel or voxel intensity value of the image is replaced with the average value of its neighbors. Mean filtering can be performed through a convolution operation with a kernel of adequate shape and size.

Median filtering (Gonzalez and Woods 2002; Parker 2010; Russ 2015) replaces the pixel or voxel intensity value of the image with the median of the neighboring values. The median filtering is more robust than the mean filtering since nonrepresentative values will have little impact on the median value. Furthermore, it can preserve sharp edges.

Gaussian filtering (Gonzalez and Woods 2002; Parker 2010; Russ 2015) is a smoothing operator that uses a Gaussian shape in the convolution process. In this filtering, the pixel or voxel values are replaced with weighted average of its neighbors following a Gaussian distribution.

Bilateral filtering (Tomasi and Manduchi 1998) is a smoothing technique that preserves image edges. Each pixel or voxel of the image is replaced with the weighted average of its neighbors, such that the weights take into account both spatial distances and intensity differences.

Figure 5 illustrates the application of smoothing filtering with mean, median, Gaussian, and bilateral filters to a given input image. The performance of noise filtering techniques varies according to the image modality. Quantitative and qualitative analysis studies have been reported in the literature (Andria et al. 2012; Manjón et al. 2008; Michailovich and Tannenbaum 2006; Salinas and Fernández 2007).

Fig. 5
figure 5

Different techniques for image noise filtering

3.2 Edge Detection

Edge detection techniques (Gonzalez and Woods 2002; Parker 2010; Russ 2015) aim to identify boundaries between regions or objects through radiometric (intensity) discontinuities in the image. Pixels or voxels with sharp changes in brightness correspond to the edges of the image.

Most edge detection methods are based on the directional maxima of the gradient magnitude computed in a local neighborhood of each pixel in the image. The gradient direction is used to estimate the local orientation of the edges.

More formally, the gradient ∇f of an image f is expressed as

$$ \nabla f=\left[\frac{\partial f}{\partial x},\frac{\partial f}{\partial y}\right] $$
(2)

The edge strength corresponds to the gradient magnitude, given by

$$ \left\Vert \nabla f\right\Vert =\sqrt{{\left(\frac{\partial f}{\partial x}\right)}^2+{\left(\frac{\partial f}{\partial y}\right)}^2} $$
(3)

The gradient direction, perpendicular to the edge, is given by

$$ \theta ={\tan}^{-1}\left(\frac{\partial f}{\partial y}/\frac{\partial f}{\partial x}\right) $$
(4)

Figure 6 shows an example of edge map calculated as the gradient magnitude in a neighborhood of each pixel in an input image.

Fig. 6
figure 6

Example of edge map obtained through gradient magnitude

The image gradient can be estimated by applying finite-difference approximations of the first-order and second-order derivatives. Examples of edge detectors based on gradient information are Prewitt and Sobel (Gonzalez and Woods 2002), whose 3 × 3 kernels are expressed in Eqs. (5) and (6), respectively, expressed as

$$ \frac{\partial f}{\partial x}\approx \left[\begin{array}{lll}-1& 0& +1\\ {}-1& 0& +1\\ {}-1& 0& +1\end{array}\right]\, \frac{\partial f}{\partial y}\approx \left[\begin{array}{ccc}-1& -1& -1\\ {}0& 0& 0\\ {}+1& +1& +1\end{array}\right] $$
(5)
$$ \frac{\partial f}{\partial x}\approx \left[\begin{array}{lll}-1& 0& +1\\ {}-2& 0& +2\\ {}-1& 0& +1\end{array}\right]\, \frac{\partial f}{\partial y}\approx \left[\begin{array}{ccc}-1& -2& -1\\ {}0& 0& 0\\ {}+1& +2& +1\end{array}\right] $$
(6)

Figure 7 illustrates the results after applying Sobel operator to the image shown in Fig. 6a. Horizontal and vertical derivative approximations are computed and combined to generate the final edge map.

Fig. 7
figure 7

Illustration of edge map computed through the Sobel operator

The Canny edge detector (Canny 1986) uses a Gaussian filter to initially smooth the image and attenuate noise. Then, the gradient is computed through a filtering such as Prewitt or Sobel, as mentioned previously. A nonmaximum suppression stage is used to maintain only the local maximum gradient values. This steps works as an edge thinning technique. Since the resulting edges can still contain pixels caused by noise or intensity variation, a double-threshold strategy is employed to remove spurious responses. Strong edge pixels correspond to those pixels with gradient value higher than the high threshold value, whereas weak edge pixels are those ones whose gradient value is smaller than the high threshold and larger than the low threshold. Pixels whose gradient value is smaller than the low threshold are suppressed. A hysteresis step is applied to keep the strong edge pixels while weak edge pixels can be connected to a strong edge pixel or removed according to its 8-neighborhood. Figure 8 illustrates the result after applying the Canny edge detector to the image shown in Fig. 6a.

Fig. 8
figure 8

Canny edge detector applied to an image

3.3 Contrast Enhancement

Image enhancement techniques (Gonzalez and Woods 2002; Parker 2010; Russ 2015) have been developed to improve the quality of images for human interpretation purposes. Contrast is one of the most common aspects involved in the subjective evaluation of image quality, which is measured as the difference in brightness or color between two adjacent regions.

Several contrast enhancement techniques (Kim et al. 1998; Singh and Bovis 2005; Stark 2000) use the image histogram to adjust its intensity values. An image histogram is a graphical representation of the number of pixels as a function of their intensity.

Examples of contrast enhancement approaches based on histograms include linear and nonlinear transformations. An important nonlinear technique is the histogram equalization, which redistributes the intensity values of pixels in the image such that the resulting image contains a uniform distribution of intensities. Figure 9 illustrates the result after applying histogram equalization to an image.

Fig. 9
figure 9

Histogram equalization technique applied to an image

4 Segmentation

Image segmentation is the process of extracting objects or regions of interest from an image. This procedure is usually performed by applying a certain criterion to a range of intensity values, such as connectivity, similarity among others. An example of medical image segmentation is the selection of bone or a specific organ, such as heart or lung, for analysis and visualization purposes. Several methods have been proposed for image segmentation in the literature (Haralick and Shapiro 1985; Pal and Pal 1993; Zhang 1996). The following subsections present some relevant image segmentation techniques.

4.1 Thresholding

A common segmentation technique is known as thresholding (Sankur and Sezgin 2001), whose purpose is to identify an intensity value (grayscale or color) to separate objects from the background. Thresholding can find a global and unique threshold value for the entire image or a local and adaptive threshold for different regions of the image. For instance, considering an interval of values between tmin and tmax, pixel values outside this range could be set to 0, whereas pixel values within the interval could be assigned to 1. Let f (x, y) and g(x, y) be the input image and the resulting image after the thresholding process, respectively. Then,

$$ g\left(x,y\right)=\left\{\begin{array}{ll}1& \mathrm{if}\;{t}_{\mathrm{min}}\le f\left(x,y\right)\le {t}_{\mathrm{max}}\\ {}0& \mathrm{otherwise}\end{array}\right. $$
(7)

Global thresholding presents some drawbacks since it is sensitive to noise and dependent on the actual image intensity range. For instance, if the patient has a metallic material prosthesis, the global thresholding applied to a computerized tomography can produce artifacts due to the presence of noise with intensity similar to bone structures. In this case, adaptive local thresholding would be more effective.

4.2 Region Growing

The region growing technique (Adams and Bischof 1994) uses a seed located on the object or region of interest, which can be manually specified by the user or automatically assigned. Then, the seed expands gradually in accordance with its neighboring pixels. The expansion process interrupts when neighboring pixels satisfy a stopping criterion. Typical connectivity types for two-dimensional images are 4 and 8 connected neighborhood, whereas for three-dimensional images are 6-, 18-, and 26-connected neighborhood.

Stopping criteria are usually based on the homogeneity of the grayscale values or textural information present in the medical images.

4.3 Watershed

The watershed segmentation technique (Beucher 1994) is based on concepts of topography and hydrography, where an image is considered as a topographic surface, such that catchment basins are formed through a flooding process. The water follows the gradient of the image and flows along a path to reach a local minimum. At the end of the flooding process, the watershed of a relief corresponds to the limits defined by adjacent catchment basins.

An example of watershed segmentation applied to medical images is the separation of normal and abnormal tissues for detecting brain tumor in MR images.

5 Registration

Image registration (Brown 1992; Zitova and Flusser 2003) is the process of integrating two or more images into a common coordinate system. Medical image registration can be usually categorized into single-modality and multimodality approaches. Single-modality methods align images acquired by the same scanner type, whereas multimodality methods align images acquired by different sensor types, such CT and MR.

The registration process is useful for aligning data of a same patient captured at different view points or along different periods of time to monitor a treatment or the progress of a disease, such as a tumor growth.

Substantial research has been conducted on the image registration field over the last decades (Crum et al. 2014; Hill et al. 2001; Lester and Arridge 1999; Maintz and Viergever 1998). Several techniques have been developed to make the task successful with little or no human intervention, more precise, faster, and less sensitive to noise or occlusions.

Two stages are usually involved in the image registration: geometric and radiometric transformations. A geometric transformation maps spatial positions in one image to another, whereas a radiometric transformation maps intensity values in one image to another.

Geometric transformations perform basic operations such as resizing, translation, rotation, and projections. In addition to such rigid transformations, nonrigid or elastic transformation can also be applied to locally align one target image with a reference image. Examples of elastic transformations include the radial basis functions, which can be further categorized into other basis functions (Buhmann 2000), such as thin-plate spline, spline with tension, and multiquadric function.

Geometric transformations in the spatial domain aim to match features or intensity patterns in the images. On the other hand, frequency-domain geometric transformations explore invariance properties to simplify the location of correspondences between a pair of images. Peaks found in phase correlation methods through the fast Fourier transform (FFT) correspond to the relative translation between a pair of images. Differences under scaling and rotation can also be determined using log-polar coordinates and exploring Fourier transform properties.

Salient features, such as corners, edges and regions, can be detected from the images and used to match them. Common metrics used in the matching process include cross-correlation and mutual information.

6 Volume Rendering

Volume rendering can be defined as a set of techniques for displaying bidimensional projections from tridimensional data sets. These techniques can be categorized into two groups: isosurface rendering and direct volume rendering .

6.1 Isosurface Rendering

Isosurface rendering methods involve the extraction of an intermediary geometric representation from the volume data set. The most commonly used representation is the triangle mesh, which is formed by vertices and triangles that connect its vertices through edges. Figure 10 shows a triangle mesh and a magnified region to allow visualization of more details. Isosurfaces are useful to represent internal organs, bones, or other anatomic structures.

Fig. 10
figure 10

Illustration of (a) an isosurface extracted from a volumetric medical data and (b) a magnified region of the mesh

The technique of Marching Cubes (Lorensen and Cline 1987) is one of the most employed methods to generate isosurfaces, where a regular 3D grid, formed by cubes, overlaps the data volume, such that the vertices on the grid match the scalar values of the volume. Each cube edge is verified if it intercepts the surface. These intersections will be part of the isosurface vertices, which will be connected by triangles following one of 256 possibilities to form the surface. The final isosurface is generated by joining the surfaces from all cubes. However, the technique of Marching Cubes presents some ambiguities in some triangle configurations that may generate holes in the isosurface. Other methods have been developed to avoid these problems, such as Marching Tetrahedra (Akio and Koide 1991) and Surface Nets (Gibson 1999).

Graphics processing units (GPUs) have been extensively used to process and display triangle meshes, as well as in 3D printing tasks. The triangle meshes can be used and edited in CAD tools to support surgical planning and prosthesis generation.

The quality of the isosurface depends on the methods used to generate it and on the segmentation process. Segmentation can generate a binary volume as result. Isosurfaces generated from binary volume usually present staircase artifacts, such as the ones shown in Fig. 11a. These artifacts are not natural to the patient’s anatomy and they emerge mainly in high-curvature regions. Some techniques have been proposed to avoid these artifacts. The method developed by Whitaker (2000) smooths the binary volume toward the gradient direction to find the isosurface with minimum area. The context aware smoothing (Moench et al. 2011) method assigns high weighting coefficients to regions with occurrence of staircase while other regions have low weighting coefficients. The weighting coefficients control the degree of smoothing, such that high weighting coefficient areas are more smoothed. This is accomplished to reduce the loss of fine details in the resulting isosurface. Figure 11b shows the result after applying the context aware smoothing to a volumetric data.

Fig. 11
figure 11

(a) Isosurface with staircase artifacts and (b) same isosurface after the application of context aware smoothing algorithm to attenuate staircase artifacts

6.2 Direct Volume Rendering

In direct volume rendering (DVR) (Roth 1982), the voxel values are mapped to colors and opacities according to a transfer function. Figure 12 illustrates an example of a transfer function. Color and opacity values are used to compose the final image according to the technique used. Volume ray casting (Levoy 1990) is one of the most common DVR techniques. In this technique, rays are traced from the observer, passing through an image plane, toward the object. The rays intercept the object voxels, such that the color and opacity values are accumulated to generate the final image. The emission-absorption (Fernando 2004) is an example of model to accumulate color and opacity values, expressed as

$$ {\displaystyle \begin{array}{l}C=\sum \limits_{i=1}^n{C}_i\prod \limits_{j-1}^{i-1}\left(1-{\alpha}_j\right)\\ {}\alpha =1-\prod \limits_{j-1}^n\left(1-{\alpha}_j\right)\end{array}} $$
(8)

where C and α are the accumulated color and opacity values computed iteratively, whereas C i and α i correspond to color and opacity, respectively, obtained from the segment i along the viewing ray.

Fig. 12
figure 12

Example of a linear transfer function, which maps scalar value 0 to black and 0% opacity. The scalar value 650 is mapped to white with 100%, whereas intermediate values are linearly interpolated to obtain color and opacity values

In volume ray casting, structures of interest may be occluded by other structures. For instance, in a CT data set, soft tissues may occlude bones or regions due to contrast. However, it is possible to modify the transfer function, which can be done through user interaction or automatic methods for generating transfer functions. Figure 13 illustrates examples of volume ray casting applied to a volumetric data.

Fig. 13
figure 13

Different views from ray casting for the same volume data set by modifying the transfer function. (a) the transfer function is adapted to reveal soft tissues; (b) the transfer function is modified to display bone structures

Maximum intensity projection (MIP) (Bruckner and Gröller 2009) addresses the problem of occlusion by taking into account the regions of interest with high scalar values. In MIP, the projection rays are not accumulated, instead the maximum scalar values are selected. One characteristic of MIP is its lack of visual depth information, which may lead to ambiguities in the visualization. MIP is useful to visualize bone structures, regions with different contrast, and tumors. Figure 14a illustrates an example of MIP rendering.

Fig. 14
figure 14

A volume data set rendered through different volume rendering techniques

Minimum intensity projection (MinIP) (Mistelbauer et al. 2013) and average intensity projection (AIP) (Wu et al. 2008) are variations of MIP. In MinIP, the projection rays select the minimum values and can be used to provide information about blood flow deviation. In AIP, the projection rays show the mean value of scalar values obtained along their path. An example of AIP rendering is shown in Fig. 14b.

Maximum intensity difference accumulation (MIDA) (Bruckner and Gröller 2009) ponders the accumulation of color and opacity values along the rays through weights, which are calculated according to the maximum values intercepted along the ray path. The greater the difference between the maximum values, the higher the weights are assigned. Prioritizing the maximum values, MIDA can attenuate the problems associated with occlusion. Since MIDA uses accumulation, such as ray casting, the generated visualization presents depth information, which may reduce artifacts related to visualization ambiguities. Figure 14c shows an example of MIDA. Similar to MIP, MIDA is useful to visualize bone structures, regions with contrast, and tumors.

Contour MIP (Csèbfalvi et al. 2001) takes into account the angle between the viewing direction and the gradient vector from voxels intercepted by the projection rays, described as

$$ s\left(P,V\right)={\left(1-|(P)\cdot V|\right)}^n $$
(9)

where ∇(P) is the gradient of voxel P, V is the viewing direction, and n is a sharpness coefficient for the contour. To compose the final image, contour MIP uses the MIP technique, since contour MIP may suffer from ambiguities in its visualization. Contour MIDA (de Moraes et al. 2015) follows the same idea as contour MIP; however, it uses MIDA to compose the final image to attenuate problems associated with ambiguities when compared to contour MIP. Both techniques emphasize voxels with higher values of gradient information, highlighting surfaces of different objects. Examples are illustrated in Fig. 14d and e.

7 3D Printing and Biofabrication

Nowadays, imaging is the basis for a series of applications that are already consolidate for a tailored treatment in medicine and odontology. Imaging is a noninvasive technique for human diagnosis and treatment. Therefore, a 3D model of an anatomy, generated from a patient CT or MRI medical scanner, can be integrated with additive manufacturing (AM) technologies for many medical purposes. This is possible due to the concurrent development of scanning, imaging, and AM technologies. AM, also known as 3D printing or rapid prototyping, is based on the paradigm of a layer-by-layer automatically controlled deposition of materials by means of a myriad of processes. AM can produce directly a physical model from a virtual computational model and is considered a groundbreaking technology due to its flexibility and capacity to produce geometrically accurate complex models. A schematic diagram for 3D printing and additive manufacturing is illustrated in Fig. 15.

Fig. 15
figure 15

3D printing and additive manufacturing techniques

Imaging and AM technologies came to the market in the 1980s decade. Since then, a great evolution on both technologies and its integration for healthcare applications is paramount. Among them can be highlighted the production of 3D physical models, called biomodels, for a better and safer treatment, as well as the development of customized prostheses, medical devices, and surgical instruments (Silva et al. 2009; Sannomiya et al. 2008; Kemmoku et al. 2010; Oliveira et al. 2014; Giacomo et al. 2014).

Today’s AM technologies use as a reference to print objects in a file format called STL (STereoLithography) created for the first machines in the 1980s decade by Albert Consulting Group under 3D Systems demand. This is a very simple representation of an object using triangle meshes to define all of its surfaces (internal and external). Each triangle is represented by means of three coordinates, one coordinate for each triangle vertex, and a normal vector pointing out of the object material, which allows to define the real boundaries of the object. The advantages of STL representation are mainly due to its easy interpretation from computing systems and cross-platform interoperability. To be used in AM, the STL file is sliced by means of parallel planes corresponding to the minimal layer thickness assured by the AM technology being used. The disadvantage is its redundant data representation composed of planar facets without any topological connectivity information. STL is still de facto standard for AM data exchange. The majority of the medical imaging software generate only STL format after all the steps explained in the previous sections of this chapter.

The AMF (Additive Manufacturing File Format) representation is an initiative of ASTM (American Society for Testing and Materials) since 2013 adopted by ISO (International Standardization Organization) as the international standard ISO/ASTM 52915 (2013). The main idea of AMF is to overcome deficiencies and challenges of the STL file format, such as the independence among different AM technologies, represent geometry with colors, texture, multimaterials, and other properties, associated with the simplicity to be implemented and interpreted by computer systems. The adaptability to support part complexity and performance in terms of file size and processing is a desirable specification. AMF is also designed for compatibility with legacy systems (e.g., STL) and potential possibilities to be adapted for new demands for future technology evolutions. One important feature is the possibility to use not only flat triangles but also curved and surface triangles. AMF uses the XML (eXtensible Markup Language) as a standard to store and transport data readable for both human and computer.

Another 3D representation for AM is the 3MF (3D Manufacturing Format) file format. It is being developed by a joint development foundation project for creating the new file standard of AM. This project was originally created by some of the AM, CAD systems and software solutions providers such as Siemens, Dassault Systèmes, 3D Systems, Stratasys, SLM, HP, Materialise, Autodesk, Shapeways, NetFabb, and Microsoft. The 3MF format is developed under the philosophy of open-source. Similarly to STL and AMF, this format is a representation of an object composed of triangle element meshes. The proposal of 3MF is to promote a rich representation of an object in details, including colors, materials, and other characteristics. The 3MF will support extensions and innovations of the AM area, and interoperability among systems (3MF Consortium 2016b, c). It is expected for the consortium members a wide adoption. Its specification and codification can be found and downloaded from GitHub (3MF Consortium 2016a).

Looking toward biological directions of AM, imaging will play a fundamental role as a basis for the reverse engineering of organs and tissues anatomy. More recently, AM started to include biomaterials and biological materials as cells and microtissues as an option for printing. This is a new area called Biofabrication as rapidly growing field of research as a solution for the shortage of tissues and organs donation for restoration and replacement of human functionalities. Then, Biofabrication has appeared and grown in the last decade with many developments, potentials, and trials but also challenges.

Very recently, the term Biofabrication was reappraised by the 2010 created International Society of Biofabrication (ISB). Many novelties and developments obliged ISB to update the meaning of this wide word considering the derivations Bioprinting and Bioassembly. Then, the new ISB definition repositions Biofabrication as “the automated generation of biologically functional products with structural organization from living cells, bioactive molecules, biomaterials, cell aggregates such as micro-tissues, or hybrid cell-material constructs, through Bioprinting or Bioassembly and subsequent tissue maturation processes” (Groll et al. 2016). Three approaches as branches of Biofabrication have evolved: scaffold-based, solid-scaffold-free based on cell and tissue spheroids deposition, and a mix of them named as third strategy or hybrid technology (Rezende et al. 2012; Danilevicius et al. 2015).

The scaffold, as a first approach, consists of a three-dimensional structure as a temporary support for cells growing and maturing of a new tissue. The scaffold must be biodegradable, offering suitable conditions in such way the cells can be seeded and posteriorly adhere, differentiate, and proliferate for new tissue formation. Since AM allows for highly precise construction of 3D complex geometries and controlled porosity, scaffolds have been fabricated in many materials for different applications such as bone, cartilage, coronary, skin, and dental, among others. Reinforcing the role of controlled geometry of scaffolds with interconnectivity and therefore the guidance for the new tissue formation, many imaging techniques have been applied in order to verify the real scaffold’s shape and geometry and to validate their applications. Noninvasive imaging techniques using CT and MRI have been researched to be used as a follow-up tool for an accurate assessment of bioresorbable vascular scaffolds (BVSs) after a percutaneous coronary intervention (Hickethier et al. 2016). Other authors have analyzed BVSs implanted in vivo by means of the optical frequency domain imaging (Naganuma et al. 2015). Even for scaffolds produced by conventional methods (non-AM), imaging has benefited the diagnosis of implanted scaffolds allowing for 3D quantification of parameters such as porosity, thickness, and surface area (Kłodowski et al. 2014).

In turn, solid-scaffold-free, based on cell or tissue spheroids controlled deposition, so-called Bioprinting or Organ Printing is defined as a layer-by-layer additive robotic computer-aided biofabrication of functional 3D organ constructs using self-assembling tissue spheroids according to a digital model or blueprint (Derby 2012; Melchels et al. 2012; Mironov et al. 2008; Rezende et al. 2015). On the other hand, Bioprinting is a dynamic strategy of biofabrication and is conceptually split into four stages: blueprint, materialization, maturation, and clinical application (Fig. 16). Each of these Biofabrication steps will, in the future, make use of imaging techniques for anatomical reverse engineering, mimicking and characterizing tissue micro and macro anatomic details, and also for quality control during maturation and clinical applications.

Fig. 16
figure 16

Stages of bioprinting, from blueprint to clinical (Rezende et al. 2015; Murphy and Atala 2014)

Different processes for bioprinting have been proposed (Atala and Yoo 2015), and they can be summarized as follows:

  • Inkjet Printing: a noncontact printing technique that takes data from a computer representing an image or character, and reproduces it onto a substrate by ejecting tiny ink drops on a drop-on-demand manner (Mohebi and Evans 2002; Chua and Yeong 2014). Presorted cells are available in a 2D–printercartridge-like. These cells are disposed drop-by-drop onto an XY Z platform.

  • Scaffold-Free Printing: this technique makes use of fluidity, where the fusion of the cell aggregates or cell rods are governed by the surface tension among two or more of those structures (Jakab et al. 2008). The live cells are encapsulated by some hydrogel, which assures mechanical protection and stability in order for the whole structure to become stronger. Thus, the cells can fuse to each other, generating the tissue and posteriorly the organ.

  • Extrusion-Based Deposition: this technique uses a syringe for extrude the material under controlled pressure and temperature onto the platform. The material is normally at semi-melted condition which eases its deposition (Skardal et al. 2010). The materials are normally polymers or hydrogels and must be soft or fluid enough to facilitate extrusion through the small diameter tip or nozzle assuring mechanical strength itself after the deposition (Atala and Yoo 2015).

  • Laser-Induced Forward Transfer: this technique utilizes a pulsed laser that is used to induce the transfer of material from a source film spread onto an optically transparent quartz support to a substrate in close proximity to or in contact with the film (Guillemot et al. 2010). This film is comprised of a printable material (biopolymer or protein) containing cells inside that are then deposited in a pattern on the substrate to create cellular structures and patterns. LIFT is supported on a transport layer, such as gold or titanium, which absorbs the laser energy and transfers it to the ribbon.

Bioprinting technology has a broad utility in various application areas, such as tissue engineering and regenerative medicine, transplantation and clinics, drug screening and high-throughput assays, and cancer research. Currently, a wide variety of tissues have been successfully bioprinted, such as thin or hollow tissues, for example, blood vessel, and tissues as such as bone, cardiac, cartilage, liver, lung, neural, pancreas, skin, and vascular (Ozbolat et al. 2016).

Image processing and computer graphics techniques for the creation of three-dimensional models or a blueprint for visualization and building of a complex organ are largely required. Therefore, for a bioprinting process the preprocessing includes data acquisition, printing model construction, printing path optimization, and materials preparation. Data acquisition refers to the process of obtaining anatomic geometries of the targeted tissues/organs and is usually implemented via medical imaging techniques, such as MRI and CT (Wu et al. 2017). So far, there is no sufficient knowledge to precisely define in which extent imaging will contribute for the Biofabrication field.

Medical imaging technology is an essential tool used by tissue engineers to provide information on 3D structure and function at the molecular, cellular, tissue, organ, and organism level. Computer-aided design/computer-aided manufacturing (CAD/ CAM) tools are also used to collect and digitize the complex tomographic and architectural information of the tissues. A comprehensive understanding of the composition and organization of the components that make up these tissues is the key requirement for reproducing the complex heterogeneous architectures that exist within functional tissues and organs (Murphy and Atala 2014).

For a complete blueprint, the most critical point is to obtain the best representation of human organs due to their complexity. For instance, replicating details of microvasculature of a 3D structure is hard task either from the anatomical or from mathematical and computational perspectives. Most of complex human organs are highly vascularized and have elaborated intraorgan branched vascular trees which dramatically increase the level of their structural organization and make task of designing corresponding organ blueprint more challenging (Rezende et al. 2015). Furthermore, so far there is no software or model powerful enough to establish a blueprint.

Additionally, biomimicry and biomimetic are also considerable resources for the making-up of the blueprint, and they can be grouped in two main categories: those based on the use of imaging resources for the digitization of biological structures and those based on the use of mathematical and software resources for the construction of geometries with properties similar to those found in biological materials (Lantada 2016).

During the last decade, increasing attention has been paid for using fractals for promoting modeling, design, and simulation tasks in several areas of Biomedical Engineering, some of them also linked to the development of novel biomedical microdevices for interacting at a cellular and even molecular level. This includes modeling the behavior of microorganisms, complex organisms and their systems, scaffolds for tissue engineering, and the surfaces of organs and tissues (Lantada 2016).

In the conventional production systems, not only a product representation associated with user requirements and applications will not suffice. It is necessary to have available production lines and methods for fast, safe, and lean production of any product. As an analogy with this field, Biofabrication will face the same needs to be competitive when all the intermediate steps and associate challenges are overcome. Hence, a biofabrication line will be necessary with all the processes planning and execution needed.

A smart platform that enables real-time monitoring, informative decision-making, and automated operation will be necessary. Some key components include a data-driven interface, automated biofabrication line, efficient communication, and novel monitoring techniques customized for biomedical applications (Wu et al. 2017).

From an engineering point of view, the successful clinical translation of organ printing technology will depend not only on the development of a single tool or device, such as a robotic bioprinter, but rather on the entire complex of related technologies and their seamless integration, i.e., an Organ Biofabrication Line (OBL). The design and development of a fully integrated OBL, or development of a series of integrated automated robotic tools, is imperative for the future commercial translation of the organ printing technology (Mironov et al. 2011).

An organ biofabrication line will combine design, computer simulation, visual information of machines, devices, human resources and processes, as well as it must include in silico assays about tissue spheroids formation, spheroids fusion, cell viability, tissue and organ development in each corresponding stage. A virtual OBL must integrate essential tools such as CAD, CAE, CAM, complex mathematical modeling, and new specific software for biological simulations, special equipment and information technologies in general (Passamai et al. 2016).

Beyond the 3D representation available and in development today for AM, the Biofabrication field will need a complete new way to represent tissues and organs, based on its functionality, multimaterials composition, and micro/macro anatomic details. Imaging will play a fundamental role for this purpose. More comprehensive design systems need to be developed in order to integrate other crucial biological factors into the process of building successful tissue constructs (Ozbolat and Gudapati 2016). A biofabrication line for organ production will be the next generation of production and, again, imaging will be necessary for characterization and quality control. Finally, new engineering tools associated with strong biological knowledge and translational methods will be necessary to support the biofabrication emerging area.

8 Conclusions

Continuous developments in medical image analysis have aided healthcare professionals in several diagnostic tasks. Technological advances in digital computing, medical imaging, biomaterials, and biofabrication have contributed to a broad range of medical applications. Critical diagnostic decisions have been improved by the integration of powerful computer-aided tools and manufacturing technologies.

In this chapter, we provided an overview of fundamental concepts related to the construction of three-dimensional models for visualization and biofabrication purposes. The understanding of such key topics is challenging due to the diverse and multidisciplinary field of biomedical engineering, however, of paramount importance to the development of innovative solutions to a variety of medical and health problems.