Skip to main content
  • Original article
  • Open access
  • Published:

Comparative analysis of alignment algorithms for macular optical coherence tomography imaging

Abstract

Background

Optical coherence tomography (OCT) is the most important and commonly utilized imaging modality in ophthalmology and is especially crucial for the diagnosis and management of macular diseases. Each OCT volume is typically only available as a series of cross-sectional images (B-scans) that are accessible through proprietary software programs which accompany the OCT machines. To maximize the potential of OCT imaging for machine learning purposes, each OCT image should be analyzed en bloc as a 3D volume, which requires aligning all the cross-sectional images within a particular volume.

Methods

A dataset of OCT B-scans obtained from 48 age-related macular degeneration (AMD) patients and 50 normal controls was used to evaluate five registration algorithms. After alignment of B-scans from each patient, an en face surface map was created to measure the registration quality, based on an automatically generated Laplace difference of the surface map–the smoother the surface map, the smaller the average Laplace difference. To demonstrate the usefulness of B-scan alignment, we trained a 3D convolutional neural network (CNN) to detect age-related macular degeneration (AMD) on OCT images and compared the performance of the model with and without B-scan alignment.

Results

The mean Laplace difference of the surface map before registration was 27 ± 4.2 pixels for the AMD group and 26.6 ± 4 pixels for the control group. After alignment, the smoothness of the surface map was improved, with a mean Laplace difference of 5.5 ± 2.7 pixels for Advanced Normalization Tools Symmetric image Normalization (ANTs-SyN) registration algorithm in the AMD group and a mean Laplace difference of 4.3 ± 1.4.2 pixels for ANTs in the control group. Our 3D CNN achieved superior performance in detecting AMD, when aligned OCT B-scans were used (AUC 0.95 aligned vs. 0.89 unaligned).

Conclusions

We introduced a novel metric to quantify OCT B-scan alignment and compared the effectiveness of five alignment algorithms. We confirmed that alignment could be improved in a statistically significant manner with readily available alignment algorithms that are available to the public, and the ANTs algorithm provided the most robust performance overall. We further demonstrated that alignment of OCT B-scans will likely be useful for training 3D CNN models.

Background

OCT background

Optical Coherence Tomography (OCT) is a major technological breakthrough in medical imaging, after the invention of computed tomography (CT) [1] and magnetic resonance imaging (MRI) [2]. OCT [3] has revolutionized the field of ophthalmology, as it can image tissue non-invasively at a micron level resolution, both for the anterior segment of the eye [4] and posterior segment of the eye [5]. It takes advantage of the different reflectivity of tissues to determine the delay time and reflection intensity of the emitted light waves, by comparing the reflected and reference light waves through a low coherence optical interferometer [6].

Within ophthalmology, OCT is particularly important for the diagnosis and management of different retinal diseases, for example age-related macular degeneration (AMD), which is the leading cause of central vision loss in persons over the age of 50 in the United States [7]. During the acquisition of an OCT image in a patient with AMD or other macular diseases, the scan area is centered on the macula, which is responsible for central high-resolution vision, and light rays are passed through the macula in an anterior–posterior fashion, producing numerous A-scans. After the image acquisition is completed, the A-scans are combined to create a single B-scan, which is a cross-sectional image of the macula. Within a scan area, multiple B-scans are typically captured, though the density of the B-scans within a given scan area can be adjusted by the camera operator. In other words, every OCT image acquisition produces a 3D image volume, composed of multiple B-scans, but the density of the B-scans (distance between each cross-sectional image) varies depending on the scanning protocol. In retina clinical practice and imaging research involving OCT images, each image acquisition (3D volume) is accessed and usually viewed one B-scan at a time, but the B-scans typically are not aligned. The misalignment of the B-scans does not pose a challenge for clinicians, who are primarily interested in the presence or absence of pathologies in individual B-scans, but this lack of alignment of B-scans could preclude analysis of OCT images en bloc as a 3D volume for machine learning purposes.

Being able to analyze OCT images en bloc as a 3D volume has significant implications for deep learning (DL) neural network-based image classification. Within the context of DL-based image classification, most published DL studies on macular OCT only utilized individual 2D B-scans as units of data input. Examples include differentiating between normal vs. AMD [8] and differentiating between choroidal neovascularization vs. diabetic macular edema vs. drusen vs. normal [9]. While choosing 2D B-scans within a 3D OCT volume as units of data for training and testing simplifies data curation and side-steps the complexity of 3D neural networks, this is not ideal, as only a portion of the available information is utilized. A more advanced strategy involves analyzing the entire 3D OCT volume en bloc. However, this will require efficient and accurate alignment of the B-scans within the same OCT 3D volume in the first place, and a tool that can achieve this is not readily available to investigators interested in 3D OCT imaging research.

Image alignment

Image alignment [10,11,12] is a method of optimally registering one or more images onto a target image. Image alignment has many practical applications in medical image processing and analysis [13,14,15]. Medical image alignment can be divided into intra-subject (images from the same patient), inter-subject (from different patients) and atlas alignment (alignment of patient data to an atlas). To align the B-scans within the same OCT volume (intra-subject), the spatial transformation may be either a rigid or non-rigid deformable transformation. In this paper, we focused on three-parameter (one rotation and two translations) rigid alignment, and evaluated five medical image alignment methods: Advanced Normalization Tools Symmetric image Normalization (ANTs-SyN) [16], FMRIB’s Linear Image Registration Tool (FLIRT) [17], Insight Toolkit (ITK) [18], Optimized Automatic Registration (OAR) [19], TOADS-CRUISE (TOADS) [20], with a cross-correlation cost function.

Our work

For macular OCT images, there is currently no consensus on how to assess the correct alignment of individual B-scans within a 3D volume. To this end, we propose a novel metric, the en face surface smoothness, which is more specific to B-scan-to-B-scan signal change and corroborates with the actual anatomy of a human macula, i.e., it is medically accurate that the inner retinal surface of a human macula, in the absence of surgical manipulation or major trauma, is smooth. We aligned the OCT images from all patients in our dataset, using the five medical image alignment methods.

The key contributions of our paper are: (1) developed of a novel metric, the en face surface smoothness, to quantify OCT B-scan alignment that is based on the actual anatomy of human maculae, and (2) compared and contrasted five commonly-used and readily-available alignment algorithms, and identified the algorithm with the best performance to align OCT B-scans within a 3D volume on the current image samples. To further demonstrate the utility of OCT B-scan alignment, we trained a 3D convolutional neural network (CNN) to detect age-related macular degeneration (AMD) on OCT images and compared the performance of the model with and without B-scan alignment. We chose to use AMD as a use case, as AMD is a leading cause of blindness in the world and OCT is indispensable in the diagnosis and management of AMD.

Methods

Dataset

The dataset [21] used in the alignment experiments was acquired at the Noor Eye Hospital in Tehran and consists of 50 normal and 48 non-neovascular AMD OCT volumes, acquired with a spectral domain optical coherence tomography system [Spectralis, Heidelberg, Germany]. For this dataset, the axial resolution is 3.5 μm with a scan area of 8.9 × 7.4 mm2. The number of A-scans was either 512 or 768 and the number of B-scans per OCT volume varied from 19, 25, 31, to 61.

The dataset used in the classification experiment was acquired at Duke University (230 OCT volumes; 115 AMD; 115 normal control) [22]. The volumes were acquired in a scan area of 6.7 × 6.7 mm centered on the fovea with a rapid scan protocol, resulting in volumes of 1000 × 512 × 100 voxels. The data was randomly split into training (56%, N = 130), validation (22%, N = 50) and test (22%, N = 50) and was partitioned at the patient level.

B-scan rigid alignment

We used five commonly used algorithms to align the B-scans in each 3D-OCT volume: ANTs, FLIRT, OAR, ITK and TOADS. Advanced Normalization Tools Symmetric image Normalization (ANTs-SyN) was proposed in 2007 by the University of Pennsylvania [23], which developed a novel symmetric diffeomorphic optimizer for maximizing the cross-correlation in the space of topology preserving maps. The FMRIB's Linear Image Registration Tool (FLIRT) alignment algorithm [10, 19, 24] was proposed in 2000 by the University of Oxford and investigated assumptions underlying the problem of aligning brain images using a cross-modal voxel similarity measure. The Insight ToolKit image alignment (ITK) [18] alignment algorithm was proposed in 2003 by the University of Pennsylvania. Optimized automatic alignment 3D (OAR) [19] was a method proposed for optimal alignment based on FLIRT. The OAR technique specifies a transformation that minimizes a cost function, which represents the quality of alignment between two images. The TOADS-CRUISE Brain Segmentation Tools (TOADS) [25] is a series of software plug-ins developed by Johns Hopkins University in 2009 for automatic segmentation of magnetic resonance brain images, which was then adopted for OCT alignment with deformable registration.

We use \(\left\{{R}_{1},{R}_{2},{R}_{3},...{,R}_{M-1},{R}_{M}\right\}\) to represent the B-scan images of a particular patient and RM/2 is the central (foveal) B-scan image. The B-scan alignment starts at the center (fovea) and so RM/2–1 and RM1/2+1 were aligned to RM/2. The alignment was propagated outward from the center and is outlined in Algorithm 1.

figure a

Building the surface map and edge map

After the B-scans were combined (aligned or not aligned) into a 3D volume, an en face surface map was created. Figure 1 is a schematic representation for how the surface smoothness metric was generated. Multiple OCT B-scans from the same volume (eye) scan were combined to form a 3D cube. The surface map was defined by the distance from the top of the data cube to the top of the nerve fiber layer (Fig. 1, white arrow). We automatically remove the background noise from each B-scan image by the mean-shift [26] clustering algorithm to locate the nerve fiber layer of each B-scan image. The retinal surface located by the clustering algorithm was subsequently fit using a third order spline with locally weighted smoothing [27] (statsmodel.nonparametric.lowess from https://www.statsmodels.org) to remove very small jagged edge artifacts from the result of the clustering algorithm. Subsequently, the surface map was created by finding the nerve fiber layer of each B-scan. It is expected that well aligned B-scans will result in a smooth surface map (Fig. 1, bottom row) compared to unaligned B-scans (Fig. 1, top row). The measure of the smoothness is defined as the mean value of the edge map, which is created by applying a discrete Laplace operator over the surface map.

Fig. 1
figure 1

Multiple OCT B-scans from the same patient scan (far left) are combined to form a 3D cube. The surface map is defined as the distance from the top of the data cube to the top of the nerve fiber layer (white arrow). It is expected that well aligned B-scans will result in a smooth surface map. The measure of the smoothness is defined as the mean value of the difference map which is created by applying a discrete Laplace operator over the surface map

The edge map was calculated by applying Laplace edge detection algorithm [28] to the surface map. The Laplace operator is the simplest isotropic differential operator that is rotationally invariant and is defined as:

$$Laplace(f)=\frac{{\partial }^{2}f}{\partial {x}^{2}}+\frac{{\partial }^{2}f}{\partial {y}^{2}}$$

The mean of the edge map was automatically calculated and used as the measure of the smoothness of the surface map.

Surface map validation

To validate the algorithm for generating the surface map, we generated two sets of surface maps for six patients (three AMD and three control patients; 360 B-scan images in total). The first set was generated by manual annotation of the nerve fiber layer (the surface layer of the retina). The second set was generated automatically using the method described above. The difference between the two sets of surface maps was represented by a histogram.

Neural network model and training

We used the C3D CNN as the backbone of our model [29]. The training parameters included 2000 epochs, learning rate of 0.0001 (learning rate is divided by 10 every 50 epochs), Adam optimizer, cross entropy loss function, and a batch size of 2. The OCT signal was normalized to zero mean and scaled to unit standard deviation. Image augmentation consisted of a random horizontal flip. The proposed method was implemented using the Pytorch framework in Python with NVIDIA Cuda v10.0 and cuDNN v7.2 acceleration libraries. All experiments were performed on a Windows 10 machine with an Intel Core i7-7700 K 3.60 GHz CPU and an NVIDIA RTX TITAN GPU with 32 GB memory.

Statistical method

We expected the smoothness of the surface map to increase, and therefore the difference in the surface map to decrease, after B-scan alignment. The mean difference in surface map smoothness before and after alignment was compared, using paired-sample t-tests with the null hypothesis that the two means were the same.

Results

Qualitative results

Figure 2 was created to demonstrate visually the kind of artifacts that could be introduced if no B-scan alignment is performed, and shows the surface maps from four sample control and four sample AMD OCT volumes. Each row represents a separate eye. Each column represents a different alignment algorithm: ANTs [16, 30, 31], FLIRT [32,33,34], ITK [35,36,37], OAR [38] and TOADS [39, 40]. The left-hand column shows the en face surface map using unregistered B-scans. In general, the alignment algorithms created a smoother inner retinal surface for all eight eyes. In addition, these are our observations. First, the horizontal streaks in the images (e.g. Control 2 and AMD 1) represented significant mis-alignment of the B-scans along the Z axis in the OCT volumes. Second, the signal intensity differences seen in Control 3 and AMD 3 were due to mis-alignment between the central and peripheral regions within the area of the macula that underwent imaging. Third, within the same eye, such as Control 3, there were significant variations in performance between the algorithms. Fourth, black points that were typically seen near the edges were artifacts created by incorrect surface depth estimates (and quantified as “edge errors” in Fig. 4). Of the five algorithms, the ANTs alignment algorithm appeared to have both the highest degree of surface smoothness and fewest edge errors.

Fig. 2
figure 2

En face surface maps of OCT volumes. Each row represented a different patient, and each column represented a different alignment algorithm. The left most column was created from unregistered B-scans. (AMD Age-related Macular Degeneration)

Quantitative results

In this section, the alignment results will be compared quantitatively. The inner retinal surface of a human macula, in the absence of surgical manipulation or significant trauma, should be smooth. Thus, the success of the alignment algorithms was quantified by the mean of the Laplace difference across an en face surface map. The smaller the mean of the Laplace difference, the smoother the surface map, and the better the alignment.

When all OCT volumes were included for analysis, the mean Laplace difference for AMD and Control groups without alignment was 27.0 ± 4.2 pixels and 26.6 ± 4.0 pixels, respectively (Fig. 3, top panel, right). Within the AMD group, ANTs and OAR performed the best, with a mean Laplace difference of 5.5 ± 2.7 pixels and 8.1 ± 4.3 pixels, respectively. The mean Laplace difference for the FLIRT, ITK and TOADS algorithm was 16.2 ± 7.4 pixels, 14.7 ± 8.0 pixels, and 15.3 ± 8.5 pixels, respectively. For the AMD group, the mean Laplace difference for all five algorithms was statistically smaller than without registration (p < 0.05). Within the control group, ANTs, and OAR performed the best, with a mean Laplace difference of 4.3 ± 1.4 pixels and 6.5 ± 2.2 pixels, respectively. The mean Laplace difference for the FLIRT, ITK and TOADS algorithm was 12.9 ± 6.1 pixels, 11.8 ± 5.3 pixels, and 11.8 ± 4.7 pixels, respectively. Similarly for the control group, the mean Laplace difference for all five algorithms was statistically smaller than without registration (p < 0.05).

Fig. 3
figure 3

Mean of the Laplace difference of the surface map over all OCT volumes for each of the registration algorithms for control and AMD patients (top). The bottom panel only included OCT volumes with 61 B-Scans. Number above each box is the p-value from a paired t-test. (The circles in the figure are outliers and are discussed at the end of the paper)

The data set used in this paper contained OCT volumes with varying B-scans densities, from 19 to 61 line scans. The higher the number of B-scans within an OCT volume, the more information it contains. Hence, we performed the same quantitative analysis, as above, to validate our approach specifically for high-density OCT images (OCT volumes with 61 B-scans) that are typically used in clinical practice or research (Fig. 3, bottom panel).

When only OCT volumes with 61 B-scans were included for analysis, the mean Laplace difference for the AMD and control groups without alignment was 23.1 ± 2.4 pixels and 25.1 ± 3.2 pixels, respectively. Within the AMD group, ANTs and OAR performed the best, with a mean Laplace difference of 3.5 ± 0.8 pixels and 4.2 ± 0.9 pixels, respectively. The mean Laplace difference for the FLIRT, ITK and TOADs algorithm was 11.7 ± 3.8 pixels, 9.6 ± 3.3 pixels, and 9.2 ± 2.3 pixels, respectively. For the AMD group, the mean Laplace difference for ANTs, OAR and TOADS was statistically smaller than without registration (p < 0.05). Within the control group, ANTs and OAR performed the best, with a mean Laplace difference of 4.0 ± 1.5 pixels and 6.0 ± 2.4 pixels, respectively. The mean Laplace difference for the FLIRT, ITK and TOADS algorithm was 12.4 ± 2.5 pixels, 9.6 ± 3.9 pixels, and 11.7 ± 2.7 pixels, respectively. For the control group, the mean Laplace difference for all five algorithms was statistically smaller than without registration (p < 0.05).

Figure 4 shows the edge errors for each algorithm for both the AMD and control groups. The top panel of Fig. 4 shows the results when all OCT volumes were included. The bottom panel of Fig. 4 shows the results when only OCT volumes with 61 B-scans were included. Overall, TOADS performed the best. For TOADS, the mean number of edge errors was 3586 ± 812, 3075 ± 829, 848 ± 346, and 701 ± 418 for the AMD (all OCT volumes), control (all OCT volumes), AMD (OCT with 61 B-scans only) and control (OCT with 61 B-scans only) group, respectively.

Fig. 4
figure 4

Mean number of edge errors over all OCT volumes for each of the registration algorithms for control and AMD patients (top). The bottom panel only included OCT volumes with 61 B-Scans

Surface map validation results

To validate the accuracy of our method for automatic surface map generation, we calculated the difference between manually generated surface maps and automatically generated surface maps, using B-scan images from 6 eyes. The difference was represented by a histogram and measured in pixels (Fig. 5). The mean and standard deviation in pixels of the difference between the six sets of surface maps was 0.28 ± 0.1, 1.4 ± 1.9, 0.45 ± 0.61, 0.5 ± 0.44, 0.99 ± 1.18, 1.21 ± 2.14, respectively.

Fig. 5
figure 5

(Left) visualization and comparisons of the surface maps created manually and automatically. (Right) Histograms showing the difference between each pair of surface maps in the number of pixels

Algorithm speed

The performance was quantified by the mean time (in milliseconds) to register a pair of B-scans images over all B-scans within the same OCT volume (Fig. 6). The fastest algorithm was OAR at approximately 2500 ms per pair. The slowest algorithm was FLIRT at approximately 5,500 ms.

Fig. 6
figure 6

Average registration time in milliseconds per pair of B-scans for each algorithm

Classification performance with and without alignment

We trained a 3D CNN to distinguish between normal and AMD OCT volumes. When our model was trained with B-scans aligned with ANTs, our model demonstrated superior performance (AUC 0.95 aligned vs. 0.89 unaligned; Table 1).

Table 1 Comparison of 3D CNN model performance with and without B-scan alignment in distinguishing between normal and AMD OCT volumes

Discussion

In this paper, we compared five commonly used alignment algorithms to register B-scans within an OCT volume, and tested our approach on a publicly-available dataset that contained both AMD and control patients. Overall, all five algorithms improved alignment, as measured by the mean Laplace difference of the en face surface map. Considering the results of all experiments, ANTs performed the best across all groups (AMD vs. control, all OCT volumes vs. OCT volumes with 61 B-scans only) in terms of producing the smoothest surface maps. In addition, we demonstrated that training a 3D CNN with B-scans aligned by ANTs produced a more robust model in detecting AMD on OCT volumes, as compared to training with unaligned B-scans.

Although ANTs did not produce the smallest number of edge errors, the proportion of edge errors as a mean percentage of total number pixels was still low at 0.04%. Hence, we believe that of the commonly available image alignment algorithms, ANTs should be used to register non-aligned B-scans within the same OCT volume. In image alignment, there are two options: rigid body transformation and nonlinear transformation [41]. For this paper, we only used rigid alignment methods, as many macular diseases, such as neovascular age-related macular degeneration and diabetic macular edema, manifest as perturbation of the retinal laminations on OCT imaging. We specifically avoided using non-linear warping techniques, as non-linear warping could inadvertently produce perturbation of the retinal layers and introduce artifactual pathologic features.

The number of B-scans within the OCT volumes varied from 19 to 61 for the dataset we used. That is, with the same scan area, the number of cross-sectional images varied from 19 to 61, and the distance between individual B-scans was inversely related to the number (density) of B-scans. The strength of our paper lies in the validation of our approaches using OCT volumes with different B-scan densities, as in real-world clinical practice, the B-scan density is often variable and determined by operator or scanning protocol preferences. In addition, the algorithms investigated in our study are already readily available to the public, so our findings are relevant to the broader research community.

Our manuscript was motivated by the fact that when OCT volumes are exported from commercially-available viewing software, an OCT volume is typically not exported en bloc but as a series of separate B-scans. However, 3D CNNs, one of the cutting-edge model architectures, require training data input in the form of 3D image volumes. Meaning, after the initial data export, the separate OCT B-scans will have to be re-combined into a 3D volume, before a 3D CNN can be trained. However, in the process of re-combining the B-scans, artifacts could be introduced if B-scan alignment is not performed. For example, if the retinal pigment epithelium (RPE) is not aligned between adjacent B-scans, an artifactual “RPE break” could be introduced. To demonstrate the utility of aligning OCT B-scans, specifically within the context of training 3D CNN models for DL-based image analysis, we trained a model to detect AMD in 3D OCT volumes. We compared the performance of our model, with and without B-scan alignment, and demonstrated that B-scan alignment produced a more robust model. Our data provides proof-of-concept evidence that aligning B-scans could be broadly useful for training any 3D CNNs that involve 3D macular OCT volumes.

Limitations

Our paper has several limitations. First, none of the five algorithms included in this study were originally developed for OCT B-scan registration, so an algorithm designed from the ground up for this specific purpose may yield even better results. Second, the range of pathologies included in this OCT dataset was limited, as only control (normal) and AMD patients were included. Specifically, drusen, defining features of AMD, are located in the outer retina and do not significantly perturb the smoothness of the inner retinal surface. For conditions that introduce significant inner retinal surface perturbation, such as trauma and localized irregular epiretinal membrane, our novel metric of surface smoothness may not be applicable.

Conclusions

In this paper, we introduced a novel metric to quantify OCT B-scan alignment and compared the effectiveness of five alignment algorithms. We confirmed that alignment could be improved in a statistically significant manner with readily available alignment algorithms that are available to the public, and the ANTs algorithm provided the most robust performance overall.

Availability of data and materials

The datasets used and/or analysed during the current study are available from the corresponding author on reasonable request.

Abbreviations

OCT:

Optical coherence tomography

AMD:

Age-related macular degeneration

ANTs-SyN:

Advanced normalization tools symmetric image normalization

MRI:

Magnetic resonance imaging

DL:

Deep learning;

FLIRT:

FMRIB’s linear image registration tool

ITK:

Insight toolkit

OAR:

Optimized automatic registration

TOADS:

TOADS-CRUISE

References

  1. Brenner DJ, Hall EJ. Computed tomography–an increasing source of radiation exposure. N Engl J Med. 2007;357(22):2277–84.

    Article  CAS  PubMed  Google Scholar 

  2. Huettel SA, Song AW, McCarthy G. Functional magnetic resonance imaging, vol. 1. Sunderland: Sinauer Associates; 2004.

    Google Scholar 

  3. Huang D, Swanson EA, Lin CP, Schuman JS, Stinson WG, Chang W, et al. Optical coherence tomography. Science. 1991;254(5035):1178–81.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  4. Budenz DL, Anderson DR, Varma R, Schuman J, Cantor L, Savell J, et al. Determinants of normal retinal nerve fiber layer thickness measured by stratus OCT. Ophthalmology. 2007;114(6):1046–52.

    Article  PubMed  Google Scholar 

  5. Fujimoto JG, Brezinski ME, Tearney GJ, Boppart SA, Bouma B, Hee MR, et al. Optical biopsy and imaging using optical coherence tomography. Nat Med. 1995;1(9):970–2.

    Article  CAS  PubMed  Google Scholar 

  6. Jia Y, Bailey ST, Wilson DJ, Tan O, Klein ML, Flaxel CJ, et al. Quantitative optical coherence tomography angiography of choroidal neovascularization in age-related macular degeneration. Ophthalmology. 2014;121(7):1435–44.

    Article  PubMed  Google Scholar 

  7. Bressler NM. Age-related macular degeneration is the leading cause of blindness. JAMA. 2004;291(15):1900–1.

    Article  CAS  PubMed  Google Scholar 

  8. Lee CS, Baughman DM, Lee AY. Deep learning is effective for the classification of OCT images of normal versus age-related macular degeneration. Ophthalmol Retina. 2017;1(4):322–327.

    Article  Google Scholar 

  9. Li F, Chen H, Liu Z, Zhang X, Wu Z. Fully automated detection of retinal disorders by image-based deep learning. Graefes Arch Clin Exp Ophthalmol. 2019;257(3):495–505.

    Article  PubMed  Google Scholar 

  10. Greve DN, Fischl B. Accurate and robust brain image alignment using boundary-based registration. Neuroimage. 2009;48(1):63–72.

    Article  PubMed  Google Scholar 

  11. Baker S, Matthews I. Equivalence and efficiency of image alignment algorithms. In: Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR 2001. ieeexplore.ieee.org; 2001. p. I–I.

  12. Studholme C, Hill DLG, Hawkes DJ. An overlap invariant entropy measure of 3D medical image alignment. Pattern Recognit. 1999;32(1):71–86.

    Article  Google Scholar 

  13. Oliveira FPM, Tavares JMRS. Medical image registration: a review. Comput Methods Biomech Biomed Engin. 2014;17(2):73–93.

    Article  PubMed  Google Scholar 

  14. Maintz JB, Viergever MA. A survey of medical image registration. Med Image Anal. 1998;2(1):1–36.

    Article  CAS  PubMed  Google Scholar 

  15. Hill DL, Batchelor PG, Holden M, Hawkes DJ. Medical image registration. Phys Med Biol. 2001;46(3):R1-45.

    Article  CAS  PubMed  Google Scholar 

  16. Avants BB, Tustison NJ, Song G, Cook PA, Klein A, Gee JC. A reproducible evaluation of ANTs similarity metric performance in brain image registration. Neuroimage. 2011;54(3):2033–44.

    Article  PubMed  Google Scholar 

  17. Smith SM, Jenkinson M, Woolrich MW, Beckmann CF, Behrens TEJ, Johansen-Berg H, et al. Advances in functional and structural MR image analysis and implementation as FSL. Neuroimage. 2004;23(Suppl 1):S208–19.

    Article  PubMed  Google Scholar 

  18. Ibanez L, Schroeder W, Ng L, Cates J. The ITK software guide. 2003; http://insight-journal.org/dspace/handle/1926/388

  19. Jenkinson M, Smith S. A global optimisation method for robust affine registration of brain images. Med Image Anal. 2001;5(2):143–56.

    Article  CAS  PubMed  Google Scholar 

  20. Vercauteren T, Pennec X, Perchant A, Ayache N. Diffeomorphic demons: efficient non-parametric image registration. Neuroimage. 2009;45(1 Suppl):S61-72.

    Article  PubMed  Google Scholar 

  21. Rasti R, Rabbani H, Mehridehnavi A, Hajizadeh F. Macular OCT classification using a multi-scale convolutional neural network ensemble. IEEE Trans Med Imaging. 2018;37(4):1024–34.

    Article  PubMed  Google Scholar 

  22. Farsiu S, Chiu SJ, O’Connell RV, Folgar FA, Yuan E, Izatt JA, et al. Quantitative classification of eyes with and without intermediate age-related macular degeneration using optical coherence tomography. Ophthalmology. 2014;121(1):162–72.

    Article  PubMed  Google Scholar 

  23. Avants BB, Duda JT, Zhang H, Gee JC. Multivariate normalization with symmetric diffeomorphisms for multivariate studies. Med Image Comput Comput Assist Interv. 2007;10(Pt 1):359–66.

    CAS  PubMed  Google Scholar 

  24. Jenkinson M, Bannister P, Brady M, Smith S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. Neuroimage. 2002;17(2):825–41.

    Article  PubMed  Google Scholar 

  25. Chen M, Lang A, Ying HS, Calabresi PA, Prince JL, Carass A. Analysis of macular OCT images using deformable registration. Biomed Opt Express. 2014;5(7):2196–214.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Comaniciu D, Meer P. Mean shift: a robust approach toward feature space analysis. IEEE Trans Pattern Anal Mach Intell. 2002;24(5):603–19.

    Article  Google Scholar 

  27. Cleveland WS. Robust locally weighted regression and smoothing scatterplots. J Am Stat Assoc. 1979;74(368):829–36.

    Article  Google Scholar 

  28. Wang X. Laplacian operator-based edge detectors. IEEE Trans Pattern Anal Mach Intell. 2007;29(5):886–90.

    Article  PubMed  Google Scholar 

  29. Tran, Bourdev, Fergus. Learning spatiotemporal features with 3d convolutional networks. Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2015, pp. 4489–4497

  30. Avants BB, Tustison N, Song G. Advanced normalization tools (ANTS). Insight J. 2009;2(365):1–35.

    Google Scholar 

  31. Tustison NJ, Cook PA, Klein A, Song G, Das SR, Duda JT, et al. Large-scale evaluation of ANTs and FreeSurfer cortical thickness measurements. Neuroimage. 2014;1(99):166–79.

    Article  Google Scholar 

  32. Modersitzki J. FLIRT with rigidity—image registration with a local non-rigidity penalty. Int J Comput Vis. 2008;76(2):153–63.

    Article  Google Scholar 

  33. Fischer B, Modersitzki J. FLIRT: a flexible image registration toolbox. In: Gee James C, Antoine Maintz JB, Vannier Michael W, editors. Biomedical image registration. Berlin: Springer; 2003.

    Google Scholar 

  34. Klein A, Andersson J, Ardekani BA, Ashburner J, Avants B, Chiang M-C, et al. Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. Neuroimage. 2009;46(3):786–802.

    Article  PubMed  Google Scholar 

  35. Plumat J, Andersson M, Janssens G, Orban de Xivry J, Knutsson H, Macq B. Image registration using Morphon algorithm: an ITK implementation. Insight J. 2009. https://doi.org/10.5429/kmysv2.

    Article  Google Scholar 

  36. Martinez-Perez M, Hughes AD, Thom SA, Parker KH. Improvement of a retinal blood vessel segmentation method using the insight segmentation and registration toolkit (ITK). Conf Proc IEEE Eng Med Biol Soc. 2007;2007:892–5.

    CAS  Google Scholar 

  37. Avants BB, Tustison NJ, Stauffer M, Song G, Wu B, Gee JC. The insight ToolKit image registration framework. Front Neuroinform. 2014;28(8):44.

    Google Scholar 

  38. Mahmoudzadeh AP, Kashou NH. Evaluation of interpolation effects on upsampling and accuracy of cost functions-based optimized automatic image registration. Int J Biomed Imaging. 2013;1(2013):395915.

    Google Scholar 

  39. Bazin P-L, Pham DL. Topology-preserving tissue classification of magnetic resonance brain images. IEEE Trans Med Imaging. 2007;26(4):487–96.

    Article  PubMed  Google Scholar 

  40. Sweeney EM, Vogelstein JT, Cuzzocreo JL, Calabresi PA, Reich DS, Crainiceanu CM, et al. A comparison of supervised machine learning algorithms and feature vectors for MS lesion segmentation using multimodal structural MRI. PLoS ONE. 2014;9(4):e95753.

    Article  PubMed  PubMed Central  Google Scholar 

  41. Woods RP, Grafton ST, Watson JD, Sicotte NL, Mazziotta JC. Automated image registration: II. Intersubject validation of linear and nonlinear models. J Comput Assist Tomogr. 1998;22(1):153–65.

    Article  CAS  PubMed  Google Scholar 

Download references

Acknowledgements

Not applicable.

Funding

Not applicable.

Author information

Authors and Affiliations

Authors

Contributions

Experimental design: BL, CKJ and TYAL. Data collection and analysis: BL, CKJ and TYAL. Manuscript drafting and revision: BL, CKJ, JHW, TN, PX and TYAL.

Corresponding author

Correspondence to T. Y. Alvin Liu.

Ethics declarations

Ethics approval and consent to participate

Ethical approval not applicable due to the retrospective nature of the current study. Informed consent was obtained from all participants during initial data collection.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jones, C.K., Li, B., Wu, JH. et al. Comparative analysis of alignment algorithms for macular optical coherence tomography imaging. Int J Retin Vitr 9, 60 (2023). https://doi.org/10.1186/s40942-023-00497-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40942-023-00497-2

Keywords