Abstract
Image registration between planning CT images and cone beam-CT (CBCT) images is one of the key technologies of image guided radiotherapy (IGRT). Current image registration methods fall roughly into two categories: geometric features-based and image grayscale-based. Mutual information (MI) based registration, which belongs to the latter category, has been widely applied to multi-modal and mono-modal image registration. However, the standard mutual information method only focuses on the image intensity information and overlooks spatial information, leading to the instability of intensity interpolation. Due to its use of positional information, wavelet transform has been applied to image registration recently. In this study, we proposed an approach to setup CT and cone beam-CT (CBCT) image registration in radiotherapy based on the combination of mutual information (MI) and stationary wavelet transform (SWT). Firstly, SWT was applied to generate gradient images and low frequency components produced in various levels of image decomposition were eliminated. Then inverse SWT was performed on the remaining frequency components. Lastly, the rigid registration of gradient images and original images was implemented using a weighting function with the normalized mutual information (NMI) being the similarity measure, which compensates for the lack of spatial information in mutual information based image registration. Our experiment results showed that the proposed method was highly accurate and robust, and indicated a significant clinical potential in improving the accuracy of target localization in image guided radiotherapy (IGRT).
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Background
To achieve the best therapeutic outcome, modern radiotherapy has attempted a variety of ways to maximize the damage to the tumor while sparing surrounding normal tissues [1, 3–5]. The accurate targeting of tumor has been playing an important role in the implementation of successful radiotherapy, which introduced the concept of image guided radiotherapy (IGRT). As one of the key steps in targeting tumor, the registration between planning CT images and CBCT (Cone beam-CT) images has been widely explored and its techniques have been improved significantly since the advent of CBCT.
Image registration techniques generally rely on the information of the images themselves. Current image registration methods fall roughly into two categories: geometric features-based and image grayscale-based [2]. To date, these image registration methods have been widely used to perform the registration between planning CT and CBCT images. Based on the deformation with intensity simultaneously corrected, a CT to CBCT deformable registration approach was proved to be robust against the CBCT artifacts and intensity inconsistency [3]. Free-form deformable registration algorithm, which resulted in a high correlation between CBCT and the new planning CT, was also successfully conducted [4]. Multiscale registration, which decomposed the registering images into a series of scales and registered the coarser scales of the images iteratively, was regarded as an effective method for the registration between CT and daily CBCT images [5]. Registration techniques based on mutual information (MI) belong to the image grayscale-based registration method and have been widely applied to multi-modal and mono-modal image registration tasks. A multi-modal retinal image registration, which was based on improved mutual information using adaptive probability density estimation, resulted in high accuracy and efficiency [6]. Three–dimensional registration techniques based on mutual information could be also applied to the alignment of brain tissues in magnetic resonance imaging time-series or PET [7, 8]. A comparison between standard mutual information and normalized mutual information indicated that normalized mutual information is more stable and robust in that it is immune to the variation of entropy [9]. The application of mutual information is a very effective strategy for image registration, but the traditional mutual information method only focuses on the image intensity information, with spatial information neglected, which leads to the instability to intensity interpolation [10]. With regard to registration of medical images, spatial information is very important and should be incorporated into grayscale-based based registration algorithms. A 3D-2D registration of CT and X-ray images incorporated the spatial information in a variational approximation and obtained a high registration accuracy [11]. Positions with large gradient usually correspond to tissue transition, which provides spatial information [12]. Therefore, wavelet transform was recently applied to image registration [13, 14]. Daubechies complex wavelet transform, which is shift invariant and provides phase information, was successfully used to achieve the fusion of multimodal medical images [15]. A flexible multiscale and shift-invariant representation of registered images was firstly obtained by using stationary wavelet transform, and then the registration through pulse-coupled neural network was performed on the new representation [16, 17].
Some studies incorporated the gradient information of medical images in the mutual information to compensate for the lack of spatial information [18, 19]. These methods produced gradient images using the sum of the squares of the corresponding sub-band coefficients. However, gradient images generated this way lack diagonal components, leading to the loss of edge information and low registration accuracy. In registrations between planning CT images and CBCT images, the noise in CBCT images usually results in poor resolution in low contrast areas of the images and blurs edges of images.
In this paper, we proposed a registration method based on stationary wavelet transform (SWT) with translational invariance. The translational invariance of the stationary wavelet is conducive to highlighting edge features of an image and improves the registration accuracy. Experiments showed that our algorithm is robust.
Materials and methods
Materials
For planning CT images and setup CBCT images in radiotherapy, Siemens large aperture CT and Varian Rapid Arc CBCT are used for image registration. The image parameters for CT are as follows: image matrix 512 × 512; pixels size 1.17 × 1.17 mm; the image parameters for CBCT are as follows: image matrix 384 × 384; pixel size 1.27 × 1.27 mm. All the participants gave their informed consent and the Ethics Committee of Beijing Xuanwu Hospital Affiliated to Capital Medical University approved the protocol of this study.
Methods
In our proposed method, the reference image and floating images were decomposed with three levels using stationary wavelet transform. Low-frequency components of wavelet produced in all levels during the decomposition were set to zero, and a gradient image of the original image was obtained by performing inverse wavelet transform on remaining high-frequency components. Then, the mutual information of the original image and the gradient image was calculated by using the normalized mutual information as the similarity measure. Finally, a new similarity measure was synthesized with a weighting function. The Powell algorithm was used for multi-parameter optimization to produce the final spatial transformation parameters for the image registration.
Stationary wavelet transform of image
Nason and Silverman introduced the stationary wavelet transform in 1995 [20]. In contrast to orthogonal wavelets, stationary wavelet, also known as non-sampling wavelet transform, has the properties of redundancy, translational invariance, capability of providing more approximate estimation of continuous wavelet transform. As an effective mathematical tool for edge detection [21–24], its advantages include the local time-frequency characteristics and multi-resolution analysis capability of wavelet transform. The jth-level decomposition of SWT is shown in Fig. 1.
The decomposition formulas of SWT are as follows:
where \( {A}_{j,{k}_1,{k}_2} \), \( {D}_{j,{k}_1,{k}_2}^1 \), \( {D}_{j,{k}_1,{k}_2}^2 \), \( {D}_{j,{k}_1,{k}_2}^3 \) are the low frequency components (LL), the horizontal high-frequency component (LH), vertical high-frequency component (HL) and diagonal components (HH) of the stationary wavelet transform, respectively. \( {h}_0^{\uparrow {2}^j} \) and \( {g}_0^{\uparrow {2}^j} \) are used to denote that 2j-1 zeros are inserted between the two points h 0 and g 0. The corresponding reconstruction algorithm (IDSWT) is
The components of the image after the SWT are shown in Fig. 2.
Synthesized gradient image based on stationary wavelet transform
The high-frequency sub-band of wavelet transform has the ability to highlight the differences between neighboring pixels in an image [25]. Large wavelet coefficient indicates the boundary of two distinct intensity regions in the original image. Stationary wavelet transform is translationally invariant, which helps to identify the image edge features. In order to improve the resolution of edge details, image with prominent edge features can be reconstructed by using the inverse SWT with the three groups of wavelet vectors (LH, HL, HH). The CBCT image and the gradient image generated with SWT are shown in Fig. 3.
Similarity measure
In this paper, mutual information is used as the similarity measure. As a basic concept of information theory, mutual information is generally used to describe the statistical correlation between two systems, or the amount of information of a system contained in another system. In multimodality image registration, when the spatial positions of two images are completely consistent, the mutual information, i.e., the information of one image expressed by another image, is maximum. The mutual information can be expressed by entropy which describes the complexity or uncertainty of a system. The entropy of the system A is defined as
The joint entropy of two systems is defined as
where a ∈ A, b ∈ B, and p A (a) is the marginal probability density function, p A,B (a, b) is the joint probability density function. The mutual information between the two systems is thus expressed as
Mutual information is sensitive to the amount of overlap between images and normalized mutual information (NMI) has been proposed to overcome this problem. It is defined as
Registration method based on stationary wavelet transform
We denoted the normalized mutual information of the original image as NMIi and that of the synthesized gradient image based on the SWT as NMIg. They were calculated using Eq. (1). Jiangang Liu merged original images with gradient information using the method described below, and the new similarity measure was given by [24]:
where
The weighting function f(v) is shown in Fig. 4, where T is a time constant used to control the shape of f(v). The weighting function is essentially a logistic function, which has the properties of saturation, differentiability and nonlinearity. It also has a maximum and a minimum. f(v) is an ideal weighting function for merging registration function. According to our experience, T = 0.04 is an appropriate value.
Our method follows this procedure:
-
Step 1: decompose the reference image R and floating image F respectively using stationary wavelet transform;
-
Step 2: assign the coefficients of low-frequency in all levels of stationary wavelet decomposition to zero and perform inverse transform to reconstruct corresponding gradient images;
-
Step 3: combine the original image and the synthesized gradient image using the weighting function to form a new similarity measure;
-
Step 4: use the Powell multi-parameter optimization algorithm to optimize the space transformation parameters (Δx, Δy, θ) and the registration is completed.
Optimization algorithm
Two-dimensional image registration is essentially a multi-parameter optimization problem, namely, searching three optimal registration parameters (two translational parameters and a rotational parameter) to maximize the mutual information. In this paper, the Powell multi-parameter optimization algorithm and the Brent one-dimensional search algorithm are used to optimize the parameters.
Results
Our algorithm was implemented in Matlab R2008a. We selected ten medical images as the reference images, and the floating images were generated with spatial transformation of the corresponding reference images. As the preset transformation parameters in X and Y directions and the rotation angle θ (As shown in Table 1) were known, judgment of the correctness and registration accuracy of the algorithm was straightforward. The smaller the gray level difference between the image F after registration and the reference image R is, the higher the registration accuracy is.
Root mean square error (MSE), which is defined as follows, is employed as the registration error [26–29].
The smaller the values of MSE are, the higher the registration accuracy is. If two images are identical, the MSE = 0. We took ten images with preset transformation parameters as the experiment data, the experiment results of which are shown in Table 1.
Judged from the above experiment results, the proposed registration method is accurate (on the order of subpixel) and robust. However, compared with mutual information registration, the proposed method is time-consuming, because it needs to calculate not only the mutual information between the original images, but also the mutual information with the gradient images, which increases the computation load.
The CBCT image is shown in Fig. 3a. Because of smaller image matrix and larger pixel size, CBCT images must be up-sampled to the size of the CT image before the registration. Due to a considerable difference between our CT and CBCT images, the initially linear shifting (80 and 80 mm along X and Y directions, respectively) of the CBCT images was performed to manually narrow the difference for the reduction of latter registration time. In order to test the CT and CBCT images registration algorithm, we generated ten images by performing manual spatial transformation to CBCT images in advance (as shown in Table 2). Among the ten images, five images were transformed in Y direction linearly, and other five images were transformed linearly in X direction. They were then registered with the corresponding layer of CT images. If the corresponding translation term of the registration is linear, and the rotation angle is close to 0°, high registration accuracy is indicated. The registration results are shown in Table 2.
In order to compare our proposed method with standard mutual information, the linearity of the transforming variation along X and Y directions was represented in Fig. 5
As can be seen from Table 2 and Fig. 5, with our proposed method, the translations in the Y direction in the first five images, and in the X direction in the last five images appear linear. Therefore, the proposed method is more robust than the traditional mutual information registration. The registration results of CBCT to CT images are shown in Fig. 6.
Discussion
Based on mutual information and stationary wavelets transform, the registration of CT and manually transformed CT images resulted in a lower MMSE compared with standard mutual information, which indicated that the algorithm we proposed was more accurate. The registration results of CT and CBCT images showed that the transformation parameters of our registration method was more linearly related to the preprocessing parameters along the corresponding directions, which indicated that our method was also more robust. The stationary wavelet transform can be applied to obtain the spatial information of the registered images and the normalized mutual information can be used as the similarity measure in the registration. The combination of the above techniques yielded an effective registration of CT and CBCT, which is indispensable for the accurate location of tumor in radiotherapy. In our future investigation, we will focus on improving the speed of the algorithm. In addition, compared with other methods, the proposed registration algorithm calculated the mutual information of the original image and the gradient image respectively, which increases the computational cost. Shortening the algorithm running time will be the focus of our further research.
There are still some limits in our work. Firstly, we just tried to use the gradient information to investigate the effects of registration between CT and CBCT images, so compared with other state-of-art method performing on multimodal image registration (such as registration of CT, MRI, PET images, etc. ) [30–34], our proposed method may not perform so well. In our future work, we will continue to improve our method and apply it on multimodal images for a further evolution.
Conclusions
In this paper, we proposed a medical image registration algorithm based on SWT and mutual information. The algorithm synthesizes a gradient image based on the translational invariance of SWT, and incorporates it into the mutual information calculation of the original image by the weighting function to obtain a new similarity measure. The proposed method effectively overcomes the weakness of mutual information registration for the lack of spatial information. Experiment results showed that the proposed method is robust and accurate. As for the registration between planning CT images and setup CBCT images in radiotherapy, SWT is data redundanct and translationally invariant, which is conducive to identify sharp variations in the image. Furthermore, image reconstruction based on SWT tends to highlight edge features, and enhances the resolution of edge details. In particular, for noisy CBCT images, we can extract more accurate gradient information from the images, thereby the accuracy of the registration can be improved.
Abbreviations
- CBCT:
-
Cone beam-CT
- IGRT:
-
Image guided radiotherapy technology
- SWT:
-
Stationary wavelet transform
- NMI:
-
Normalized mutual information
- 3D CRT:
-
Three-dimensional conformal radiotherapy
- IMRT:
-
Intensity modulated radiation therapy
- MI:
-
Mutual information
References
Faisal, A., Parveen, S., Badsha, S., Sarwar, H., and Reza, A. W., Computer assisted diagnostic systems in tumor radiography. J. Med. Syst. 37(3):9938, 2013.
Wang, X. H., Istepanian, R. S., and Song, Y. H., Microarray image enhancement by denoising using Stationary Wavelet Transform. IEEE Trans. Nanobiosci. 2(4):184–190, 2003.
Zhen, X., Gu, X. J., Yan, H., Zhou, L. H., Jia, X., and Jiang, S. B., CT to cone-beam CT deformable registration with simultaaneous intensity correction. Phys. Med. Biol. 57(21):6807–6826, 2012.
Piper, J. W., Evaluation of A CT to cone-beam CT deformable registration algorithm. IJROBP 69(3):418–419, 2007.
Paquin, D., Levy, D., and Xing, L., Multiscale registration of planning CT and daily cone beam CT images for adaptive radiation therapy. Med. Phys. 36(1):4–11, 2009.
Legg, P. A., Rosin, P. L., Marshall, D., and Morgan, J. E., Improving accuracy and efficiency of mutual information for multi-modal retinal image registration using adaptive probability density estimation. Comput Med Imaging Graph 37(7–8):597–606, 2013.
Liu, B., Bai, X. Z., Zhou, F. G., Han, H. B., and Hou, C., Mutual information based three-dimensional registration of rat brain magnetic resonance imaging time-series. Comput. Electr. Eng. 39(5):1473–1484, 2013.
Sandiego, C. M., Weinzimmer, D., and Carson, R. E., Optimization of PET-MR registrations for nonhuman primates using mutual information measures: A Multi-Transform Method (MTM). NeuroImage 64(1):571–581, 2013.
Gao, Z. Y., Gu, B., and Lin, J. R., Monomodal image registration using mutual information based methods. Image Vis. Comput. 26(2):164–173, 2008.
Maintz, P. J., and Viergever, M. A., Mutual information-based registration of medical images: A survey. IEEE Trans. Med. Imaging 22(8):986–1004, 2003.
Zheng, G. Y., Effective incorporating spatial information in mutual information based 3D-2Dregistration of a CT volume to X-ray images. Medical Image Computing and Computer-Assisted Intervention-MICCAI 2008. Lect. Notes Comput. Sci 5242:922–929, 2008.
Tsao, J., Interpolation artifacts in multimodality image registration based on maximization of mutual information. IEEE Trans. Med. Imaging 22(7):235–260, 2003.
Mallat, S. G., and Hwang, W. H., Singularity detection and processing with wavelets. IEEE Trans. Inf. Theory 38(2):617–643, 1992.
Mallat, S., and Zhong, S., Characterization of signals from multiscale edges. IEEE Trans. Pattern Anal. Mach. Intell. 14(7):710–732, 1992.
Singh, R., and Khare, A., Fusion of multimodal medical images using Daubechies complex wavelet transform–A multiresolution approach. Inf. Fusion. 19:49–60, 2014.
Chai, Y., Li, H. F., and Qu, J. F., Image fusion scheme using a novel dual-channel PCNN in lifting stationary wavelet domain. Opt. Commun. 283(19):3591–3602, 2010.
Chai, Y., Li, H. F., and Guo, M. Y., Multifocus image fusion scheme based on features of multiscale products and PCNN in lifting stationary wavelet domain. Opt. Commun. 284(5):1146–1158, 2011.
Maintz, J. B. A., van den Elsen, P. A., and Viergever, M. A., Comparison of edge-based and ridgebased registration of CT and MR brain images. Med. Image Anal. 2(1):151–161, 1996.
Pluim, J. P. W., Maintz, J. B. A., and Viergever, M. A., Image registration by maximization of combined mutual information and gradient information. IEEE Trans. Med. Imaging 19(2):809–814, 2000.
Nason, G. P., and Silverman, B. W., The stationary wavelet transform and some statistical applications in wavelet and statistics. In: Antoniadis, A. (Ed.), Lecture Notes in Statistics. Spinger Verlag, Berlin, pp. 281–299, 1995.
Zikov, T., Bibian, S., Dumont, G. A., and Huzmezan, M., Quantifying cortical activity during general anesthesia using wavelet analysis. IEEE Trans. Biomed. Eng. 53(4):617–632, 2006.
Galiana-Merino, J. J., Rosa-Herranz, J. L., and Seismic, S. P., Phase picking using a Kurtosis-Based criterion in the Stationary Wavelet Domain. IEEE Trans. Geosci. Remote Sens. 46(11):3815–3826, 2008.
Demirel, H., and Anbarjafari, G., Image resolution enhancement by using discrete and stationary wavelet decomposition. IEEE Trans. Image Process. 20(5):1458–1460, 2011.
Liu, J. G, Tian, J., Multi-modal Medical Image Registration Based on Adaptive Combination of Intensity and Gradient Field Mutual Information. In Proceedings of the 28th IEEE EMBS Annual International Conference. pp. 1429–1432, New York City, USA: IEEE Computer Society Press, 2006.
Avci, D., Leblebicioglu, M. K., Poyraz, M., and Esin, D., A new method based on adaptive discrete wavelet entropy energy and neural network classier (ADWEENN) for recognition of urine cells from microscopic images independent of rotation and scalling. J. Med. Syst. 38(2):7, 2014.
Tang, M., and Chen, F., A qualitative meta analysis review on medical image registration evalution. Procedia Eng. 29:499–503, 2012.
Ito, K., Suzuki, A., Shima, S. N., Aoki, T. Performance evaluation using mandelbrot images. In Proceeding of 16th IEEE International Conference on Image processing, pp. 4333–4336, 2009.
Lu, X., Zhang, S., Su, H., and Chen, Y., Mutual information-based multimodal image registration using a novel joint histogram estimation.Comput. Med. Imaging Graph. 32(3):202–209, 2008.
Nejati, M., and Pourghassem, H., Multiresolution image registration in digital X-Ray angiography with intensity variation modeling. J. Med. Syst. 38(2):10, 2014.
Pluim, J., Maintz, J., and Viergever, M., Image registration by maximization of combined mutual information and gradient information. In Medical Image Computing and Computer-Assisted Intervention – MICCAI, pp. 103–129. Springer, 2000.
Pluim, J., Maintz, J., and Viergever, M., Image registration by maximization of combined mutual information and gradient information. IEEE Trans. Med. Imaging 19:809, 2000.
Heinrich, M. P., Jenkinson, M., Bhushan, M., Matin, T., Gleeson, F. V., Brady, S. M., and Schnabel, J. A., MIND: Modality independent neighbourhood descriptor for multi-modal deformable registration. Med. Image Anal. 16:1423–1435, 2012.
Heinrich, M., Jenkinson, M., Brady, M., and Schnabel, J., Textural mutual information based on cluster trees for multimodal deformable registration. IEEE International Symposium on Biomedical Imaging, ISBI. IEEE, pp. 1471–1474. Barcelona, 2012b.
Park, S. B., Rhee, F. C., Monroe, J. I., and Sohn, J. W., Spatially weighted mutual information image registration for image guided radiation therapy. Med. Phys. 37:4590–4601, 2010.
Acknowledgments
This work was supported by National Natural Science Foundation of China (No. 30670576 ), Beijing Natural Science Foundation (No. 4122018 ) and Clinical Cooperative Research Program of Capital Medical University (Grant No.10JL24).
Competing interests
The authors declare that they have no competing interests.
Author contributions
All authors contributed to the work significantly. Deng Jun-min: Conceived the whole study scheme, algorithm and wrote the manuscript. Yue Hai-zhen: Developed the method and original software. Zhuo Zhi-zheng: Revised the manuscript and performed the literature search. Yan Hua-gang: Performed the data analyses and performed the literature search. Liu Di: Collected and selected the patient, planned the study protocol. Li Hai-yun: Conceived the whole study and reviewed the manuscript critically. All authors read and approved the manuscript.
Author information
Authors and Affiliations
Corresponding author
Additional information
This article is part of the Topical Collection on Patient Facing Systems
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution License which permits any use, distribution, and reproduction in any medium, provided the original author(s) and the source are credited.
About this article
Cite this article
Deng, Jm., Yue, Hz., Zhuo, Zz. et al. A Stationary Wavelet Transform Based Approach to Registration of Planning CT and Setup Cone beam-CT Images in Radiotherapy. J Med Syst 38, 40 (2014). https://doi.org/10.1007/s10916-014-0040-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10916-014-0040-2