| تعداد نشریات | 43 |
| تعداد شمارهها | 1,792 |
| تعداد مقالات | 14,626 |
| تعداد مشاهده مقاله | 38,951,902 |
| تعداد دریافت فایل اصل مقاله | 15,163,669 |
ادغام تصاویر ساختاری و عملکردی بر مبنای تحلیل کارتون-بافت و شبکه عصبی پالس کوپل شده | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| هوش محاسباتی در مهندسی برق | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| دوره 16، شماره 1، اردیبهشت 1404، صفحه 41-58 اصل مقاله (1.05 M) | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| نوع مقاله: مقاله پژوهشی انگلیسی | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| شناسه دیجیتال (DOI): 10.22108/isee.2025.140251.1671 | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| نویسندگان | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| روژین زمانزاده1؛ حمیدرضا شاهدوستی* 2 | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| 1دانشکده مهندسی برق، دانشگاه صنعتی همدان، همدان، ایران | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| 2دانشکده مهندسی برق، دانشگاه صنعتی همدان، همدان، ایران | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| چکیده | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ادغام تصاویر MRI و PET (یا SPECT) نمایش همزمان جزئیات ساختاری و عملکردی که توسط تصاویر اولیه ثبت شده است را امکان پذیر می کند. در این مقاله یک مدل جدید جهت ادغام تصاویر MRI و PET (یا SPECT) پیشنهاد میشود که در این روش ویژگی های استخراج شده توسط تجزیه کارتون-بافت به عنوان ورودی به شبکه عصبی پالس کوپل شده داده میشود. بدین منظور ابتدا تصاویر پزشکی به مولفه های کارتون و بافت تجزیه شده به طوری که مولفه کارتون شامل بخشهای تکه ای یکنواخت با مرزهای مشخص و مولفه بافت جزئیات تصویر شامل لبه ها و بافت را شامل میشود. در گام دوم مولفه های کارتون هر دو تصویر بر اساس مقایسه انرژی ضرایب آنها با یکدیگر ادغام شده و مولفه های بافت هر دو تصویر بر اساس اعمال کردن ضرایب EOG (انرژی گردایان تصویر) به شبکه عصبی پالس کوپل شده ادغام میشوند. آزمایشات با استفاده از چندین تصویر پزشکی مغز انسان نشانده کیفیت روش پیشنهادی بر مبنای ارزیابی کمی و بصری است. با بهره گیری از یک تجزیه موثر، روش پیشنهادی مرزهای مشخص و جزئیات واضح تصاویر اولیه را داراست درحالیکه از اعوجاج رنگ مصون مانده است. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| کلیدواژهها | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ادغام تصاویر پزشکی؛ شبکه عصبی پالس کوپل شده؛ تجزیه کارتون-بافت؛ تصویربرداری تشدید مغناطیسی | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| اصل مقاله | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
1. Introduction[1]Many health care applications, such as medical diagnostics, patient health monitoring, computer-aided surgery, and drug evaluation, require images with both functional and anatomical information. But any single modality of medical imaging cannot provide images with complete and accurate information about both the anatomy and function of organs. In fact, various types of medical imaging sensors provide different kinds of information that are complementary and redundant. Structural medical images, such as MRI (magnetic resonance imaging), ultrasound, and CT (computed tomography) images, represent the anatomy of organs, whereas functional medical images, such as PET (positron emission tomography) and SPECT (single-photon emission computed tomography), represent the function of organs. Medical image fusion techniques are employed to integrate the information provided by individual imaging sensors and to obtain an image with high functional and anatomical information content [1]. Each of the image fusion categories has its advantages and drawbacks. The first category, including SD (spatial domain)-based methods, has a basic ability in spatial information visualization, such as edges and spatial details [2]. Furthermore, the methods of this group are relatively fast and easy to implement. However, they do not take the local dissimilarities between the source images into account, such that they suffer from color distortion [3]. In contrast, the second category, including TD (transform domain)-based methods, has a superior color information consistency but with higher spatial distortion [3]. In addition to these general categories, fusion of medical images can be performed at three levels, i.e., the score level (or decision level), object-level (or feature level), and sensor level (or pixel level). Sensor-level fusion directly synthesizes the pixels of the source images to obtain the fused image. At this level of fusion, a perfect registration of source images is required. Object-level fusion extracts informative features of the source images (for example, by neural networks, clustering algorithms, or template methods [4]) and subsequently combines them to obtain a single fused feature vector. Score level fusion methods usually process the source images individually to extract useful information and then combine the information using a classifier or voting method. The sensor level, at which our method is performed, is the most common type of fusion level owing to the fact that it is computationally efficient. The next section reviews the literature in the area of medical image fusion at the sensor level. During the last two years, the use of cartoon-texture decomposition in the field of image fusion has shown very promising results [5-8]. The fusion methods based on cartoon-texture analysis transfer the source images into a new space by using the cartoon-texture decomposition, in which spatial details (such as edges) and color information are discriminated properly, and thus the fusion procedure can be efficiently performed. The cartoon component reveals information about the isophotes, geometric structures, and smooth pieces of the source images, while the texture component reveals information about textures, fine details, oscillating patterns, and noise [5]. Due to the success of the cartoon-texture decomposition in different image fusion tasks, such as multi-focus image fusion and remote sensing image fusion, we use this decomposition for the medical image fusion task. The pulse-coupled neural network, proposed by Eckhorn et al. in 1990, is a biologically inspired spiking neural network [9]. This neural network, characterized by the pulse synchronization of neurons and global coupling, was inspired by pulse bursts in monkey and cat visual cortex. The pulse-coupled neural network is a two-dimensional, laterally connected, single-layered neural network using pulse-coupled neurons, which are connected with pixels of the image [10]. This neural network is widely used in various image processing applications. Particularly, several image fusion approaches based on the pulse-coupled neural network have been proposed [11-15]. This neural network is usually employed in the transform domain, e.g., PCA (principal component analysis) [10] and shearlet [11]. According to the experimental results reported in the literature, one can conclude that pulse-coupled neural network-based fusion approaches can achieve state-of-the-art performance. However, these algorithms have some significant limitations. For example, image contrast is reduced, and sometimes fine spatial information is lost [16-18]. The main reason is that during most of the pulse-coupled neural network-based fusion methods, the neuron is motivated by only one coefficient in the transform domain. To cope with this problem, this paper exploits the EOG (energy of image gradient) of the coefficients to motivate the neurons of the pulse-coupled neural network. The main contribution of this paper is that a novel multimodal image fusion framework is established. The proposed framework makes use of the cartoon-texture decomposition to map the source images into a new space in which spatial information and color information are discriminated from each other. The EOGs of the coefficients of texture components are computed and fed to the pulse-coupled neural network for coefficient selection. In addition, the regional energy of the cartoon components is used for fusing the cartoon components of the source images. Simulation results demonstrate that a clear fused image containing a large amount of spectral and spatial information can be produced by the proposed method. Particularly, the proposed method using cartoon-texture decomposition can achieve better performance than methods using multiscale transforms, such as contourle. The outline of the rest of the paper is as follows. Section 2 reviews the literature in the area of medical image fusion at the sensor level. In Section 3, the principle of cartoon-texture image decomposition is briefly presented. Section 4 describes the pulse-coupled neural network model. Section 5 presents the proposed fusion framework. In Section 6, several experiments are performed to evaluate the performance of the proposed medical image fusion method. Then, the results of the experiments are analyzed to show the advantages of the proposed method. Finally, concluding remarks are drawn in Section 7.
2. Review of Fusion Methods at the Sensor LevelThe following subsections review the two classes of image fusion approaches at the sensor level, namely, TD-based methods and SD-based methods.
2.1. TD-Based Fusion Methods In the TD-based fusion methods, the source images are mapped into a new space in which the information of the source images is represented by using a joint frequency-space domain. Then, the fused image in the transform domain is obtained using a specific fusion rule, such as the maximum selection rule, which takes the activity level of coefficients into account. Finally, the fused image in the spatial domain is obtained by applying the inverse transform [19]. To this end, Ref. [20] introduced a new image fusion framework for medical images based on non-subsampled contourlet transform (NSCT). In this method, the source images are transformed by the NSCT (non-subsampled contourlet transform) to obtain low- and high-frequency components. Then, this method makes use of two different fusion rules based on phase congruency and directive contrast to obtain fused low- and high-frequency coefficients. Finally, the synthesized image is constructed by applying the inverse NSCT to all composite coefficients. An approach based on non-Gaussian statistical modeling of discrete wavelet coefficients was proposed in [21]. This approach exploits families of alpha-stable and generalized Gaussian distributions to model image wavelet coefficients. Then, this method estimates distribution parameters through the expectation maximization (EM) algorithm. Finally, this technique incorporates the model into a weighted average image fusion algorithm. Recently, edge-preserving filters have been employed to provide multiscale representations of images. A novel edge-preserving image fusion approach was proposed by Bai et al. using the multiscale toggle contrast operator [22]. This method extracts the multiscale dilation and erosion features of the source images through the multiscale toggle contrast operator to obtain edge information of the source images. The fused image is constructed by combining the dilation and erosion features into a base image. Experimental results showed that clear and well-preserved edge features are produced by this algorithm. The multiscale geometrical bilateral filter (MGBF) transform was introduced in Ref. [23], which combines the ability of edge-preserving of the bilateral filter and that of representing directional information of the directional filter bank (DFB). This fusion method decomposes the source images into the directional subbands and approximation subbands using the MDBF. Then the fused subbands are obtained by the max fusion rule. Applying the inverse MDBF gives the fused image. A fast and effective image fusion method was proposed based on guided filtering in Ref. [24]. The guided filtering-based weighted average technique used in this fusion framework can take advantage of spatial consistency for the fusion of approximation and detail subbands. Several fusion methods were proposed based on cartoon-texture decomposition, which can be considered as edge-preserving filters [5-8]. These methods show promising results in the field of image fusion. There are a number of fusion methods that combine different transforms to achieve superior performance. For example, an image fusion approach based on a combination of wavelet and curvelet transform was proposed in [25]. Owing to the fact that wavelets cannot effectively represent long edges and curvelets cannot effectively capture small features, combinations of these two transforms can improve the quality of fusion. Ref. [26] introduced a hybrid multiresolution method combining the wavelet transform and the NSCT to fuse images. This method makes use of some complementary characteristics between the two mentioned transforms simultaneously. In the fusion method proposed by Kavitha et al. [27], the wavelet transform was combined with the discrete ripplet transform (DRT) to obtain an efficient hybrid transform.
2.2. SD-Based Fusion MethodsA straightforward fusion technique is to obtain each pixel of the fused image by calculating the weighted average of the corresponding pixels in the source images. Usually, the activity level of pixels determines the weights. For example, the fusion methods proposed in [28] and [29] employ the support vector machines (SVMs) and neural networks to weight the source pixels with the highest level of activity. The method proposed in Ref. [30] divides the source images into fixed-size blocks and calculates the spatial frequency (SF) in each block to measure the activity level. In this way, the proposed method takes advantage of spatial context information for fusion. To further improve the fusion performance, Ref. [31] proposed a new method that adaptively determines the optimal block size by means of a quad-tree structure. Another strategy adopted to take the spatial correlation of neighboring pixels into account is the post-processing of the pixel-wise fusion map. For instance, a new method was introduced by Zhang et al. [32], which consists of three steps. In the first step, the pixels with a high activity level are detected using the Laplacian operator. Second, the KNN (K-nearest neighbor) matting method is employed to obtain a globally optimal fusion map. The third step obtains the fused image according to the optimal fusion maps. Apart from the methods calculating a weighted average of source pixels, there are a number of total variation (TV) based methods which perform fusion in the spatial domain. Ref. [33] introduced a TV-based fusion technique in which the fusion task is considered as an inverse problem and a locally affine model is utilized as the forward model. Then, a seminorm TV-based approach, along with the PCA, is employed to iteratively calculate the fused image. Piella proposed a variational method for image fusion [34]. This method can fuse any number of images while preserving salient features. The geometry of source images is described using the structure tensor. In addition, this method combines the geometry merging of the source images with intensity correction and perceptual enhancement to provide a ‘sharp’ and ‘natural’ fused image. A general variational method for image fusion was proposed in [35], which formulates the synthesized image as a convex combination of the source images and makes use of concepts from perceptually inspired contrast enhancement criteria, such as non‐linearity and locality.
3. Cartoon-Texture DecompositionThe cartoon-texture decomposition tries to separate an image into a sketchy approximation with sharp edges and boundaries, which is called the cartoon part, and a subband with small-scale oscillatory patterns, which is called the textural component. The general scheme by which the image is decomposed into the cartoon component denoted by and the texture component denoted by is a minimization problem [36]: where the values of the functions are greater than or equal to zero for all inputs and are spaces of the functions in such a way that and if and only if . Furthermore, is a trade-off parameter. The variational space is the appropriate one for the sparse representation of the cartoon part. Therefore, the function is selected to be the TV of the cartoon part, which eliminates smooth features while preserving boundaries. In addition, the textural part can be represented by the L2-TV model. This means that the cartoon-texture decomposition makes use of the total variation and L2-norm to obtain the components [37]. A mathematically convex and tractable model for the cartoon-texture decomposition is as follows [38]: where and is the divergence of . Here, the tuning parameters and are greater than zero. Moreover, denotes the p-norm of and is the ith element of . Fig. 1 shows a PET image and a SPECT image decomposed by this model (p is set to 1 in this figure). The low-resolution pseudo-color image (PET or SPECT) is decomposed by Eq. (2) in our method. Here, we consider the following energy function when decomposing the high-resolution gray-scale MRI image: where is the cartoon component of the MRI image and is the cartoon component of the ith band of the pseudo-color image obtained by applying Eq. (2). The coefficients (for i=1, 2, 3) are calculated through a linear combination of spectral bands of the pseudo-color image. To this end, we exploit the adaptive method proposed by Rahmani et al. [39] to obtain the coefficients. This new term adapts the scale of the cartoon part of the MRI image to that of the pseudo-color image. So, the proposed model for the MRI image decomposition is as follows:
where is a trade-off parameter. In what follows, we describe the numerical algorithm used to solve Eqs. (2) and (4). Considering that Eq. (2) is a special case of Eq. (4), we only solve Eq. (4) using the numerical algorithm. One can rewrite Eq. (4) as a separable convex optimization problem subject to linear constraints by making use of the auxiliary variables as follows: The optimization problem in Eq. (5) is defined as the augmented Lagrangian function [40]: where is the Lagrange multiplier and coefficients (i=1, 2, 3, 4) denote penalty parameters. A variable splitting approach, such as ADMM-GBS (alternating direction method of multipliers with Gaussian back substitution) [41], is used to solve this equation. Ref. [40] has analyzed the convergence of this approach.
Fig. 1. Cartoon-Texture Decomposition: (a) SPECT Image, (b) Cartoon Component, (c) Texture Component, (d) PET Image, (e) Cartoon Component, (f) Texture Component.
4. Pulse-Coupled Neural NetworkThe pulse-coupled neural network is a model based on the monkey’s and cat’s primary visual cortex. In the pulse-coupled neural network model proposed by Eckhorn et al. [9], a pulse-coupled neuron denoted by Nij, is made up of three functional units (see Fig. 2): 1- the receptive fields, 2- the modulation product, and 3- the pulse generator. Each neuron receives input signals not only from other neurons, but also from external sources by means of two channels in the receptive fields. The first channel receives the feeding input Fij, and the second channel receives the linking input Lij. The function of a neuron can be represented by the following equations [11]: In fact, the pulse-coupled neural network combines the linking input Lij and the feeding input Fij in a second-order manner to obtain the value of total internal activity , and then compares with the value of dynamic neuromime threshold to obtain the output . If is sufficiently large, then the neuron fires a pulse. The parameters , and are the decay constants of the pulse-coupled neural network model, and the parameters , , and are the magnitude scaling constants. Furthermore, and are the synaptic weight strengths used to weight the outputs of other neurons in the feeding receptive field and linking receptive field, respectively. In addition, the parameter is the linking strength. In image processing applications, the input to the F channel of the (i,j)th neuron, is the value of its corresponding pixel, which is denoted by : The output of the pulse-coupled neural network is an image denoted by , whose size is the same as that of the input image. The value of the (i,j)th pixel of the image is obtained by taking the summation of its corresponding neuron outputs over n [15]:
Fig. 2. The Pulse-Coupled Neural Network Model.
5. Proposed MethodFor fusion of the MRI and PET (or SPECT) medical images, the proposed image fusion framework is as follows: Step 1: Use the given MRI and PET (or SPECT) images as inputs, which are denoted by and , respectively ( is the ith band of the pseudo-color image). Step 2: Decompose the ith band of the pseudo-color image using Eq. (2) to obtain the cartoon component and the texture component .
Step 3: Calculate the non-negative coefficients (for i=1, 2, 3) using the algorithm proposed in Ref. [39]. This algorithm minimizes the following objective function to obtain coefficients : Step 4: Decompose the MRI image using Eq. (4) to obtain the cartoon component and the texture component . Step 5: To obtain the cartoon component of the ith fused band, the regional energy is used as the activity level measurement: where denotes the cartoon part (either or ), denotes the local energy (either or ), and is a 3×3 weighting window. After calculating the local energy of the cartoon components of both the MRI and PET (or SPECT) images, the maximum selection rule is applied to select the coefficient with the highest activity measure:
where is the cartoon component of the ith fused band. Step 6: For the texture components, compute the EOG as follows: in which, denotes the texture component (either or ). Step 7: Use the EOG of texture components as the input of the pulse-coupled neural network and motivate the neurons according to Eqs. (7) to (11). Step 8: Calculate the firing time of the pulse-coupled neural network (sum of the = 1 or ) at the nth iteration, as follows: Step 9: If the iteration number n exceeds the maximum iteration number nmax , fuse the texture components by means of the firing time calculated in the previous step. where is the texture component of the ith fused band. Step 10: Apply the inverse cartoon-texture transform to the fused cartoon and texture coefficients to obtain the final fused image:
6. Experimental Results and ComparisonsIn this section of the paper, experiments on several medical source images are presented for testing the performance of the proposed fusion framework. Firstly, the datasets used in our experiments are introduced. The parameters are then tuned, owing to the fact that the parameters have a significant effect on the fusion performance of the proposed approach. Subsequently, five quality indices, applied to evaluate the performance of the synthesized images, are given. Finally, the performance of the proposed framework is presented and compared with several pulse-coupled neural network-based fusion frameworks, including Method 1, Method 2, and Method 3: Method 1: This fusion method exploits NSCT and pulse-coupled neural network to make use of their strengths [16]. Spatial frequency in the NSCT domain is calculated and used as the input to motivate the network, and then coefficients in the NSCT domain with large firing times are chosen as the coefficients of the fused image. Method 2: This fusion method is based on NSST (non-subsampled shearlet transform) and pulse-coupled neural network [11]. Firstly, the source images are decomposed using the NSST to obtain several subimages. Then, the regional energy is employed to fuse the low-frequency subbands. Finally, high-frequency subbands are fused by means of the pulse-coupled neural network mode. Method 3: The discrete multi-parameter fractional random transform (DMPFRNT) and pulse-coupled neural network are exploited in this fusion technique [42]. This method converts the source images into the DMPFRNT domain. High amplitude spectrum (HAS) and low amplitude spectrum (LAS) components in the DMPFRNT spectrum domain contain different information about source images. In this fusion approach, the local standard deviation of the amplitude spectrum is selected as the link strength of the pulse-coupled neural network. Method 4: In this method, two input images are fed into the branches of a twin convolutional neural network such that both branches share the same weights and architecture. Each one includes three convolutional layers, and the feature maps essentially act as the role of activity level measures [43]. Two pairs of representative data sets are selected to be used in visual tests, which are medical images from the human brain:
This data set consists of a pair of MRI and PET images (sagittal view), belonging to a 38-year-old man in excellent health. Both the MRI and PET images in dataset A are of size 256×256.
This data set consists of a pair of MRI and SPECT (SPECT-TI) images, belonging to a 51-year-old woman with neoplastic disease (brain tumor). The type of tumor in this patient is a glioma. Both the MRI and SPECT images in dataset B are of size 256×256. These datasets were downloaded from the website of the Atlas project, whose web address is: http://www.med.harvard.edu/AANLIB. The source images used in our experiments have been previously registered. Without loss of generality, we empirically take , , , , , , and in the cartoon-texture decomposition algorithm. In addition, the maximum number of iterations is set equal to 55 when decomposing the source images by the variable splitting approach. Furthermore, the parameters of the pulse-coupled neural network are set as , , , , , , and . For visual evaluation, the fused images corresponding to “dataset A” and “dataset B” obtained by different fusion methods are represented in Figs. 3(a)-(f) and 4(a)-(f), respectively. Figs. 3(a) and 3(b) show the PET image and MRI image of “dataset A”, respectively. Figs. 3(c) to (f) represent the combined images of the first dataset given by the proposed method, Method 1, Method 2, and Method 3, respectively. Figs. 4(a) and 4(b) show the MRI image and SPECT image of “dataset B”, respectively. Figs. 4(c) to (f) represent the combined images of the second dataset given by the proposed method, Method 1, Method 2, and Method 3, respectively. To evaluate the images more accurately, two zoomed areas corresponding to the “dataset A” are shown in Fig. 5, and two zoomed areas corresponding to the “dataset B” are shown in Fig. 6. The first zoomed area contains functional information, while the second zoomed area contains anatomical information. Compared with the reference images, the fused images obtained by the proposed method successfully preserve the informative features and combine them to generate a clearer image in all areas of the scene. For example, the pink color in the region of calcarine fissure is the same as that in the reference image (see the white arrows in Figs. 3(a) and 3(c)), and the structural information represented in Fig. 5(b) is more similar to the reference image compared to other methods. Some black artifacts exist around the fused images generated by Method 1 (see the red arrow in Fig. 4(d)). These black artifacts do not exist in the source images and thus impair image quality severely. In addition, one can see that Method 1 suffers from color distortion such that the desired color is not reproduced in the fused results obtained by Method 1. The colors in the fused images of Method 2 become darker than those of the reference images. This phenomenon is more obvious in Fig. 5(d). Furthermore, as can be seen from Figs. 5(i) and 6(i), Method 2 somewhat blurs the structural information of the MRI image. The color distortion is more severe in the fused results obtained by Method 3. For example, red colors produced in Fig. 5(e) do not exist in the reference image shown in Fig. 5(a). So, the fused results of Method 3 can mislead the radiologist in the diagnosis in clinical applications. Moreover, fusion results of Method 3 cannot meet the expectations in terms of anatomical information content (see Figs. 5(j) and 6(j)). Figs. 3(g) and 4(g) indicate that the Method cannot retain spatial information well, such that local spatial information is not clearly visible (see the white arrow in Fig. 3(g)). To make the experiments comprehensive, two new data sets called “dataset C” and “dataset D” are used in quantitative evaluations (see Fig. 7). Quantitative assessment is essential in the field of image fusion [44]. The five indices utilized for quantitative evaluation are:
When the reference image is available, the performance of fusion methods can be assessed using the PFE index: where is the fused image and is the reference image. To obtain the final value, the PFE value between the fused image and pseudo-color image is calculated, and that between the fused image and the MRI image is computed. Then, the average of the obtained values is reported. This index will be zero when the reference and the fused image are the same, and the value of this index will increase when the synthesized result is deviated from the source image [45].
Fig. 3. Fusion Results of the First Dataset: (a) PET Image, (b) MRI Image, (c) Proposed Method, (d) Method 1, (e) Method 2, (f) Method 3, (g) Method 4. Fig. 4. Fusion Results of the Second Dataset: (a) MRI Image, (b) SPECT Image, (c) Proposed Method, (d) Method 1, (e) Method 2, (f) Method 3, (g) Method 4. Fig. 5. Zoomed Areas of the First Dataset: (a) PET Image, (b) Proposed Method, (c) Method 1, (d) Method 2, (e) Method 3, (f) Method 4, (g) MRI Image, (h) Proposed Method, (i) Method 1, (j) Method 2, (k) Method 3, (L) Method 4.
The index fusion symmetry FS [46] was proposed to indicate the symmetry of the fusion process with respect to the source images. This index is mathematically formulated as:
where denotes the mutual information between the fused image and the source image A and denotes the mutual information between the fused image and the source image B.
The FEQI index considers the significance of edge information of the fused image, and is formulated as follows: where , , and denote the edge information of , , and , respectively, and WFQI measures the amount of information of the source images, which has been transferred into the synthesized image (to see more details, see Ref. [47]). The contribution of the edge information depends on the parameter . In our experiments, the parameter is set to 1.
This newly proposed index is used to evaluate the perceptual quality of images [48]. Owing to the fact that image gradients are very sensitive to image distortions, the GMSD index makes use of the global variation of a gradient-based quality index to predict the overall quality of the image. The GMSD index is computed as follows: where is the fused image, is the reference image, is the total number of pixels in the image, and: Fig. 6. Zoomed Areas of the Second Dataset: (a) SPECT Image, (b) Proposed Method, (c) Method 1, (d) Method 2, (e) Method 3, (f) Method 4, (g) MRI Image, (h) Proposed Method, (i) Method 1, (j) Method 2, (k) Method 3, (L) Method 4. Table (1): UIQI values obtained by different fusion schemes.
Table (2): PFE values obtained by different fusion schemes.
Table (3): FS values obtained by different fusion schemes.
Table (4): EFQI values obtained by different fusion schemes.
Table (5): GMSD values obtained by different fusion schemes.
Table (6): UIQI values obtained by different misalignments.
Table (7): Running time of different fusion schemes (in Seconds).
Table (8): Sensitivity analysis of different parameters (impact on UIQI index using standard deviation).
Fig. 7. Datasets C and D: (a) MRI Image of dataset C, (b) PET Image of dataset C, (c) MRI Image of dataset D, (d) PET Image of dataset D. in which denotes the convolution operator. The GMSD value between the fused image and the pseudo-color image, and that between the fused image and the MRI image, are first obtained. Then, the average of these values is used as the final index.
Proposed by Wang and Bovik, this index measures the similarity between the fused image and the source image by considering three factors, i.e., loss of correlation, contrast distortion, and luminance distortion [49]: where and are the mean value of the source image and that of the fused image, respectively, and are the variance of the source image and that of the fused image, respectively, and is the covariance between the source image and the fused image. Again, this index is obtained for the two source images, and the average of the obtained values is reported. Values of different objective indices obtained using various fusion methods are tabulated in Tables (1) to (5) for four different pairs of medical images. According to these tables, all the objective indices prove that the fused image obtained by the proposed method is strongly correlated with the reference images and more informative features, e.g., edges and textures are preserved during the fusion process, which means that the proposed fusion approach is the best one among the implemented algorithms. In addition, Table (6) shows the effect of different types of misalignments on the results using the UIQI index. From this table, it can be seen that misalignments can decrease the performance of the proposed method. Moreover, Table (7) reports the execution time of different fusion algorithms. According to this table, the shortest execution time belongs to Method 1, which indicates the superiority of this method in computational complexity over other methods. Finally, Table (8) reports the sensitivity of the proposed model to its parameters. As can be seen from the table, the model is more sensitive to parameters of the pulse-coupled neural network such as , , , , and . According to this table, the maximum standard deviation belongs to , which indicates that the model has the highest sensitivity to this parameter. 7. ConclusionAn efficient MRI and PET/SPECT image fusion scheme in the cartoon-texture domain was proposed. The proposed method decomposes the source images using the cartoon-texture transform to obtain their components. For this purpose, the ADMM-GBS technique, one of the variable splitting algorithms, is employed. Subsequently, the proposed method fuses the cartoon components based on the regional energy of coefficients. Then, the texture components are combined by means of the pulse-coupled neural network whose inputs are the EOGs of texture coefficients. The fusion performance of the proposed framework was tested and compared against some pulse-coupled neural network-based fusion methods using visual inspection, as well as objective measures. Visual analysis of the results demonstrated that the proposed approach can preserve the structural information of the MRI image and metabolic information of the PET/SPECT image with good contrast. The quantitative results obtained showed that the proposed technique preserves important features of the source images in terms of classical criteria such as UIQI and modern perceptual criteria such as GMSD.
[1] Submission date:29, 12, 2023 Acceptance date:07, 07, 2025 Corresponding author: Hamid Reza Shahdoosti, Department of Electrical Engineering, Hamedan University of Technology, Hamedan, Iran | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| مراجع | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
[2] M. Imani, H. Ghassemian, "Pansharpening optimisation using multiresolution analysis and sparse representation", International Journal of Image and Data Fusion, Vol. 8, No. 3, July 2017. http://doi.org/10.1080/19479832.2017.1334713 [3] H. R. Shahdoosti, N. Javaheri, "Pansharpening of clustered MS and Pan images considering mixed pixels", IEEE Geoscience and Remote Sensing Letters, Vol. 30, No. 6 March 2017. http://doi.org/10.1109/LGRS.2017.2682122 [4] P. Ganasala, V. Kumar, "CT and MR image fusion scheme in nonsubsampled contourlet transform domain", Journal of digital imaging, Vol 27, No. 3, June 2014. http://doi.org/10.1007/s10278-013-9664-x [5] Y. Zhang, "Multi-focus image fusion based on cartoon-texture image decomposition", Optik. Int. J. Light Electron Opt., September 2015. https://doi.org/10.1016/j.ijleo.2015.10.098 [6] M. Lotfi, H. Ghassemian, "A new pansharpening method based on cartoon + texture decomposition", In 2016 8th International Symposium on Telecommunications (IST), pp. 468-471, September 2016. http://doi.org/10.1109/ISTEL.2016.7881865 [7] Z. Liu, Y. Chai, H. Yin, J. Zhou, Z. Zhu, "A novel multi-focus image fusion approach based on image decomposition", Information Fusion, Vol 35, No 2, July 2017. https://doi.org/10.1016/j.inffus.2016.09.007 [8] M. Lotfi, H. Ghassemian, "A pansharpening method based on modified cartoon plus texture decomposition", Remote Sensing Letters, Vol. 9, No. 3, October 2018. https://doi.org/10.1080/2150704X.2017.1415470 [9] R. Eckhorn, H. J. Reitboeck, M. Arndt, P. Dicke, "Feature linking via synchronization among distributed assemblies: Simulations of results from cat visual cortex", Neural computation, Vol. 2, No.3, August 1990. http://doi.org/10.1162/neco.1990.2.3.293 [10] Y. Zhang, L. Chen, Z. Zhao, J. Jia, J. Liu, "Multi-focus image fusion based on robust principal component analysis and pulse-coupled neural network", Optik-International Journal for Light and Electron Optics, Vol. 125, No. 17, Appril 2019. https://doi.org/10.1016/j.ijleo.2014.04.002 [11] S. Singh, D. Gupta, R. S. Anand, V. Kumar, "Nonsubsampled shearlet based CT and MR medical image fusion using biologically inspired spiking neural network", Biomedical Signal Processing and Control, Vol. 18, No. 6, May 2015. https://doi.org/10.1016/j.bspc.2014.11.009 [12] P. Geng, X. Zheng, Z. G. Zhang, Y. J. Shi, S. Q. Yan, "Multifocus image fusion with PCNN in Shearlet domain", Research Journal of Applied Sciences, Engineering and Technology", Vol. 4 No. 15, October 2012. https://doi.org/10.1155/2021/5439935 [13] W. Li, X. F. Zhu, "A new algorithm of multi-modality medical image fusion based on pulse-coupled neural networks", In International Conference on Natural Computation, pp. 995-1001, 2005. https://doi.org/10.1007/11539087_131 [14] W. Li, X. F. Zhu, "A new image fusion algorithm based on wavelet packet analysis and PCNN", Proceedings of Machine Learning and Cybernetics, pp. 5297-5301, 2005. http://doi.org/10.1109/ICMLC.2005.1527879 [15] W. Huang, Z. Jing, "Multi-focus image fusion using pulse coupled neural network", Pattern Recognition Letters, Vol. 28, No. 9, June 2007. https://doi.org/10.1007/978-3-319-90716-1_14 [16] Q. Xiao-Bo, Y. Jing-Wen, X. I. Hong-Zhi, Z. Zi-Qian, "Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain", Acta Automatica Sinica, Vol. 34, No. 12, October 2009. https://doi.org/10.1016/S1874-1029(08)60174-3 [17] Z. Wang, Y. Ma, "Medical image fusion using m-PCNN", Information Fusion, Vol. 9, No. 2, July 2008. https://doi.org/10.1016/j.inffus.2007.04.003 [18] M. M. Deepika, V. Vaithyanathan, "An efficient method to improve the spatial property of medical images", Journal of Theoretical and Applied Information Technology, Vol. 35, No. 2, May 2012. https://doi.org/10.1007/s10278-022-00769-7 [19] H. R. Shahdoosti, A. Mehrabi, "Multimodal image fusion using sparse representation classification in tetrolet domain", Digital Signal Processing, Vol. 79, Augest 2018. https://doi.org/10.1016/j.dsp.2018.04.002 [20] G. Bhatnagar, Q. J. Wu, Z. Liu, "Directive contrast based multimodal medical image fusion in NSCT domain", IEEE Transactions on Multimedia, Vol. 15, No. 5, May 2013. http://doi.org/10.1109/TMM.2013.2244870 [21] A. Loza, D. Bull, N. Canagarajah, A. Achim, "Non-Gaussian model-based fusion of noisy images in the wavelet domain", Computer Vision and Image Understanding", Vol. 114, No. 1, Februry 2010. https://doi.org/10.1016/j.cviu.2009.09.002 [22] X. Bai, F. Zhou, B. Xue, "Edge preserved image fusion based on multiscale toggle contrast operator", Image and Vision Computing, Vol. 29, No. 12, Janury 2012. https://doi.org/10.1016/j.imavis.2011.09.003 [23] J. Hu, S. Li, "The multiscale directional bilateral filter and its application to multisensor image fusion", Information Fusion, Vol. 13, No. 3, September 2012. https://doi.org/10.1016/j.inffus.2011.01.002 [24] S. Li, X. Kang, J. Hu, "Image fusion with guided filtering", IEEE Transactions on Image Processing, Vol. 22, No. 7, June 2013. http://doi.org/10.1109/TIP.2013.2244222 [25] S. Li, B. Yang, "Multifocus image fusion by combining curvelet and wavelet transform", Pattern Recognition Letters, Vol. 29, No. 9, May 2008. https://doi.org/10.1016/j.patrec.2008.02.002 [26] S. Li, B. Yang, "Hybrid multiresolution method for multisensor multimodal image fusion", IEEE Sensors Journal, Vol. 10, No. 9, October 2010. http://doi.org/10.1109/JSEN.2010.2041924 [27] C. T. Kavitha, C. Chellamuthu, R. Rajesh, "Medical image fusion using combined discrete wavelet and ripplet transforms", Procedia Engineering, Vol. 38, Februry 2012. https://doi.org/10.1016/j.proeng.2012.06.102 [28] S. Li, J. T. Kwok, Y. Wang, "Multifocus image fusion using artificial neural networks", Pattern Recognition Letters, Vol. 23, No. 8, July 2002. https://doi.org/10.1016/S0167-8655(02)00029-6 [29] S. Li, J. Y. Kwok, I. H. Tsang, Y. Wang, "Fusing images with different focuses using support vector machines", IEEE Transactions on Neural Networks, Vol. 15, No. 6, March 2004. http://doi.org/10.1109/TNN.2004.837780 [30] S. Li, J. T. Kwok, Y. Wang, "Combination of images with diverse focuses using the spatial frequency", Information fusion, Vol. 2, No. 3, October 2001. https://doi.org/10.1016/S1566-2535(01)00038-0 [31] I. De, B. Chanda, "Multi-focus image fusion using a morphology-based focus measure in a quad-tree structure", Information Fusion, Vol. 14, No. 2, April 2013. https://doi.org/10.1016/j.inffus.2012.01.007 [32] X. Zhang, H. Lin, X. Kang, S. Li, "Multi-modal image fusion with KNN matting", In Chinese Conference on Pattern Recognition. Springer, Berlin, Heidelberg, pp. 89-96, November 2014. https://doi.org/10.1007/978-3-662-45643-9_10 [33] M. Kumar, S. Dass "A total variation-based algorithm for pixel-level image fusion", IEEE Transactions on Image Processing, Vol. 18, No. 9, December 2009. http://doi.org/10.1109/TIP.2009.2025006 [34] G. Piella, "Image fusion for enhanced visualization: A variational approach", International Journal of Computer Vision, Vol. 83, No. 1, 2009 (https://doi.org/10.1007/s11263-009-0206-4). [35] D. Hafner, and J. Weickert, "Variational Image Fusion with Optimal Local Contrast", In Computer Graphics Forum, Vol. 35, No. 1, pp. 100-112, Februry 2016. https://doi.org/10.1111/cgf.12690 [36] Y. Meyer, "Oscillating patterns in image processing and nonlinear evolution equations", The 15th Dean Jacqueline B. Lewis Memorial Lectures, American Mathematical Soc. 2001. http://doi.org/10.1090/ULECT/022 [37] J. F. Aujol, G. Gilboa, T. Chan, S. Osher, "Structure-texture image decomposition-modeling, algorithms, and parameter selection", International Journal of Computer Vision, Vol. 67, No. 1, pp. 111-136, June 2006. https://doi.org/10.1007/s11263-006-4331-z [38] L. A. Vese, S. J. Osher, "Modeling textures with total variation minimization and oscillating patterns in image processing", Journal of Scientific Computing, Vol. 19. No. 3, May 2003. https://doi.org/10.1023/A:1025384832106 [39] S. Rahmani, M. Strait, D. Merkurjev, M. Moeller, T. Wittman, "An adaptive IHS pan-sharpening method", IEEE Geoscience and Remote Sensing Letters, Vol. 7, No. 4, Augest 2010. http://doi.org/10.1109/LGRS.2010.2046715 [40] M. K. Ng, X. Yuan, W. Zhang, "Coupled variational image decomposition and restoration model for blurred cartoon-plus-texture images with missing pixels", IEEE Transactions on Image Processing, Vol. 22, No. 6, September 2013. http://doi.org/10.1109/TIP.2013.2246520 [41] B. He, M. Tao, X. Yuan, "Alternating direction method with Gaussian back substitution for separable convex programming", SIAM Journal on Optimization, Vol. 22, No. 2, June 2012. https://doi.org/10.1137/11082234 [42] J. Lang, Z. Hao, "Novel image fusion method based on adaptive pulse coupled neural network and discrete multi-parameter fractional random transform", Optics and Lasers in Engineering, Vol. 52, No. 6, Februry 2014. https://doi.org/10.1016/j.optlaseng.2013.07.005 [43] Y. Liu, X. Chen, J. Cheng, H. Peng, "A medical image fusion method based on convolutional neural networks", In 2017 20th International Conference on Information Fusion, IEEE, pp. 1-7, July 2017. http://doi.org/10.23919/ICIF.2017.8009769 [44] M. Sheikhan, A. Chamankar, "Optimizing image fusion process using geometrical searching algorithm", Computational intelligence in Electrical Enginering, Vol 6, No, 3, 2017. https://journals.ui.ac.ir/article_15431.html [In Persian] [45] V. P. S. Naidu, "Discrete cosine transform-based image fusion", Defence Science Journal, Vol. 60, No. 1, July 2010. https://doi.org/10.1016/j.sigpro.2013.09.001 [46] M. Manchanda, R. Sharma, "A novel method of multimodal medical image fusion using fuzzy transform", Journal of Visual Communication and Image Representation, Vol. 40, December 2016. https://doi.org/10.1016/j.jvcir.2016.06.021 [47] Y. Jiang, M. Wang, "Image fusion with morphological component analysis", Information Fusion, Vol. 18, No. 9, October 2015. https://doi.org/10.1016/j.inffus.2013.06.001 [48] W. Xue, L. Zhang, X. Mou, A. C. Bovik, "Gradient magnitude similarity deviation: A highly efficient perceptual image quality index", IEEE Transactions on Image Processing, Vol. 23, No. 2, May 2014. https://doi.org/10.1109/TIP.2013.2293423 [49] Z. Wang, A. C. Bovik, "A universal image quality index", IEEE Signal Processing Letters, No. 9, No. 3, March 2002. https://doi.org/10.1109/97.995823 | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
آمار تعداد مشاهده مقاله: 39 تعداد دریافت فایل اصل مقاله: 22 |
||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||