Introduction
Medical images are important tools for detecting and diagnosing different medical conditions and ailments [1], [2]. However, the quality of medical images can often be degraded during the capture procedure due to factors such as noise interference, poor illumination, and artifacts. This may lead to the misdiagnosis of medical conditions, which can further exacerbate life threatening situations [3]. Consequently, it is imperative that effective approaches be developed for enhancing the quality of medical images [4].
There are different image enhancement approaches that aim to transform an input image towards obtaining a better, more detailed, and less noisy output image [5]–[12]. However, these image enhancement methods often produce either poorly or overly enhanced images, thus warranting the need for more effective techniques [13], [14]. To this effect, one early image enhancement method is the histogram equalization (HE) technique, which operates by adjusting an image’s contrast by either decreasing or increasing the global contrast of the image. Such an approach is often plagued by the problem of over-enhancement, particularly when low occurring intensities are transformed and combined with neighboring high occurring intensities [9]. As a solution, the bi-histogram equalization (BHE) was introduced for its ability to maintain the quality of an original image. Nevertheless, BHE becomes limited when the pixel distribution of an image is asymmetrical, which often occurs under real-life conditions. Other existing techniques include the logarithm transformation, gamma correction, and adaptive histogram equalization (AHE) methods, which were proposed to have less computational complexities. However, they become less effective under complex lighting differences [15]. Other approaches have been proposed such as the linear contrast stretching (LCS) method, which makes use of the concept of linear transformation wherein the gray level’s dynamic range is increased. However, the need to configure manually the threshold value of the LCS often limits its performance by making it susceptible to wrong threshold values, thus leading to poor output images [16].
In terms of medical image enhancement (MIE), the authors in [3] presented an enhancement technique comprising of two modules. The fuzzy dissimilarity histogram (FDH) is conducted in the first module, which uses the fuzzy neighborhood dissimilarity metric to evaluate each pixel via fuzzy rules. The contrast and entropy parameters were used to obtain the fuzzy dissimilarity clip limit using a fuzzy inference system (FIS). Then, the enhanced image was derived can be obtained by equalizing the FDH. Contextual intensity transformation (CIT) is applied in the second module to the FDH-based equalized output to generate a final enhanced medical image. Some studies were carried out using images from the BRATS 2015 dataset and satisfactory results were obtained using the peak signal-to-noise ratio (PSNR), entropy, and the contrast ratio. In [17], a morphological transformation process was conducted on medical images to improve their contrast. A disk-shaped mask was used in the bottom-hat and top-hat transformation process, which improves the method’s performance. The enhanced medical images were further generated by an iterative exfoliation process. The experimental results obtained demonstrated that the contrast of medical images were improved. In another article [18], a feedback graph attention convolutional neural network approach was presented that considers the global structure of an image by building a graph network from image sub-regions. The model comprises of three main parts, which include the parallel graph similarity branch, a feedback mechanism that refines the low-level representations using a recurrent structure, and an elimination of the artifacts to generate higher quality images. The results obtained showed satisfactory performance when compared to other approaches. In [4], an adaptive MIE technique was presented, wherein training samples were clustered into neighborhoods. The low-resolution space was then sectioned by exchanging the dictionary atoms with the cluster patch centers for computing K-nearest patches. These were then projected to map the low resolution patches upon the high resolution area. The authors conducted experiments using two medical datasets and showed that satisfactory results were obtained. Furthermore, in [19], a contrast optimization approach based on genetic algorithm was proposed by introducing a kernel that detects the edges of an image. The results obtained from their experiments demonstrated that the proposed approach was effective.
Owing to the subjective nature of the image enhancement process, several metaheuristic optimization algorithms (MOA) have found great application for image enhancement and with remarkable results to show [20]. The use of MOAs for MIE purposes has been demonstrated several times in the literature to outperform the existing methods mentioned above [11], [20]. Consequently, the literature has literally exploded with a long list of different MOAs, with many being deployed for MIE functions [21]. However, the question arises as to whether there exist any significant difference in both the qualitative and quantitative performances of these different MOAs. Hence, this paper investigates the performance of some selected well-known MOAs for MIE, namely, the artificial bee colony (ABC), cuckoo search optimization (CSO), differential evolution (DE), firefly, genetic algorithm (GA), particle swarm optimization (PSO), covariance matrix adaptive evolutionary strategy (CMAES), whale optimization algorithm (WOA), and the grey wolf optimization (GWO) algorithms. Firstly, our concern was to ensure that all MOAs were compared fairly, and to this effect we adopted the fitness computation rate (FCR) instead of the number of generations. By using the FCR, we can ensure that the objective function is computed the same number of times for all MOAs, thus guaranteeing that the computational resources of the MIE machine are consumed almost equally by all MOAs. Next, we deployed the transformation function proposed in [20] coupled with a robust evaluation function used in [11] to serve as the objective function for MIE purposes. Finally, a number of performance parameters were used to analyze the different MOAs in the qualitative and quantitative sense. As a result of our findings, the current article’s contributions can be described as follows:
We explore whether there exists any significant advantage(s) in using different MOAs for MIE. To this effect, nine state-of-the-art MOAs were considered, which were selected across both evolutionary and swarm-based classes. This distinguishes the current article as having researched a considerable number of MOAs only for the purpose of MIE.
By using the FCR as a metric of evaluation instead of the number of generations, we discovered that, although the GWO and WOA algorithms yielded slightly better empirical results than the other methods, nevertheless, there was little/no significant difference in both qualitative and quantitative measures among the different MOAs.
Physical timing performances were measured across the different MOAs with little/no statistical significant difference to show. Thus, this confirmed that using a fixed FCR translates to the same timing performance across most MOAs.
The remainder of the paper is structured as follows: Section II presents the method used in this study, which includes the transformation and evaluation functions used, a description of the different MOAs, the datasets and performance evaluation metrics utilized. Section III presents the results and discussion from the different experiments carried out and conclusions are drawn in Section V.
Method
In order to produce an enhanced output image, most image enhancement techniques typically use an effective transformation function to map the intensity values of an original input image [22]. Such a transformation function will often be characterized by parameters whose values are required to be optimally determined in order to improve performance. To this effect, metaheuristic-based approaches are normally applied to optimize such parameters of a transformation function via an objective function (typically called an evaluation function) [20]. Consequently, we first describe the specific transformation function deployed in our study. Then we discuss the evaluation function used in each MOA. We describe each MOA, focusing specifically on the modifications made to ensure that the FCR is accurately measured. Then the performance evaluation metrics used and the medical image database considered in our study are described.
A. Transformation Function
The transformation function proposed in [20] was used in our study. Following its formulation, the function has remained highly effective and thus widely applied in the literature. The function is stated as follows:\begin{align*}&\hspace {-.5pc}T(f(i,j)) = g(i,j) = k\left ({{\frac {M}{\sigma (i,j) + b}} }\right) \\&\times \left [{ {f(i,j) {-} c \times m(i,j)} }\right] + m{(i,j)^{a}} \tag{1}\end{align*}
\begin{equation*} M = \sum \nolimits _{i = 0}^{H-1} {\sum \nolimits _{j = 0}^{V-1} {f\left ({{i,j} }\right)} }, \tag{2}\end{equation*}
B. Evaluation Function
An evaluation function is a mathematical expression that computes the quality of an enhanced image
Further, the metrics considered in designing the evaluation function include the number of edge pixels, entropic measure, the number of foreground pixels, and the peak signal to noise ratio (PSNR). The steps involved in computing the function are as follows:
The number of edge pixels
in the enhanced image is computed using a Sobel threshold$N_{g} $ , which is automatically computed from the original image$S_{f} $ using a Sobel edge detector.$f(i,j)$ The Sobel threshold
is then used in the Sobel edge detector to obtain the edge intensities$S_{f} $ of the enhanced image. It is noted that in addition to being invariant,$E_{g} (i,j)$ was considered for computing$S_{f} $ in order to ensure a fair comparison between the original image and the different instances of the enhanced image. Thus, the number of edge pixels$E_{g} (i,j)$ in the enhanced image is obtained as:$N_{g} $ \begin{equation*} N_{g} =\sum _{i=1}^{H}\sum _{j=1}^{V}E_{g} (i,j)\tag{3}\end{equation*} View Source\begin{equation*} N_{g} =\sum _{i=1}^{H}\sum _{j=1}^{V}E_{g} (i,j)\tag{3}\end{equation*}
The variance
of$\vartheta _{g} (i,j)$ , and the variance$g(i,j)$ of$\vartheta _{f} (i,j)$ are computed within a neighborhood (window) having$f(i,j)$ pixels. And a threshold value$n\times n$ is automatically computed for$\eta _{f} $ using Otsu’s threshold algorithm.$\vartheta _{f} (i,j)$ Then, a binary function
is used to classify the pixels belonging to the foreground objects in the enhanced image as follows:$D_{g} (i,j)$ for\begin{align*} D_{g} (i,j)=\begin{cases} \displaystyle 1 & if{\mathrm{ \; }}\vartheta _{g} (i,j)\ge \eta _{f}\\[-3pt] \displaystyle 0 & if{\mathrm{ \; }}otherwise \end{cases}\tag{4}\end{align*} View Source\begin{align*} D_{g} (i,j)=\begin{cases} \displaystyle 1 & if{\mathrm{ \; }}\vartheta _{g} (i,j)\ge \eta _{f}\\[-3pt] \displaystyle 0 & if{\mathrm{ \; }}otherwise \end{cases}\tag{4}\end{align*}
and$i=1,2,\ldots,H$ .$j=1,2,\ldots,V$ Thereafter, the number of pixels
belonging to the foreground objects in$\phi _{g} $ is computed as$g(i,j)$ \begin{equation*} \phi _{g} =\sum _{i=1}^{H}\sum _{j=1}^{V}D_{g} (i,j)\tag{5}\end{equation*} View Source\begin{equation*} \phi _{g} =\sum _{i=1}^{H}\sum _{j=1}^{V}D_{g} (i,j)\tag{5}\end{equation*}
The entropic measure
of$\beta _{g} $ is also given as$g(i,j)$ where\begin{align*} \beta _{g} =\begin{cases} \displaystyle -\sum _{m}\Omega _{m} \log (\Omega _{m}) & {\mathrm{ for\; }}\Omega _{m} \ne 0 \\ \displaystyle 0 & {\mathrm{ for\; }}\Omega _{m} = 0 \end{cases}\tag{6}\end{align*} View Source\begin{align*} \beta _{g} =\begin{cases} \displaystyle -\sum _{m}\Omega _{m} \log (\Omega _{m}) & {\mathrm{ for\; }}\Omega _{m} \ne 0 \\ \displaystyle 0 & {\mathrm{ for\; }}\Omega _{m} = 0 \end{cases}\tag{6}\end{align*}
is the frequency of pixels having gray levels in the histogram bin$\Omega _{m} $ .$m=1,\ldots,256$ The PSNR
of$\rho _{g} $ is obtained as$g(i,j)$ where\begin{equation*} \rho _{g} =10\log _{10} \left [{\frac {\left ({L-1}\right)^{2} }{MSE} }\right]\tag{7}\end{equation*} View Source\begin{equation*} \rho _{g} =10\log _{10} \left [{\frac {\left ({L-1}\right)^{2} }{MSE} }\right]\tag{7}\end{equation*}
is the maximum pixel intensity value in$L$ and the mean squared error$g(i,j)$ is given as$MSE$ \begin{equation*} MSE=\frac {1}{H\times V} \sum _{i=1}^{H}\sum _{j=1}^{V}\left |{ f(i,j)-g(i,j)}\right |^{2}\tag{8}\end{equation*} View Source\begin{equation*} MSE=\frac {1}{H\times V} \sum _{i=1}^{H}\sum _{j=1}^{V}\left |{ f(i,j)-g(i,j)}\right |^{2}\tag{8}\end{equation*}
Based on the parameters computed in (3)–(7), the evaluation function
is stated as$\mathcal {E} $ where\begin{equation*} {\mathcal { E}} = 1-\exp \left ({-\frac {\rho _{g} }{100} }\right)+\frac {N_{g} +\phi _{g} }{H\times V} +\frac {\beta _{g} }{8}\tag{9}\end{equation*} View Source\begin{equation*} {\mathcal { E}} = 1-\exp \left ({-\frac {\rho _{g} }{100} }\right)+\frac {N_{g} +\phi _{g} }{H\times V} +\frac {\beta _{g} }{8}\tag{9}\end{equation*}
is a linear combination of the normalized values of the different metrics described in (3)–(7). By normalizing each metric in (8), each parameter are bounded between 0 and 1. Thus, based on this linear combination, the evaluation function is described by a defined scale bounded between a minimum value of zero and maximum value of four. A minimum value of zero represents an entirely dark enhanced image, while a maximum value of four represents an entirely bright enhanced image.${\mathcal {E}} $
C. Metaheuristic Optimization Algorithms
This section discusses the rationale for selecting the algorithms explored in this study. The FCR concept is discussed, as well as the various MOAs and their associated input/output parameters.
1) Choice of MOAs
Some of the earliest MOAs are the evolutionary techniques, with the GA making its debut in 1975 [23]. Since then, a plethora of MOAs have been proposed in the literature. To this end, with an inclusion of the GA, we look at three generations of MOAs starting from the 1990s to the present, with a ten-year gap between each generation. The CMAES algorithm was chosen in part because it has been one of the top performers at the Genetic and Evolutionary Computation Conference (GECCO) between 2009 to 2018 [24], a conference which showcases the most recent high-quality advances in genetic and evolutionary computation. We also considered algorithms with less than five occasions where the objective function is computed within a single iteration of the algorithm. Thus, in the order of the year of development and the number of times the fitness function (
2) Fitness Computation Rate
The FCR is defined here as the number of times the fitness (i.e objective) function is computed during the execution of an MOA. In our study, such a fitness/objective function is termed the evaluation function, and both terms are used interchangeably.
In many MOA-based articles, the number of generations is often used as the termination criteria for an MOA. However, in recent publications, such as in [25]–[28], it has been argued that the number of generations does not guarantee fair comparison between different MOAs because an MOA may compute the fitness function more times than other MOAs. Consequently, it is fairer to use the FCR since it ensures that the number fitness evaluations consumed is kept constant while executing the different MOAs. However, because the FCR is often difficult to compute mathematically, it is ideal to measure it algorithmically during runtime [26], [27], which is what we have done for each MOA considered in our study as presented in their different pseudocodes (see Algorithms 1–9).
Algorithm 1 The ABC Algorithm for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
**Phase 1: Recruited bees stage**
Get new solution
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Phase 2: Onlooker bees stage**
for
Select a
Improve the
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Phase 3: Scout bees stage**
for
Obtain a new
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 2 The CSO Algorithm for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
**Phase 1: Get Cuckoo stage**
Get new parameters
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Phase 2: Abandon nest stage**
for
Empty a fraction of the worst solutions based on
Update the new solutions
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 3 The DE Algorithm for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
Select three solutions
Get new solution
Obtain a uniform random number
if
Maintain new solution
else
Maintain previous solution
end if
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 4 The Firefly Algorithm for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
for
Compute the distance
Get new solution
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 5 The Genetic Algorithm for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
Performance selection using the roulette wheel mechanism **Stage 2: Crossover phase**
Apply crossover based on selected parents to obtain new solution
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
Select new parents randomly
Apply mutation using (18) to obtain new solution
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 6 The Particle Swarm Optimization for MIE
See Table 1
**Initialization phase**
for
Generate random parameter values
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
**Main loop begins here**
while
for
Update velocity of particle using (20)
Obtain new solution
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 7 The CMAES for MIE
See Table 1
**Initialization phase**
Initialize the CMAES parameters
**Main loop begins here**
while
for
Sample new solutions
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
Sort the population of solutions
Save present mean as
Update the mean to obtain new mean values
Update
Update
Update
Update
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 8 The Whale Optimization Algorithm for MIE
See Table 1
**Initialization phase**
**Main loop begins here**
while
Randomly initialize
for
if
if —A— < 1 then
Update the position of current search agent to obtain
else if —A—
Select a random search agent (
Update the position of current search agent to obtain
end if
else if
Update the position of the current search agent to obtain
end if
Perform image transformation using (2) based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
end for
Update
end while
Save best value and terminate process if
Post process and visualize the results
Algorithm 9 The Grey Wolf Algorithm for MIE
See Table 1
**Initialization phase**
Initialize
Generate initial random solutions
while
for
Perform image transformation based on
Compute fitness
If
Save best solutions (i.e.
Terminate process if
Update
Update
Update
end for
Update
Update
Update
Update new solution
end while
Save best value and terminate process if
Post process and visualize the results
3) Common and Specific Input and Output Parameters
The different MOAs considered in our study are governed by different input parameters, some of which are common across all MOAs. Thus, we define these parameters in Table 1 including their respective values obtained after a manual fine-tuning process. These parameter values are thus the best performing values, which were used for the MIE comparative exercise.
For the common inputs, there were a total of four different parameters to be optimized in our study (i.e.
4) Artificial Bee Colony
The ABC algorithm is modeled based on the intelligent foraging behavior of the honey bee swarm [29]. It follows the process by which sets of honey bees (i.e swarm) accomplish their tasks successfully through social cooperation. Technically, the ABC algorithm employs three types of bees in the swarm, namely: employed (recruitment) bees, onlooker bees, and scout bees. The concept of a swarm corresponds to the total population
The ABC algorithm initializes by generating a random population of solutions, then image transformation is performed per \begin{equation*} X^{new}_{p} = X^{pr}_{p} + \phi _{p} \times (X^{pr}_{p} + X^{pr}_{k})\tag{10}\end{equation*}
\begin{equation*} Pr_{p} = \frac {E_{p}}{\sum _{i}E_{i}}\tag{11}\end{equation*}
In the third (scout bees) stage, the algorithm determines a new \begin{equation*} X^{new}_{p} = L_{d} + \phi _{p,d} \times (U_{d} - L_{d})\tag{12}\end{equation*}
5) Cuckoo Search Optimization
The CSO algorithm is discussed following its presentation in [30], and interested readers can access further details in the same reference. The algorithm mimics the biological hatching process of the cuckoo bird. Biologically, some cuckoo bird species typically lay their egg(s) in a foreign bird’s nest in order to be hatched by the foreign bird. In translating this concept into an algorithm, the cuckoo’s egg is denoted as a solution where all the eggs in the nest of the foreign bird constitute a population of solutions (i.e. many eggs). Thus, the algorithm iterates by replacing the not-so-good solutions in the population with better solutions. To achieve this, it uses a \begin{equation*} X_{p}^{new}=X_{p}^{pr}+\alpha \otimes L\dot {e}vy(\lambda) \tag{13}\end{equation*}
The CSO algorithm for MIE is presented in Algorithm 2, with an aim to show how the FCR is computed within the algorithm. Similar to most MOAs, an initial population of solutions is randomly generated in the initialization phase (see Algorithm 1) wherein each
6) Differential Evolution
The DE algorithm proposed in [31] works using a population of candidate solutions, called agents. These agents explore the search space towards an optimal solution by combining the positions of existing agents within the population. Similar to other MOAs, it adopts an initialization phase where starting solutions are generated randomly. The algorithm then selects three agents (i.e. solutions) \begin{equation*} X^{new}_{p} = a_{p} + \beta \times (b_{p} - c_{p})\tag{14}\end{equation*}
Then, in the second (crossover) stage, the algorithm updates the
7) Firefly Algorithm
The firefly algorithm works by modeling the flashing behavior of fireflies. It was proposed in [32] based on the concept that fireflies (i.e. solutions) will be attracted within the population based on an attractiveness property, which is proportional to their brightness. Thus, less brighter fireflies will migrate towards brighter ones. Essentially, the algorithm updates any pair of fireflies \begin{align*} X^{new}_{p} \!=\! X^{pr}_{p} + \beta \times exp[-\gamma \times r^{2}_{p,k}] \times (X^{pr}_{p} - X^{pr}_{k}) + \alpha \times \epsilon \\{}\tag{15}\end{align*}
First, an initial population of solutions is generated similar to the other MOAs considered in our study. The distance \begin{equation*} r_{p,k} = \frac {norm(X_{p}^{pr} - X_{k}^{pr})}{d_{max}}\tag{16}\end{equation*}
\begin{equation*} d_{max} = norm(U - L)\tag{17}\end{equation*}
8) Genetic Algorithm
The GA is modeled after the natural selection process and it is intended to produce high-quality solutions to optimization problems [23]. Similar to other MOAs, it begins with the initialization phase where candidate solutions are generated randomly and their respective fitness values are computed.
In the first phase (i.e. the selection phase), the fittest solutions are selected as the parent indices based on the roulette wheel mechanism. The selected parents (i.e. candidate solutions) are then subjected to the crossover operator in the crossover stage in order to produce new offsprings (i.e. new solutions)
In the mutation phase, a number of solutions are randomly selected and their values are changed to form new solutions as follows \begin{equation*} X^{new}_{p} = X^{pr}_{p} + \sigma \times randn(D)\tag{18}\end{equation*}
\begin{equation*} \sigma = 0.1 \times (U - L)\tag{19}\end{equation*}
9) Particle Swarm Optimization
The PSO algorithm models the movement of particles (i.e. candidate solutions) over a search space towards converging to an optimal solution. The evolution of each solution (or particle) in the population (i.e. a swarm) is governed by its local best position and guided towards better solutions via a velocity parameter. Essentially, the algorithm updates a particle’s velocity using \begin{align*}&\hspace {-.5pc}V^{new}_{p} = \omega \times V^{pr}_{p} + c1 \times rand(D) \times (X^{best}_{p} - X^{pr}_{p}) \\&+ c2 \times rand(D) \times (G^{best} - X^{pr}_{p})\tag{20}\end{align*}
\begin{equation*} X^{new}_{p} = X^{pr}_{p} + V^{new}_{p}\tag{21}\end{equation*}
Then, the new solutions
10) Covariance Matrix Adaptive Evolutionary Strategy
The covariance matrix adaptation evolution strategy (CMAES) is a type of numerical optimization method that falls under the umbrella of evolutionary algorithms and evolutionary computation. In the CMAES algorithm proposed in [33], new candidate solutions are sampled from a multivariate normal distribution in
The adaptation of the covariance matrix is analogous to learning a second order model of the underlying objective function in classical optimization, comparable to the approximation of the inverse Hessian matrix in the quasi-Newton technique. In contrast to most conventional approaches, the CMAES method makes less assumptions about the underlying objective function. The approach does not need derivatives because it just employs a ranking (or, equivalently, sorting) of candidate solutions.
The main loop of the algorithm is divided into three sections: 1) sampling of new solutions, 2) re-ordering (sorting) of sampled solutions based on fitness, and 3) updating of internal state variables based on the re-ordered samples. The algorithm’s pseudocode is given in Algorithm 7. The sequence of the update assignments matters:
in$M_{p} $ , the mean of the distribution and the current preferred solution to the optimization problem,$\mathbb {R}^{D}$ , the step-size$\sigma _{p}>0 $ , a symmetric and positive-definite$C $ covariance matrix with the initialize matrix given as an identity matrix$n\times n$ and$C_{0}=I$ in$p_{\sigma } $ ,$\mathbb {R}^{D}$ in$p_{c} $ , two evolution paths, initially set to the zero vector.$\mathbb {R}^{D}$
The iteration starts with sampling \begin{align*} X_{p}^{new}\sim&{\mathcal {N}}(M_{p},\sigma _{p}^{2}C_{p}) \tag{22}\\\sim&M_{p}+\sigma _{p}\times {\mathcal {N}}(0,C_{p}) \tag{23}\end{align*}
After generating new solutions across the entire population, then the solutions are sorted in descending order of fitness values. The previous mean values \begin{align*} M^{new}=&\sum _{p=1}^{\mu }w_{p}\,X^{sort}_{p:P }\tag{24}\\=&M^{pr}+\sum _{p=1}^{\mu }w_{p}\,(X^{sort}_{p:P } - M^{pr}) \tag{25}\end{align*}
11) Whale Optimization Algorithm
The whale optimization algorithm (WOA) is a relatively newer optimization algorithm proposed in [34] that mimics the natural hunting mechanism of humpback whales. Humpback whales can detect the presence of prey and encircle it. Because the optimum design’s position in the search space is unknown at the outset, the WOA algorithm for MIE presented in Algorithm 8 assumes that the current best candidate solution is the target prey or is near to it. When the best search agent is determined, the other search agents will attempt to update their locations in relation to the best search agent. The following equations illustrate this behavior:\begin{align*} {D}=&|{C}\bullet { {X_{p}^{pr}}}-{X^{best}}| \tag{26}\\ {X}^{new}_{p}=&{ {X_{p}^{pr}}}-{A}\bullet {D}\tag{27}\end{align*}
\begin{align*} {A}=&2{a}\bullet { {r_{1}}}-{a} \tag{28}\\ {C}=&2{ {r_{2}}}\tag{29}\end{align*}
\begin{align*} {D}=&|{C}\bullet { {X_{rand}}}-{X^{best}}| \tag{30}\\ {X}^{new}_{p}=&{ {X_{rand}}}-{A}\bullet {D}\tag{31}\end{align*}
\begin{equation*} {X}^{new}_{p}={D'}e^{bt}cos(2\pi t)+{ {X^{best}}}\tag{32}\end{equation*}
12) Grey Wolf Optimization Algorithm
The GWO algorithm proposed in [35] simulates the natural leadership structure and hunting mechanism of grey wolves. For replicating the leadership structure, four sorts of grey wolves are used: alpha, beta, delta, and omega. In addition, three key hunting processes are incorporated to optimize performance: seeking for prey, surrounding (i.e. encircling) prey, and attacking prey. When building GWO, the fittest solution is designated as the alpha (
Grey wolves, as previously said, encircle victims during the hunt. The following equations are presented to quantitatively model their encircling behavior:\begin{align*} {D}=&|{C}\bullet { {X_{p}^{pr}}}-{X^{best}}| \tag{33}\\ {X}^{new}_{p}=&{ {X_{p}^{pr}}}-{A}\bullet {D}\tag{34}\end{align*}
\begin{align*} {A}=&2{a}\bullet { {r_{1}}}-{a} \tag{35}\\ {C}=&2{ {r_{2}}}\tag{36}\end{align*}
\begin{equation*} a = 2 - \frac {2*FCR}{FCR_{max}}\tag{37}\end{equation*}
\begin{align*} { {D_{\alpha }}}=&|{ {C_{1}}}.{ {X_{\alpha }}}-{ {X^{pr}_{p}}}| \tag{38}\\ { {D_{\beta }}}=&|{ {C_{2}}}.{ {X_{\beta }}}- {{X^{pr}_{p}}}| \tag{39}\\ { {D_{\delta }}}=&|{ {C_{3}}}.{ {X_{\delta }}}- {{X^{pr}_{p}}}| \tag{40}\\ { {X_{1}}}=&{ {X_{\alpha }}}-{ {A_{1}}}.({ {D_{\alpha }}}) \tag{41}\\ { {X_{2}}}=&{ {X_{\beta }}}-{ {A_{2}}}.({ {D_{\beta }}}) \tag{42}\\ { {X_{3}}}=&{ {X_{\delta }}}-{ {A_{3}}}.({ {D_{\delta }}}) \tag{43}\\ {X^{new}_{p}}=&{\frac {{ {X_{1}}}+{ {X_{2}}}+{ {X_{3}}}}{3}}\tag{44}\end{align*}
In the attacking prey phase, when the target (i.e. prey) stops moving, the grey wolves conclude the hunt by attacking it. To mathematically depict approaching the prey, the value of
To summarize the algorithmic process as presented in Algorithm 9, the search process begins with the GWO algorithm producing a random population of grey wolves (potential solutions). Alpha, beta, and delta wolves determine the likely position of the prey during the duration of iterations. Each potential solution changes its position with relation to the prey. The value of
D. Dataset
Medical images selected from the MedPix database were used in our study. The MedPix database is a free open online database containing medical images, clinical topics, teaching cases, and textual metadata. It consists of about 12,000 patient case scenarios, 9000 topics, and about 59,000 images. The images within the database are grouped into categories such as the brain and neuro, cardiovascular, chest, pulmonary, eye and orbit, and abdomen [36]. In this study, we have randomly selected images as representative samples from across the human body, which are classed under three use cases described as follows:
Use case 1: This image comprises an oral contrast-enhanced axial computed tomographic (CT) scan of a distended appendix as shown in Figure 1(a). The appendix can be seen in the area identified by the blue dotted arrow in Figure 1(a), which depicts a distended appendix having thickened enhancing walls and periappendiceal inflammatory changes with fat stranding. This image (of Figure 1(a)) was obtained freely from the MedPix database and can be searched for and downloaded using the code name “synpic28644”.
Use case 2: This image presents the CT image of a brain presenting the case of an arteriovenous malformation of the right occipital lobe. The original image is shown in Figure 3(a) with the region of interest enclosed by the dotted yellow box wherein the vascular mass can be seen. It is noted that this image (of Figure 3(a)) can be downloaded freely from the MedPix database using the code name “synpic51882”.
Use case 3: This use case presents a lateral digital subtraction angiography view of the external carotid artery, showing the filling of the right external carotid artery branches obtained after the injection of some contrast medium. The original image from this process is shown in Figure 5(a), which can be searched for and downloaded from the MedPix database using the code name “synpic15935”.
Use case 2: Original and enhanced CT images of the right occipital lobe of a brain.
Use case 3: Original and enhanced DSA images of the right carotid artery injection of a brain.
E. Performance Metrics
In order to adequately assess the effectiveness of any MIE technique, it is essential to define the quantitative measures used in the evaluation process. To this effect, the metrics considered include the number of edges, number of pixels in the fore-ground, entropic measure, and the peak signal to noise ratio, which can be obtained as previously stated in (3), (5), (6), and (7), respectively. Also, the absolute mean brightness error (AMBE) \begin{equation*} \xi =\left |{\delta (f(i,j))\left.{ -\delta (g(i,j))}\right |}\right.\tag{45}\end{equation*}
\begin{align*} \delta (f(i,j))=&\frac {1}{HV} \sum _{i}\sum _{j}\delta (f(i,j)) \tag{46}\\ \delta (g(i,j))=&\frac {1}{HV} \sum _{i}\sum _{j}\delta (g(i,j))\tag{47}\end{align*}
Results and Discussion
The MOAs considered in our study were evaluated using different medical images selected from different regions of the body. The qualitative and quantitative results obtained are presented and discussed in this section. By qualitative results, we imply the visual assessment of the enhanced images from the viewpoint of a radiologist or any human observer. Such an assessment is based upon the salient differences between the different enhanced output images and the original input image.
In terms of quantitative outcomes, we begin by discussing the fitness evolution trend of each method per input image. The fitness trends are provided in terms of the best, worst, and mean fitness values derived as a function of the FCR from 1000 independent trials (i.e. Monte Carlo experiments). In addition, the values of each measure produced by each method are displayed and analyzed in relation to the corresponding enhanced images. For each approach, we then show the final optimized parameter values that were utilized in the transformation function to generate the various enhanced images. The physical timing performance of each method and a thorough statistical significance tests of the results per use-case are discussed in the last part of this section.
A. Use-Case 1: Image of Acute Appendicitis
We applied each algorithm to use case 1, where the area of interest to a radiologist is depicted by the magnified region in Figure 1(a), which depicts a distended appendix having thickened enhancing walls and periappendiceal inflammatory changes with fat stranding. Following the application of the different algorithms to the image of Figure 1(a), we present the enhanced output images per MOA in Figures 1(b)–1(j), respectively.
1) Qualitative Analysis
It can be seen that the enhanced images in Figures 1(b)–1(j) demonstrate an improved contrast and brightness level as against the original image. Visually, the original image presents a blurrier and less detailed CT scan. Such improvement in the contrast level in Figures 1(b)–1(j) implies that the boundaries of the different organs were better delineated in the enhanced images than in the original image.
Specifically, the zoomed area of each image shows that the thickened walls of the distended appendix were better enhanced with an improved intensity and brightness around the boundaries of the wall protruding to the caecum. This ensures that the radiologist is better informed about the degree of distension around the appendix as compared to using only the original image. Furthermore, the oral contrast liquid used during the diagnostic procedure as seen by the whitish dispersion in the original image are more vivid in the enhanced images. Additionally, the details of the fat stranding in the appendix region are better pronounced in the enhanced images than in the original image, thus allowing a radiologist to provide an improved diagnosis.
Looking closely at the different enhanced images, it can be seen that the ABC algorithm produced an overtly brighter image than the other methods, which caused the region of interest to be more obscured than expected. At first glance, it may be difficult to visually determine the best enhanced image, however, by enlarging the images and visually parsing through them in a slideshow manner, we came to the following opined conclusions:
The ABC algorithm produced a slightly blurrier image than the other algorithms, with a generally lesser feel of texture and provision of details. Thus, the ABC algorithm performed least in this use case.
The enhanced images produced by the DE and Firefly algorithm were visually indistinguishable, thus implying that both algorithms achieved similar performance levels.
It is generally difficult to distinguish visually between the performances of the different MOAs as it concerns the enhancement of the present use case image.
Following the above observations, we shall next examine and analyse the quantitative results per MOA towards arriving at some substantial conclusions.
2) Quantitative Analysis
a: Fitness Evolution Performance
Figure 2 presents the fitness evolution graphs of the different MOAs as a function of the FCR. There are three different results presented in this regard, including the best, worst, and mean fitness evolution performance results shown in Figures 2(a)–2(c), respectively. We note that Monte Carlo (MC) simulation consisting of 1000 different MC trials were used to arrive at the different results of Figure 2. In particular, Figure 2(a) represents the best fitness evolution results obtained per algorithm from within a set of 1000 different MC trials. It is worth noting that the different enhanced images presented in Figure 1 are the outcomes of this best fitness value graphed in Figure 2(a). On the other hand, the output images produced by the worst and mean fitness performance processes of each algorithm are not presented here since the evolution trends of Figures 2(b) and 2(c) are sufficient to describe their different performance levels. The worst performance graph of Figure 2(b) corresponds to the lowest fitness values obtained after conducting 1000 MC trials per algorithm. Whereas, we computed the mean results of Figure 2(c) by averaging the fitness evolution graphs over 1000 MC trials.
Generally, it can be seen from Figure 2 that the performance rate of each algorithm improves with an increase in the FCR. Such an increase in the performance of each algorithm is expected since only better solutions (i.e better enhanced images) are retained whenever the fitness function is computed. In particular, it can be seen in Figure 2(a) that after computing the fitness function 200 times, the different algorithms can be rated as follows: the CMAES algorithm achieved the highest fitness value, followed closely by the GWO and WOA, then the PSO followed by the DE, Firefly, GA, CSO, and the ABC algorithm. We can also see in Figure 2(a) that there exists a small difference (
Based on an average of 1000 different MC trials, Figure 2(c) shows that the GWO algorithm provides the best performance followed by the WOA algorithm. Essentially, the figure reveals that there exists only a marginal difference (< 0.008) in the fitness values of the best and least performing algorithm. This thus leads to the conclusion that there may be no significant difference in using one MOA over another in this use case.
b: Other Performance Metrics
Although the fitness evolution performance graphs of Figure 2 have provided an overall assessment of the different algorithms, nevertheless, it is worthwhile to further analyse other specific outcomes regarding the quality of the enhanced images. Consequently, a few notable metrics were measured and presented in Table 2. It can be seen in Table 2 that the CMAES algorithm achieved the best results (see bold values in Table 2) in terms of its detailed variance and number of edges, whereas the GA and WOA achieved the best results in terms of the peak SNR and entropy, respectively.
Interestingly, the AMBE result of the ABC algorithm (i.e. AMBE = 0.0235) in Table 2 indicates that the ABC algorithm contributed the largest brightness value compared to the other algorithms. Such a high AMBE value corroborates the visual conclusions drawn in section III-A1, which states that the ABC algorithm produced an overtly brighter enhanced image. This high AMBE value ultimately caused the ABC algorithm to over brighten the region of interest as compared to the other algorithms (see Figure 1(b)), thus diminishing its overall visual performance.
Again, we cannot over emphasize the fact that there was really no significant difference in the individual metric values achieved by the different algorithms. For example, there was no significant difference in the supposed better performance of the DE algorithm over the Firefly algorithm (see close values of both algorithms in Table 2). These close values of both algorithms typically corroborates our visual conclusions drawn in section III-A1, which affirms that there was no significant difference between the DE and Firefly algorithms.
Additionally, our conclusions about the insignificant differences in the performance levels of the different algorithms can be further emphasized based on the number-of-edges metric (see third column in Table 2). In this case, although the DE and Firefly algorithms may have enhanced more edges than the ABC algorithm, it should nevertheless be noted that these edges could as well have been enhanced in the wrong regions of interest. For example, the dividing line seen clearly in the caecum region of Figure 1(a) (see red line) becomes difficult to discern in all the enhanced images due to excessive brightening of the pixel edges. Thus, it is possible that introducing better results in one dimension of an enhanced image does not necessarily imply a better overall enhanced image. Consequently, we suggest that one single MOA may not necessarily provide the best enhanced image with regards to all constituting metrics of assessment.
We present in Table 3 the final optimized parameter values of the transformation function as estimated by the different MOAs. These values are reported here so that the final enhanced images shown in Figures 1(b)–1(g) can be independently reproduced for validation purposes. These values are also presented as they provide some plausible explanation for the different outcomes obtained per algorithm. For example, we can explain why the ABC algorithm produced an overtly bright image since it converged to a larger
B. Use-Case 2: Image of Right Occipital Lobe
Each algorithm was applied to use case 2 and the corresponding enhanced images outputted per algorithm are presented in Figures 3(b)–3(j).
1) Qualitative Analysis
We observe from Figures 3(b)–3(j) that the output CT images are better enhanced compared to the original image in Figure 3(a). Specifically, the finer details within the vascular mass bounded by the yellow-dotted box are well delineated in the enhanced images. Thus, this reveals to the radiologist a more vivid display of the degree of growth of the vascular mass as compared to using the original image for diagnostic purposes.
In terms of the performance of the different MOAs, it can be seen from closer examination of the enhanced images that the GA algorithm introduced the largest degree of brightness. Whereas, the CMAES algorithm produced the darkest enhanced image. Such a dispersion in the brightness level of the different enhanced images did not affect the region of interest, neither was any other region of the CT image affected. Thus, such negligible differences in the contrast and brightness levels of Figures 3(b)–3(g) makes it qualitatively difficult to determine the best enhanced image with regards to the region of interest.
2) Quantitative Analysis
Here, an attempt is made to quantitatively appraise the performance levels of the different MOAs based on the following specific metrics:
a: Fitness Evolution Performance
First, it is worth mentioning that the enhanced images of Figures 3(b)–3(j) correspond to the outcome of the best fitness evolution per MOA out of 1000 MC trials graphed in Figure 4(a). From Figure 4(a), it is noted that the different MOAs converged to very close fitness values after computing the fitness function 200 times. Such close fitness values (< 0.045) between the best (i.e the CMAES algorithm) and the least (i.e the CSO algorithm) affirms why it was qualitatively difficult to determine the best enhanced image in Section III-B1. From the best performance graph of Figure 4(a), it is noted that the CMAES algorithm suffices as the best performing algorithm towards the end range of the FCR. This implies that if we had shortened the FCR (probably to obtain output images faster), the WOA algorithm instead of the CMAES algorithm would have yielded the best enhanced image within such a simulation period.
Nevertheless, we note that such a high fitness value of the CMAES algorithm may not always be guaranteed in all trials, particularly as noted in the worst performance curve of Figure 4(b). Note that the curve of Figure 4(b) corresponds to the worst output result obtained out of 1000 different MC trials. Thus, in this single trial, it can be seen that the WOA algorithm yielded the highest fitness value and closely followed by the PSO, DE, and GA algorithms. Since such extreme cases (i.e best and worst case of Figures 4(a) and 4(b), respectively) may not provide the best assessment across 1000 different MC trials, thus, we provide the mean performance curves computed over 1000 MC trials as shown in Figure 4(c). Such a reliable fitness evolution statistics of Figure 4(c) reveals that the GWO algorithm generally yielded the highest fitness value (i.e best enhanced image), whereas the CSO and CMAES algorithms sufficed as the least performers.
b: Other Performance Metrics
We further examined the enhanced images of Figures 3(b)–3(j) based on the performance metrics that constitute the overall fitness value. These results are documented in Table 4. Interestingly, the GA algorithm yielded the largest AMBE value, which is the metric that quantifies the degree of brightness of an enhanced image. Such a large AMBE value of the GA algorithm accurately affirms the conclusion drawn in Section III-B1, which states that the GA algorithm yielded the brightest enhanced image. Similarly, we note that the CMAES algorithm produced the least AMBE value, which affirms that it yielded the darkest enhanced image (see Figure 3(h)).
Although the CMAES algorithm may have yielded the largest value for the detailed variance, number of edges, and the peak SNR, nevertheless, going by the fitness value, it was not ranked as the best performing algorithm going by the mean fitness graph of Fig 4(c). This is due to the fact that, as a statistical method, such MOAs cannot be evaluated by a single trial of the MIE procedure, and so numerous trials are frequently necessary to pick the optimum performance. We also noticed that the metric values acquired across the multiple MOAs were not significantly different, which resulted in significantly near fitness scores across all methods. As a result, such near fitness values confirm the difficulties in visually differentiating the best improved image in Figure 3. Finally, Table 4 shows that no one MOA attained the greatest value across all measures, showing that it is nearly impossible to claim any MOA as the top performer across all potential use cases.
The optimized parameters obtained by the different MOAs are reported in Table 5 for the sake of repeatability. We remark that these are the parameter values used to generate the enhanced images shown in Figures 3(b)–3(j). For instance, we noticed that the GA algorithm’s large
C. Use-Case 3: Image of Right Common Carotid Artery Injection
This use case was processed using the different MOAs and the respective enhanced images by the different algorithms are presented in Figures 5(b)–5(j).
1) Qualitative Analysis
A radiologist’s major concern in analyzing the various images of Figure 5 is to identify whether there is any indication of opacification of the internal carotid artery (ICA) or its branches. Although there was no indication of opacification because the contrast medium can be seen to be clearly disseminated over the ICA, any viewer can perceive that the original image has a blurrier sight than the enhanced images. As a result, several of the arteries in Figure 5(a) are made nearly invisible to the human eye when compared to the enhanced images in Figures 5(b)–5(j). To that end, there is no doubt that the enhanced images from the various MOAs provide a radiologist with a more detailed perspective than viewing the original image.
It may be challenging to determine the best improved image based just on visual examination. However, when it comes to the worst performance, it is clear that the contrast of Figure 5(b) generated by the ABC algorithm is the blurriest. It displays a lot of weak arterial lines at the image’s border, resulting in fewer characteristics for analysis. To summarize, while the original image was definitely enhanced across the multiple MOAs, there is very little to separate visually the performance of the different methods.
2) Quantitative Analysis
The quantitative performance of the different MOAs are analysed based on the following metrics:
a: Fitness Evolution Performance
Figures 6(a), 6(b), and 6(c) show the findings of the best, worst, and mean fitness values as a function of the FCR per MOA. We examine Figure 6(a) because the enhanced images in Figure 5 are derived from these data after computing the fitness function 200 times. Figure 6(a) shows that remarkably near fitness values were obtained across the different MOAs, justifying the indistinguishable visual performance of the various enhanced images. Nonetheless, the CMAES, DE, and PSO algorithms are shown to have produced the best fitness values, thus resulting in the better looking enhanced images of these algorithms as shown in Figures 5(h), 5(d), and 5(g), respectively.
Figure 6(b) presents the lowest fitness performance of the different algorithms obtained out of 1000 MC trials. At an FCR of 200, we found that the GWO algorithm produced the largest fitness value, followed closely by the WOA, ABC, and Firefly algorithms. This means that determining which MOA would perform best based on a single run of the algorithm is often challenging since randomized results may suffice during the initial iterations of the different algorithms. However, after an average of 1000 MC trials, we can observe in Figure 6(c) that the GWO and PSO algorithms surpassed the DE and other algorithms in terms of their fitness values. Despite these findings, we infer that the fitness values obtained by the algorithms are not significantly different, which explains the difficulty in visually choosing the best improved image.
b: Other Performance Metrics
Table 6 shows the various performance measures that make up the fitness values. Based on the detailed variance values and the number of edges, an interesting finding is made, revealing that the ABC algorithm yielded the lowest values. These values explain why the ABC algorithm produced the most blurry enhanced image. Again, the ABC algorithm’s lower visual performance is supported by its lowest entropy value, which explains the absence of adequate information (i.e. blurriness) in its enhanced image. The CMAES, DE, Firefly, and PSO algorithms, on the other hand, provided the highest performance values across all measures, which confirms their somewhat improved visual performance in Figures 5(h), 5(d), 5(e) and 5(h), respectively.
Table 7 presents the optimum parameter values of the transformation function as computed by the different MOAs and utilized to create the enhanced images of Figures 5(b)–5(j). These parameters are provided so that the various improved images of Figure 5 may be recreated and confirmed independently by any interested researcher. In particular, both the DE and PSO algorithms converged to similar parameter values, which explains why they produced comparable improved images with identical output performance metrics.
D. CPU Timing Performance
The timing performance of the different algorithms were compared and reported per use case in Table 8. For this purpose, all simulations were conducted using a PC running an Intel(R) Core i5-7500 CPU processor @ 3.40 GHz with an installed memory (RAM) of 16GB. To ensure that the timing experiments were relatively valid, the following measures were taken:
All algorithms were encoded and simulated in MATLAB 2017a.
All use-case images were resized to the same size (i.e.
)$256\times256$ The population size and FCR were fixed for all MOAs.
The start time for measuring the timing performance of each algorithm was taken at the initialization of each algorithm and was ended strictly after the main loop was completed.
It was ensured that only the MATLAB software was kept running as the only foreground process during each simulation period in order to ensure that no extra processing time was incurred by the PC.
Finally, the timing results reported in Table 8 were obtained over an average of 1000 MC trials per use-case.
In terms of the run-time of each MOA, we observed that the GA algorithm converged the quickest in the first and third use case, whereas the CSO algorithm came quickest in the second use case. Despite these algorithms being the fastest, we note that there was less than an average difference of 0.1 - 0.3 seconds in the computing times of the different algorithms. This implies that when the FCR is kept fixed and equal for all MOAs, then there exists little or no significant difference in their respective real time computational speeds.
Statistical Significance Tests Based on the Use-Cases
We performed statistical significance tests on the mean outcomes over 1000 independent trials. Instead of merely using the best results, we used the mean outcomes, which ensures that a more rigorous analysis is achieved. To this end, we ran the Brown-Forsythe and Welch ANOVA tests on the mean outcomes per use-case, as well as a posthoc comparison of the various MOAs using Games-Howell’s multiple comparison tests. Our collected findings are therefore provided in Tables 10–13 for the different use-cases. The Games-Howell’s test was chosen because it is appropriate for comparing two or more sample populations and, unlike the Tukey’s test, it does not presuppose homogeneity of variances or equal sample sizes, which better suits our test results.
The first column of Tables 10–13 mentions the paired algorithms compared with, while the second column shows the difference in the means of both algorithms. A positive or negative mean difference in the second column indicates that the mean of the first listed algorithm is bigger or smaller than the mean of the second algorithm, respectively. The confidence interval of the mean difference between the matched algorithms is shown in the third column. The fourth column indicates whether the estimated
Table 10 for use-case 1 shows that, with the exception of the CMAES, WOA, and GWO algorithms, there was no statistical significance between the different MOAs. In this situation, the GWO method is acknowledged to have obtained the best mean outcomes and is thus thought to be significantly distinct from all other MOAs. Nonetheless, no significant difference was found between the GWO and WOA, indicating that both algorithms function equally. As a result, only the GWO and WOA algorithms produced a highly significant difference in empirical performance when compared to the other approaches.
Table 11 shows that, with an exception of the GWO, WOA, and CMAES algorithms, there was no significant difference between the other MOAs investigated in our study for use-case 2. We notice that such a difference in mean values might be either positive or negative; hence, the GWO algorithm obtained positive differences when compared to the others, suggesting that it performed better than the others.
The findings for use-case 3 shown in Table 12 indicate that there was no significant statistical difference between the other algorithms, save for the GWO and WOA algorithms. Nevertheless, it should be noted that these results were computed throughout the complete FCR range of values (i.e. 1 - 200), thus demonstrating that the algorithms performed similarly in half of the comparison list (see Table 12). Essentially, it should be highlighted that, on average, across 1000 separate trials, the GWO and WOA algorithms performed significantly better than the other approaches. It also shows that there was no substantial difference between the GWO and WOA algorithms, meaning that either algorithms can be employed. It also shows that there was no statistically significant difference in the performance of the ABC, CSO, DE, firefly, GA, and PSO algorithms. As a result, any algorithm from this collection of approaches can be put for usage.
Finally, we performed statistical significance tests between the run-times of the various algorithms, and the results are shown in Table 13. Because the number of samples per technique was fewer than 50 (see Table 8), the Dunnett’s T3 posthoc test was utilized, showing itself as a superior strategy than the Games-Howell’s test in this circumstance. It is clear that there was no substantial variation in the run-time of the different algorithms, which strongly supports the notion that adopting equal FCR across all approaches ensures fairness in comparison. As a result, when it comes to selecting an MOA for MIE on a run-time basis, any MOA can be chosen without sacrificing timing performance.
Conclusion
This article has investigated whether there exist any significant difference in the performance of different metaheuristic optimization algorithms (MOAs) as applied for medical image enhancement (MIE). To this effect, well-known MOAs were investigated for MIE based on some selected medical images obtained from the Medpix database. Fairness was achieved by terminating the different algorithms based on the fitness computation rate (FCR), which ensures that the objective function is consumed equally by all MOAs during their different algorithmic processes. From an empirical standpoint, and based on an average of 1000 Monte Carlo independent trials, the GWO and WOA algorithms achieved slightly better results than the other methods. However, there was no substantial difference in performance among the other MOAs, including the ABC, DE, CSO, GA, Firefly, and PSO algorithms. Furthermore, there was no significant difference in the performance of the GWO and WOA algorithms, thus suggesting that any of both algorithms can be used for MIE purposes. In terms of qualitative (i.e. visual) evaluation, there was little to differentiate the performance of the various algorithms. Furthermore, statistical analyses showed that there was no statistically significant difference in the computing times of the various methods. As a result, there may be no discernible difference in adopting one MOA over another for MIE applications. Nonetheless, the aim of this study merely seeks to inspire new and wider investigations based on a larger corpus of images, as well as based on the analysis of several other MOAs, with the goal to either substantiate or dispute the conclusions found herein.
ACKNOWLEDGMENT
The authors would like to thank the Tshwane University of Technology for their support. Opinions, findings, and conclusions or recommendations expressed in any publication generated by the NRF supported research are those of the author(s) alone, and the NRF accepts no liability whatsoever in this regard.