A Chaotic Oppositional Whale Optimisation Algorithm with Firefly Search for Medical Diagnostics

2022-08-24 12:58MilanTairNebojsaBacaninMiodragZivkovicandVenkatachalam
Computers Materials&Continua 2022年7期

Milan Tair, Nebojsa Bacanin, Miodrag Zivkovicand K.Venkatachalam

1Singidunum University, 32 Danijelova Str., 11010, Belgrade, Serbia

2Department of Applied Cybernetics, Faculty of Science, University of Hradec Králové, 50003, Hradec Králové,Czech Republic

Abstract: There is a growing interest in the study development of artificial intelligence and machine learning, especially regarding the support vector machine pattern classification method.This study proposes an enhanced implementation of the well-known whale optimisation algorithm, which combines chaotic and opposition-based learning strategies, which is adopted for hyper-parameter optimisation and feature selection machine learning challenges.The whale optimisation algorithm is a relatively recent addition to the group of swarm intelligence algorithms commonly used for optimisation.The Proposed improved whale optimisation algorithm was first tested for standard unconstrained CEC2017 benchmark suite and it was later adapted for simultaneous feature selection and support vector machine hyper-parameter tuning and validated for medical diagnostics by using breast cancer, diabetes,and erythemato-squamous dataset.The performance of the proposed model is compared with multiple competitive support vector machine models boosted with other metaheuristics, including another improved whale optimisation approach, particle swarm optimisation algorithm, bacterial foraging optimisation algorithms, and genetic algorithms.Results of the simulation show that the proposed model outperforms other competitors concerning the performance of classification and the selected subset feature size.

Keywords: Whale optimisation algorithm; chaotic initialisation; oppositionbased learning; optimisation; diagnostics

1 Introduction

Constructing algorithms to solve non-deterministic polynomial time hard problems (NP-hard)is not typically hard to do.However, executing such an algorithm can, in extreme cases, take thousands of years on contemporary hardware.Such problems are regarded as nearly impossible to solve with traditional approaches for finding a deterministic algorithm.Metaheuristics, as stochastic optimisation approaches, are useful for solving such problems.These approaches do not guarantee an optimum solution.Instead, they provide tolerable sub-optimal outcomes in a reasonable amount of time[1].The optimisation process aims to discover a solution that is either optimum or near-to-optimal regarding the stated goals and given constraints.Since the turn of the century, many populationba stochastic metaheuristics have been developed to address challenging optimisation issues in both combinatorial and global optimisation [2].

Stochastic population-based metaheuristics begin the search process using a set of random points, but the process is led and directed by some mechanism fromiteration to iteration.Natural systems and processes can be used to inspire the concepts that underpin the search mechanism.Some algorithms, such as genetic algorithms (GA) frequently become trapped in local optima.Hybridisation is one method for overcoming the aforementioned difficulties with metaheuristics.

The hybridisation should not be based on a random mix of hybridising algorithms, but rather on a specific analysis of the benefits and drawbacks of each method.Under some situations, the drawback of one algorithm (slow convergence, tendency to quickly become trapped in the local optimum, etc.)is overcome by combining with the structural portion of other algorithms.This can help fix common issues of the original optimal solution search method.

In contrast to deterministic optimisation methods that ensure the optimum answer, heuristic algorithms aim to discover the best possible solution, without ensuring that the optimal solution is obtained [3,4] Many practical and benchmarking problems cannot be solved using deterministic approaches as their solving is dawdling.Since heuristic algorithms are robust, academics worldwide have been interested in their improvement.This has led to the development of metaheuristics for a variety of problems.Metaheuristics is used to create heuristic techniques for solving different kinds of optimisation problems.Two main groups of metaheuristics are those which are inspired and which are not nature-inspired.Metaheuristics inspired by nature can be additionally categorised as evolutionary algorithms (EA) and swarm intelligence.Swarm intelligence has recently become a major research field and they are proven as methods that are able to achieve outstanding results in different fields of optimisation [5-8].

Besides metaheuristics, machine learning is another important domain of artificial intelligence.Notwithstanding good performance of machine learning models for various classification and regression tasks, there are still some challenges from this domain that should be addressed.Among two most important challenges are feature selection and hyper-parameters’optimisation.Performance of machine learning models to the large extend depend on the un-trainable parameters, which are known as the hyper-parameters and finding optimal or suboptimal values of such parameters for concrete problem is NP-hard task itself.Moreover, machine learning models utilize large datasets with many features and typically only subset of such features has significant impact on the target variable.Therefore, the goal of feature selection is to choose a subset of features and to eliminate redundant ones.

Study shown in this article proposes a new approach based on the well-known whale optimisation algorithm (WOA), which belongs to the group of swarm intelligence metaheuristics.Proposed improved WOA was first tested for standard unconstrained CEC2017 benchmark suite and it was later adapted for simultaneous feature selection and support vector machine (SVM) hyper-parameter tuning and validated for medical diagnostics by using breast cancer, diabetes, and erythematosquamous datasets.The contributions of this research are twofold.First, an improved version of the WOA, that addresses known deficiencies of the basic version by implementing the concepts of chaotic initialisation and opposition-based learning (OBL), and by incorporating the firefly search(FS), has been proposed, Second, proposed method was adopted for simultaneously optimisation of SVM hyper-parameters and feature selection and achieved better performance metrics than other contemporary classifiers for three medical benchmark datasets.

The rest of the manuscript is structured as follows.Section 2 presents an overview of the recent swarm intelligence applications.Section 3 introduces the WOA metaheuristics, their improvements,and proposes the hybrid SVM optimisation approach.Section 4 exhibits the experimental findings and the discussion.The last section brings the final remarks, proposes a future work, and concludes the article.The authors hope to provide an accurate feature detection method used for their diagnostics.

2 Background and Related Work

Swarm intelligence is based on natural biological systems, and function with a population of selforganising agents who interact with one another locally and globally with their environment.Although there is no central component that governs and directs individual behaviour, local interactions between agents result in the formation of globally coordinated behaviour.The swarm intelligence algorithm can produce quick, inexpensive and high-quality solutions for challenging optimisation problems [9].

In swarm intelligence algorithms, a swarm is a population consisting of simple homogeneous agents.The swarmmight have direct or indirect interactions.Direct interactions refer to audio or visual situations, touch, etc.In indirect interactions, agents are affected by the environment and changes made to it, which can be a result of other agents’actions, or from external influences.

The operators of evolutionary algorithms, notably selection, crossover and mutation, also employ swarm metaheuristics.The crossover operator is employed to look for quicker convergence in a single sub-domain.Mutations help avoid local optima and are used for random searching.The selection operator enables the system to move to the desired conditions.

Popular implementations of swarm intelligence algorithms are the particle swarm algorithm(PSO)[10], the ant colony optimisation (ACO) algorithm [11], the wasp swarm (WS) algorithm [12], the stochastic diffusion search (SDS) algorithm [13], the artificial bee colony (ABC) algorithm [14],the firefly search (FS) algorithm [15], whale optimisation algorithm (WOA) [16], the multi-swarm optimisation (MSO) algorithm [17], the krill herd (KH) algorithm [18], the dolphin echolocation (DE)algorithm [19], etc.Each of these algorithms, as well as many others which fall into this category, have strengths and weaknesses for different optimisation problems.One algorithm may outperform the other when applied to one problem and significantly under perform on another.

As already noted in Section 1, swarm intelligence has seen successful application to different areas of optimisation.Research proposed introduces a variety of the swarm intelligence algorithm, tested on Himmelblau’s nonlinear optimisation and speed reducer design problems, and used to attempt to solve problems such as the travelling salesman problem, the robot path planning problem etc.The study presented in [20] reviews earlier works and concludes that there is strong applicability of swarm intelligence algorithms for solving complex practical problems.Authors give overviews of the application of the swarm intelligence algorithm for swarm robotics.A chapter in [21] presents interesting insights and predictions about the use of the swarm intelligence algorithm for swarm robotics in the future.Study presented in [22] provides detailed reviews of the swarm intelligence algorithm and its practical implementation, from early conceptions in the late 80s onward.Aside from application in robotics, swarm intelligence algorithms have found application in computer networks[23,24], power management [25], engineering [26,27], Internet of things [28], social network analysis[29], medical diagnostics [30,31], data mining [32], etc.

Algorithms such as PSO and GA are known to get trapped in sub-optimal search space regions.Algorithms based on maritime animal behaviour are a relatively recent addition to the field of swarm intelligence algorithm research.Aside from the WOA, there are several notable algorithms reported in the literature, inspired by animal behaviour.Review [33] gives a review of other evolutionary algorithms inspired by animals for a specific problem optimisation.However, their review covers a range of maritime animal-inspired algorithms.Study [34] gives an overview of other animal behaviour inspired algorithms, not limited to maritime animals.Research proposed in [35] modifies the salp swarm algorithm optimisation [36,37], based on the salps sailing and foraging behaviour.Work presented introduces a dolphin inspired swarm optimisation algorithm with mimics the dolphins hunting technique, using their available senses.Research presented in demonstrates an optimisation algorithm modification on five real-life engineering problems, etc.

In recent years, SI was frequently applied in different scientific fields with promising results,usually targeting NP-hard problems from the fields of computer science and information technology.Applications include global numerical optimisation problem wireless sensor network efficiency [38],localisation and prolonging the overall lifetime [39,40] and cloud task scheduling [41,42].

Finally, swarm intelligence has also been successfully combined with machine learning models and in this way hybrid methods were devised.Some example of such approaches include: training and feature selection of artificial neural network (ANN) [43-49], assisting prediction of COVID-19 cases[50,51], hyper-parameters tuning [52-54] etc.

3 Proposed Methodology

The methodology of this research can be expressed in seven steps, of which the first step was performed manually when all input datasets were acquired from different sources, and the second-tolast step was also automated for the ease of multiple re-runs when needed, like others those before it,but its execution is done manually upon successful completion of the following steps:

■Download and normalise the datasets;

■Randomly split the datasets into 10 sets of training and testing data;

■Execute the optimisation algorithm for the given number of iterations;

■Randomly split the originally downloaded datasets into 10 new validation sets;

■Perform predictions with the trained model using the validation sets;

■Evaluate the prediction results;

■Compare results of this study to results from previous research.

3.1 Overview of Basic Whale Optimisation Algorithm

Whales are highly intelligent animals that have twice the number of brain cells similar to human spindle cells, which are responsible for emotions, decision making and social behaviour.Whales are often found in social groups, of families.Most whales prey on fish.Fish are much faster than whales.Because of this, some whales have developed a hunting method that allows them to hunt schools of fish.This method is often called the bubble-net feedingmethod.Researchers have found two variations of this method, which whales use.A part of this method is an approach where whales encircle their prey in a spiral manoeuvre.They dive down below the school of fish and start circling in a spiral path,slowly working their way to the surface, all the while creating a wall of bubbles above their traversed path.This wall of bubbles directs fish in an ever-smaller area as the circular path becomes smaller and smaller.Finally, the whale can capture large parts of the school as it reaches the surface.Whales do not know where the school of fish is located, ahead of time.This method helps them locate the fish after the manoeuvre.This feeding method is a direct inspiration of the WOA, which is mathematically modelled and used to perform optimisation [55].This method can be used to search for the optimal location within the search space, and can be presented by these equations:

In Eqs.(1)-(5),⊗is an element-wise multiplication operator,tis the current iteration,andare current position vectors of the best-known solution and the position vector,is a random vector,andare coefficient vectors,ais a control parameter, andtmaxis the maximum number of iterations.This model can be applied to an n-dimensional search space and it has two mechanisms, with equal probability of utilisation.The first is the shrinking encircling manoeuvre and the second is the spiral updating position manoeuvre.The former is achieved by linearly decreasing the control parameterafrom2 to 0 over the courseof iterations,while the latter is achieved bycalculating the distance between the whale and its prey, according to the equation for the distance.

By assigning random values toin ranges [-∞, -1] and [1,+∞] an optimal solution search can be performed.In this search behaviour, a random whale is selected, instead of the one with the best search performance.This allows the WOA to search for a global optimum.This can be expressed as:

The pseudo-code of the WOA can be found in a paper whereas a modified WOA with simultaneous hyper-parameter tuning is implemented.

The model representing this experimental setup is further explained in the model section.The proposed model balances the searching and utilising features of the algorithm.These results can be further improved by using multi-swarm mechanisms, as reported in the original paper.In addition to this modification, the paper suggests modifying the population generation process.Instead of using a random feature selection for each agent in a population, it suggests generating the population-based on chaos theory.In this approach, a chaos progression, sensitive to initial conditions, is used to modify the best agents from the previous population when generating the next population.

This can be achieved by introducing a self-adapting chaotic disturbance mechanism, expressed as:

In Eq.(10),Chaosrepresents the chaos value, andRis used to ensure a better balance between the algorithm’s searching and utilising features.Ris decreased from 1 to 0 in each of the iterations.This population generation method was proven to give better performances [56-59].

3.2 Drawbacks of the Original WOA and Proposed Improved Method

The basic WOA implementation has impressive performances, and it is considered to be a very powerful optimiser.However, the extensive experimentations with different benchmark functions sets have revealed drawbacks that can hinder the performances of the WOA under some circumstances.Namely, the basic WOA suffers from scarce exploration and inappropriate intensification-diversification trade-off, which can limit the algorithm’s power to explore different promising areas of the search domain and lead to getting stuck in the local optima.Consequently, in some runs, the basic WOA will not converge to the optimal areas of the search domain, obtaining poor mean values.To address these deficiencies, the proposed enhancedv WOA metaheuristics incorporate three mechanisms: chaotic population initialisation, OBL, and FS [60].

The first proposed enhancement is the chaotic population initialisation, which was first proposed in [61].The majority of the metaheuristics depend on random generators, but recent publications suggest that the search can be more intense if chaotic sequences are utilised Numerous chaotic maps exist today, including circle, Chebushev, iterative, sine, tent, logistic, singer and many more.After thorough experimentations with all mentioned maps, the most promising results were obtained with a logistic map, which has been consequently chosen for implementation in WOA.

The second mechanism that was incorporated in the proposed modified WOA is OBL, first introduced in [62].The OBL procedure gained popularity fast as it can significantly improve both intensification and diversification processes.TheOBL procedure is applied in two stages, first when the initial population is spawned and when a new population is spawned after completing the evaluation and finding the best agentX*of the finished iteration.Since the populationPis a set containing a number of agentsX, it can be said thatP= (X1,X2, ...Xn), wherenis the number of all agents in a single population.As each agent represents an attempted solution (s1,s2, ...sN), ofNsolutions, for a set of parameters (p1,p2, ...pM), optimised forMdimensions, at every given iterationt, a population contains agents with features for optimisation of a combination of parameters for a given solution.Creating agents can be expressed as:

In Eq.(11),lpandupare lower and upper bounds of the parameterpdimensions,while φ represents a pseudo-random value.To ensure chaotic initialisation, using the OBL procedure,for every generated agent, another one is generated for the oppositepdimension of the originalXs.This guarantees a chaotic component and this process can be described with the following expression:

In Eq.(13), the new population becomes the original populationP, afterwards joined with the opposing populationPo.This union represents the new population of agents for the next iteration.

To incorporating the FS mechanism in the proposed modified WOA,(t+ 1) is updated for a given solutions, denoted as(t+ 1), according to the following expression:

In Eq.(14),κ is a generalised Gaussian distribution, α is a control parameter of the FSmechanism,whose dynamic change guarantees improved performance.It is gradually decreased from the initial value, down to its minimum, over the course of the iterations.Parameters β0and γ are control parameters of randomisation.is the current candidate solution agent,is another candidate with a modification of one or more parameters,rs,pis the distance between those two candidate agents,which is calculated as:

In Eq.(15), the distance is calculated for each dimension of the given solutionsand parameterpon all agents of the population, where |P| is the number of agents in the population.

Inspired by introduced enhancements, proposed method is named chaotic oppositional WOA with firefly search (COWOAFS-SVM).

The proposed modified WOA can be expressed with the pseudo-code depicted in Algorithm 1.

Algorithm 1: The COWOAFS-SVM pseudo-code Initialize required control parameters α,β0and γ for the FA.P←Initialisation of the starting population of agents.Po←Chaotic initialisation of the opposing population, based on the original population P.P←∪Po, join both populations into a single and evaluate each agent.X*←Obtain the best agent using the fitness function.while j<tmaxdo α←According to Eq.(5).for each agent do X(t + 1)←According to Eqs.(7) or (9), dependent on A (per Eq.(3)).end for Evaluate X(t + 1)and update X*, using the fitness function.P←Spawn and mutate population, based on X*, according to Eqs.(10)-(13).end while

Experimental result cross-validation analysis is performed.These results are averaged to attain the result for each of the datasets and aggregated by each of the features of these datasets.Results are presented in the experimental results section and further explained in the discussion section of the paper.

4 Experimental Setup, Comparative Analysis and Discussion

For this research, the original algorithm and the algorithm created using the meta-heuristic approach with feature selection and simultaneous hyper-parameter tuning driven by the modified swarm intelligence algorithm are evaluated.This approach was proven to give favourable results in earlier research.Finally, both these algorithms were tested using 30 standard unconstrained benchmarking functions.To test the modified algorithm, first a set of standard benchmark functions retrieved from the well-known Congress on Evolutionary Computation (CEC) benchmark function suite,and afterwards, the proposed method vas validated for SVM hyper-parameter tuning and feature selection.In both experiments, proposed improved WOA was tested with the parameters as suggested in the original paper.

4.1 Standard CEC Function Simulations

The proposed method is tested on a set of standard bound-constrained benchmarks from the CEC benchmark suite [63].Statistical results of these benchmarks are presented in the results section.Functions with a single global optimum (F1-F7) are used to evaluate the capability of exploitation.Functions with many local optimums (F8-F23) are used to evaluate the capability of exploration.Functions (F24-F29) are used to evaluate the algorithms ability to escape local minima, as they are a set of very challenging functions.The proposed algorithm was run 30 times for each benchmark function, every time with a different population.Algorithms included in comparative analysis were implemented for the purpose of this study and tested with default parameters.Opponent algorithms include: Harris hawks’optimisation (HHO) [64], improved HHO (IHHO) [65], differential evolution(DE) [66], grasshopper optimisation algorithm (GOA) [67], grey wolf optimizer (GWO) [68], moth mothflameoptimisation (MFO) [69],multi-verse optimizer (MVO) [70], particle swarm optimisation(MVO)[70],particle swarm optimisation(PSO),whale optimisation algorithm(WOA), sine cosine algorithm (SCA) [71] and firefly algorithm (FA).

This research utilizes the same approach presented.The research presented in reported the results with N=30 and T=500.Since the proposed COWOAFS-SSVM algorithm uses more FFE in every run, the max FFE was utilised as the finish criteria.All other methods use one FFE per solution during the initialisation and update phase, and to provide fair analysis, max FFE was set to 15,030 (N+N·T).The values of the control parameters for the other methods can be obtained from.

The reported results from the Tab.1 show that the proposed COWOAFS-SVM metaheuristics obtained the best results on 30 benchmark functions, and was tied for the first place on seven more occasions.By conducting a comparison of the results of COWOAFS-SVM against the basicWOA and FA, it is clear that hybrid algorithm outperformed basic versions on every single benchmark instance.

Table 1: CEC 2017 comparative analysis results

Table 1: Continued

Table 1: Continued

On some functions, the proposed COWOAFS-SVM drastically improved the results of the basic variants; such is the example of F1, where it obtained 1000 times better results than the basic WOA and FA.Overall results indicate that the proposed COWOAFS-SVM outperformed significantly all approaches covered by the research.

Dispersion of solutions over 50 runs for some algorithms and chosen functions are shown in Fig.1.

Figure 1: Dispersion of solutions of benchmark functions F4, F5, F7, F9, F10and F12

To statistically validate the significance of the proposed method, the Friedman test and the two variance ranking analysis were performed.Tabs.2 and 3 depict the results obtained by twelve observed algorithms on the set of 30 CEC2017 benchmark functions for Friedman test ranks and aligned Friedman ranks, respectively.It is clear from both tables that the proposed COWOAFS-SVM significantly outperformed all other algorithms.

Table 2: Friedman test ranks for the compared algorithms over 30 CEC2017 functions

Table 2: Continued

Table 3: Aligned Friedman test ranks for the compared algorithms over 30 CEC2017 functions

Table 3: Continued

Finally, the Iman and Davenport test has also been conducted, as it can provide better results thanx2.The overview of the Iman and Davenport results is shown in Tab.4.The obtained result of 38.39 was compared against the F-distribution critical value (F(9.9e10)=1.82) and indicate that the result is significantly higher.

Table 4: Result summary of the Friedman and Iman-Davenport statistical test (α= 0.05)

Additionally, Friedman statistics (184.40) is larger than thex2critical value with ten degrees of freedom (1.82) at the level of significance 0.05.As the result, null hypothesis can be rejected, indicating that the proposed algorithm performs significantly better than other methods in the comparative analysis.

As the null hypothesis has been rejected by both statistical tests that were executed, the nonparametric Holm step-down procedure was also performed, and results are shown in Tab.5.

Table 5: Holm step-down procedure results

Table 5: Continued

4.2 Hyper Parameters Tuning and Feature Selection Experiments

This section first describes the SVM parameters that have been subjected to the optimisation process, followed by the short description of the datasets, pre-processing and metrics utilised to measure the performances.Lastly, in this section, results of the conducted experiments are presented,with the comparative analysis against other current methods, tested on the same problem instances.

4.2.1 Experimental Setup

For the purpose of this study, all methods included in comparative analysis are implemented and tested with control parameters, as suggested in the original paper [30].All methods were tested over the course of 150 iterations and the results are average of 10 independent runs, as per the restraints of the experimental setup, suggested in [30].The experiment was conducted in following the specifications laid out in the proposed methodology, starting with the acquisition, normalisation and preparation of all input datasets for the experiment, following a series of steps, as explained earlier, securing separate training and testing sets, using the former in the proposed algorithm and the later to evaluate the prediction results.

This section proposes the application of the devised COWOAFS-SVM algorithm to improve the efficacy of the SVM model (with and without the feature selection) on three standard benchmark medical diagnostic datasets, as described in the referred paper [5].The datasets included in the research deal with common illness, namely breast cancer, diabetes and erythemato-squamous (ES) disease.

Similarly as in [5], only two SVM hyper-parameters were considered: penalty parameter (C) and kernel bandwidth of kernel function (γ).Each agent is encoded as an array of two elements (C and γ).The search boundaries for C parameter are in range {2-7.5, 2-3.5,..., 27.5}, and for γ in range {2-7.5,2-3.5,..., 27.5}.

This article presents the experimental results of the modified WOA on several datasets, inherited from the original study.Datasets used for this experiment cover diseases affecting many people, such as breast cancer, diabetes and erythemato-squamous.Breakthroughs in diagnostic and detection methods for these diseases are unquestionably beneficial.All features commonly extracted from input sources, used in the diagnostic of these diseases are available in digitised forms from different databases:

■The breast cancer dataset was acquired from the Wisconsin breast cancer database.

■The database dataset was acquired from Pima Indian heritage and contains anonymized data about their female patients.

■The dermatology database has clinical and histopathological data, acquired, normalised, and prepared in the same manner as explained in [30].

The original study gives full descriptions and details of all three datasets.

Values from these datasets are normalised to [-1, 1] to avoid calculation issues and dominance and drag that much lower or much greater values may have on results.This normalisation strategy is performed on the dataset at the very beginning.Each dataset was split into ten segments to ensure more reliable and stable experimental results.Each run was repeated ten times to improvethese results.

The final results are averaged.To assure comparability of the results of this study with the results reported in the earlier study, this setup was limited to the same parameters of the population, iteration count, repetitions and segmentation of the datasets.This way, the hardware impact is eliminated and only the final performance of the optimisation of the target function is measured.

The algorithm performance evaluation was inherited from the original research [30], to ensure result comparability.These metrics are:

In Eqs.(16)-(18), formulas are given for the classification accuracy, and measurements are the sensitivity and the specificity, all referring to the tested algorithm.ValuesTP,TNandFNarenumbers of true positives, true negatives and false negatives, respectively.All values are expressed in percentage.

4.2.2 Experimental Results

The results of this experimental research are shown in this section.The performances of the proposed COWOAFS-SVM method were compared to the results published in [30].To provide valid grounds for the comparisons, this research utilised the same experimental setup and datasets as in[30].The results for MWOA-SVM, WOA-SVM, BFO-SVM, PSO-SVM, GA-SVM and pure SVM were also taken from [30].

Results Pertaining to Breast Cancer Diagnostics

The breast cancer dataset is comprised of 699 entries, 458 of which were benign, while 241 were malignant.The performances of the proposed COWOAFS-enhanced SVM method without feature selection are shown in Tab.6 and with feature selection in Tab.7.The results show average values obtained through 10 independent runs of ten-fold CV.The results indicate that the SVM structure generated by the COWOAFS-SVM algorithm outperforms all other SVM approaches, in terms of accuracy, area under curve, sensitivity and specificity.COWOAFS-SVM achieved average accuracy of 96.84%, together with the standard deviation that is smaller than the results of the MWOA-SVM and other compared methods, indicating that the COWOAFS-SVM is capable of producing consistent results.

Table 6: Breast cancer-results without feature selection

Table 7: Breast cancer-results with feature selection

Similarly as in [30], the performances of the proposed COWOAFS-SVM method were further enhanced by performing the optimisation of the SVM structure and feature selection together.The proposed method once again outperformed all competitors, by achieving the accuracy of 97.04%,the best values for area under curve, sensitivity and specificity, while having the smallest standard deviation.

Results Pertaining to Diabetes Diagnostics

The second dataset related to the diabetes diagnostics consisted of 768 entries, 268 belonging to the diabetic patients, while 500 were normal healthy instances.The results of the proposed approach without feature selection are shown in Tab.8.It can be clearly seen that the proposed COWOAFSSVM approach outperformed all other competitors, for which the results were taken from paper [30].The average metrics of ten runs show the accuracy of 77.59%, which is slightly better than MWOASVM, and significantly better than other compared methods.Other metrics are also in favour of the proposed COWOAFS-SVM method, including the smallest standard deviation values.

Table 8: Diabetes-results without feature selection

Tab.9 depicts the results obtained by the proposed COWOAFS-SVM approach with feature selection, from ten runs of ten-fold CV.Once again, the results clearly indicate that the proposed method outperformed other Metaheuristic based approaches in terms of accuracy, area under curve,sensitivity and specificity, while having the smallest values of the standard deviation.

Table 9: Diabetes-results with feature selection

Results Pertaining to ES Disease Diagnostics

The ES disease dataset consisted of 358 instances with 34 features and six classes of diseases.As in paper [30], the onlymetric that was utilised is the accuracy, because of themulti-classification nature of the challenge.Tabs.10 and 11 contain results attained using the proposed COWOAFS-SVM method compared to competitor approaches, averaged on 10 runs of 10 ten-fold CV, without and with feature selection, respectively.

In both cases, the proposed method outperformed other competitors both in terms of highest accuracy achieved and smallest standard deviation, indicating consistent performances of the approach.

Table 10: ES-results without feature selection

Table 11: ES-results with feature selection

4.3 Discussion

The presented research proposed a novel WOA method, enhanced with chaotic initialisation,opposition-based learning and firefly search, which was later utilised to improve the performances of the SVM for medical diagnostics.The experimental findings listed in the preceding section plainly show that the proposed COWOAFS-SVM enhanced SVM model obtained superior performances than other modern approaches subjected to the comparative analysis [5].The proposed COWOAFSSVM is capable of constructing a superior SVM with higher accuracy and consistency than other competitor methods.It is particularly interesting to compare the results of the proposed COWOAFSSVM against the CMWOAFS-SVM, which obtained the best results in [5].It is possible to conclude from the presented results that the proposed COWOAFS-SVM achieved significantly enhanced performances than CMWOAFS-SVM on all datasets used in the experiments.The improvement were observed both for the accuracy and other indicators (area under curve, sensitivity and specificity), as well as for the consistency of the results, that is reflected in the smaller average standard deviation.

To gain better insights into methods’performance, convergence speed graphs for all methods included in comparative analysis with respect to accuracy are shown in Fig.2.

Figure 2: Average convergence speed graphs for classification error rate metrics for 150 iterations

5 Conclusion

In this paper, an improved model for simultaneous feature selection and hyper-parameter tuning for SVMis proposed.The algorithm that was utilised for improving SVM performances was based on the WOA SI metaheuristics.The proposed method was then used to improve the SVM for medical diagnostics of breast cancer, diabetes, and erythemato-squamous illnesses.COWOAFS-SVM was evaluated and compared to other state-of-the-art metaheuristics based approaches, and achieved superior performances both in terms of higher accuracy and consistency reflected through smaller values of the standard deviation on all three observed datasets.The achieved promising results indicate the possible direction of the future research, which will focus on validating the proposed model on other datasets from different application domains.The other direction of the possible future work is to apply the proposed COWOAFS-SVM metaheuristics in solving other challenging problems from wide spectrum of application domains, including cloud computing, optimisation of the convolutional neural networks, and wireless sensor networks.

Funding Statement:The authors state that they have not received any funding for this research.

Conflicts of Interest:The authors state that they have no conflicts of interest regarding this research.