Acessibilidade / Reportar erro

A Hybrid Feature Ranking Algorithm for Assisted Reproductive Technology Outcome Prediction

Abstract

In recent years, the emerging technology of machine learning has made vast strides in medicine. Machine learning-based clinical decision support systems assist doctors make efficient diagnoses and offer better prescriptions. Today, one of the greatest challenges for doctors worldwide is the treatment of infertility, with even the most sophisticated technology offering limited success. Currently, the Assisted Reproductive Technology (ART) in use is highly sophisticated technology that offers a success rate of 20%, depending on a slew of factors with complex relationships. With their capacity to analyze large and complex datasets, the application of machine learning techniques to predictions can maximize the ART success rate. This research work attempts a dynamic model for ART outcome prediction using incremental classifiernamed Ensemble of Heterogeneous Incremental Classifier (EHIC) in Machine Learning. In this paper,a new feature ranking algorithm named Voted Information Gain Attribute Rank Estimation Algorithm (VIGAREA) is proposed to enhance the performance of EHIC. The proposed VIGAREA is a combination of a number of feature selection methods and information gain ratio of each variable. It has the capability to rank the features based on its significance. The methodology and the way how the proposed VIGAREA is developed is presented. Experimental results proved that the EHIC with the proposed VIGAREA achieves the highest prediction with the ROC area of 95.5% for the ART dataset used for the research. The effectiveness of the proposed VIGAREA is checked with a range of miscellaneous feature selection methods and found that the proposed feature ranking method VIGAREA performs optimally. Further, the performance of the proposed model is compared to that of existing models, and the findings show that the former outperforms the latter.

Keywords:
Assisted Reproductive Technology (ART); Ensemble Learner; Incremental Classifiers; Feature Ranking

HIGHLIGHTS

  • Proposed a hybrid feature ranking algorithm named VIGAREA

  • The proposed feature ranking algorithm gives importance to the individual feature as well as interaction between the features.

  • The proposed VIGAREA can combine any number and type of feature selection methods along with information gain ratio

HIGHLIGHTS

  • Proposed a hybrid feature ranking algorithm named VIGAREA

  • The proposed feature ranking algorithm gives importance to the individual feature as well as interaction between the features.

  • The proposed VIGAREA can combine any number and type of feature selection methods along with information gain ratio

INTRODUCTION

Machine learning (ML) is a combination of advanced statistics and Artificial Intelligence (AI), with the ability to extract previously unknown and potentially useful patterns from large databases, to be eventually converted into knowledge [11 Guh RS, Wu TC, Weng SP. Integrating genetic algorithm and decision tree learning for assistance in predicting in vitro fertilization outcomes. Expert Syst. Appl. 2011 Apr 1;38(4):4437-49.]. ML initially had takers in marketing and finance, and has since been applied in several fields, including healthcare. Presently, the status of human health worldwide calls for predictions to be made on one’s physical condition in order to prevent disease, or determine the effects of treatment on people. Currently, a serious healthcare issue that needs to be addressed globally is human infertility. Infertility is witnessed, even in young couples, on account of numerous causes. Occasionally, however, the healthcare systems can attribute no specific reason for infertility [22 Zarinara A, Zeraati H, Kamali K, Mohammad K, Shahnazari P, Akhondi MM. Models predicting success of infertility treatment: a systematic review. J. Reprod. Infertil. 2016 Apr;17(2):68.]. Infertile couples face problems that affect their social and marital lives [22 Zarinara A, Zeraati H, Kamali K, Mohammad K, Shahnazari P, Akhondi MM. Models predicting success of infertility treatment: a systematic review. J. Reprod. Infertil. 2016 Apr;17(2):68.]. Infertility is conclusively addressed by Assistant Reproductive Technology (ART), a complex and biotechnologically advanced technique that helps couples conceive when conventional medical and surgical corrective procedures have failed [33 Harpreet Arora. Medically Reviewed by Review Team of Medindia Dec 2018 [cited 2019 Feb] Available from: https://www.medindia.net/patientinfo/assisted-reproductive-technologies.htm#,
https://www.medindia.net/patientinfo/ass...
]. However, ART treatments are expensive, time-consuming, complex and painful, with a very low success rate, given the involvement of a large number of variables. The predictive analytics of machine learning might be highly useful in this respect. Because of the existence of multiple determinants called features, and the existence of non-linear correlations between the features, the ART process requires advanced predictive models that are able to learn from a massive quantum of data and update themselves by obtaining data with new combinations of features [44 Uyar A, Ciray HN, Bener A, Bahceci M. 3P: Personalized pregnancy prediction in IVF treatment process. InInternational Conference on Electronic Healthcare 2008 Sep 8 (pp. 58-65). Springer, Berlin, Heidelberg.]. Such predictive models assist doctors and patients to take corrective action, which improves the success rate. These factors call for a dynamic machine learning model for ART outcome prediction.

On the other hand, the predictive ability of a machine learning model depends on the features in the dataset and the algorithm used to build the model. The prediction ability is affected when the dataset has irrelevant features. Therefore, it is important to find influencing features in the dataset with a feature selection algorithm [44 Uyar A, Ciray HN, Bener A, Bahceci M. 3P: Personalized pregnancy prediction in IVF treatment process. InInternational Conference on Electronic Healthcare 2008 Sep 8 (pp. 58-65). Springer, Berlin, Heidelberg.]. There exist a number of feature selection algorithms that work based on filter and wrapper methods. Filter methods work independent of the learning algorithm and computes the importance of variables by taking into account the relationship between a feature and a target variable. Wrapper methods works by evaluating all the possible combinations of features and select a subset of feature that is optimized for a given classification algorithm. Filter methods only consider the importance of a specific feature and not the results of interactions between features, whereas wrapper methods concentrate on the performance of the learning algorithm and overlook the importance of a specific feature [55 Mwadulo MW, A review on feature selection methods for classification tasks, I. J. Comput. Appl. Technol. Res.,2016; 5(6): 395- 402.

6 Polikar R, Upda L, Upda SS, Honavar V. Learn++: An incremental learning algorithm for supervised neural networks. IEEE Trans. Syst. Man Cybern., part C (applications and reviews). 2001 Nov;31(4):497-508.
-77 Srivastava T, Basics of Ensemble Learning Explained in Simple English, 2015 Aug [cited 2020 Apr 19] Available from :https://www.analyticsvidhya.com/blog/2015/08/introduction-ensemble-learning/]. A hybrid feature selection algorithm that considers both the importance of individual features and the interactions between the features may support in getting a better dynamic model for ART outcome prediction which was built by combining two incremental machine learning classifiers, the Instance-Based Learner (IB1) [14 14 Aha DW, Kibler D, Albert MK. Instance-based learning algorithms. Machine learning. 1991 Jan;6(1):37-66.-1515 Cohen WW. Fast effective rule induction. In Machine learning proceedings 1995. 1995 Jan 1; Morgan Kaufmann pp. 115-123.] and the Averaged One Dependence Estimators (A1DE) updatable learner [16 16 Kevin_Jorda, Bayesian-machine-learning: Mid Term, Dec 2016, [cited 2019 Dec 20] Available from: https://quizlet.com/112857387/bayesian-machine-learning-flash-cards/
https://quizlet.com/112857387/bayesian-m...

17 Webb GI, Boughton JR, Wang Z. Not so naive Bayes: aggregating one-dependence estimators. Machine learning. 2005 Jan;58(1):5-24.

18 Wang L, Wang S, Li X, Chi B. Extracting credible dependencies for averaged one-dependence estimator analysis. Math. Prob. in Eng. 2014 Jun 17;2014.
- 1919 Picek S, Heuser A, Guilley S. Template Attack vs Bayes Classifier. IACR Cryptol. ePrint Arch. 2017;2017:531.], by means of the ensemble method and was published in [88 Ranjini K, Suruliandi A, Raja SP. An Ensemble of Heterogeneous Incremental Classifiers for Assisted Reproductive Technology Outcome Prediction. IEEE Trans. Comput. Soc. Syst. 2020 Nov 3;8(3):557-67.].

Hence, this research work propose a hybrid feature ranking algorithm named the Voted Information Gain Attribute Rank Estimation Algorithm (VIGAREA) that is capable of listing the influencing features by combining the results of different feature selection algorithm and information gain ratio of each feature. The performance of the proposed feature ranking algorithm is evaluated by comparing it with that of other feature selection methods.

Related work

Lu and coauthors [99 Lu H, Chen J, Yan K, Jin Q, Xue Y, Gao Z. A hybrid feature selection algorithm for gene expression data classification. Neurocomputing. 2017 Sep 20;256:56-62.] introduced a hybrid feature selection algorithm that combines the Mutual Information Maximization (MIM) and the Adaptive Genetic Algorithm (AGA) for gene expression data classification to eliminate the redundant samples and to reduce the dimension of the gene expression data.

Ahmad and Hayat [1010 Ahmad J, Hayat M. MFSC: Multi-voting based feature selection for classification of Golgi proteins by adopting the general form of Chou's PseAAC components. J. Theor. Biol. 2019 Feb 21;463:99-109.] developed a high throughput computational model to identify the subGolgi proteins. Since the available dataset is imbalanced, Synthetic Minority Oversampling Technique (SMOTE) is utilized to balance the dataset and in addition, a condense feature space is formed by fusing the high rank features of eleven different feature selection techniques. The high rank features are selected through majority voting algorithm; Ahmed and coauthors [1111 Ahmed N, Rafiq JI, Islam MR. Enhanced human activity recognition based on smartphone sensor data using hybrid feature selection model. Sensors. 2020 Jan;20(1):317.] developed a hybrid feature selection method, which includes a filter and wrapper method for Human Activity Recognition (HAR) based on smart phone sensor data. The process uses a Sequential Floating Forward Search (SFFS) to extract desired features for better activity recognition. Shaban and coauthors [1212 Shaban WM, Rabie AH, Saleh AI, Abo-Elsoud MA. A new COVID-19 Patients Detection Strategy (CPDS) based on hybrid feature selection and enhanced KNN classifier. Knowledge-Based Systems. 2020 Oct 12;205:106270.] introduced a new COVID-19 diagnose strategy called COVID-19 Patients Detection Strategy (CPDS) in which, a new Hybrid Feature Selection Methodology (HFSM) is developed, that elects the most informative features from the chest Computed Tomography (CT) images of COVID-19 patients and non COVID-19 peoples. Ranjiniand coauthors [1313 Ranjini K, Suruliandi A, Raja SP. Machine learning techniques for assisted reproductive technology: A review. J. Circuits, Syst. Comput. 2020 Sep 15;29(11):2030010.], the authors of the current paper, examined the application of machine learning techniques to ART and discussed, qualitatively and quantitatively, the classifiers used for ART.

Motivation and Justification of the work

Assisted Reproductive Technology (ART) is a medical procedure that addresses infertility [55 Mwadulo MW, A review on feature selection methods for classification tasks, I. J. Comput. Appl. Technol. Res.,2016; 5(6): 395- 402.], and includes all treatments that handle human sperms, occytes or embryos in-vitro to establish pregnancy. In-Vitro Fertilization (IVF) and Intra-Cytoplasmic Sperm Injection (ICSI) are commonly used ART methods. In these procedures, several eggs are collected from a woman’s ovary and fertilized with sperm to produce embryos. Of the developed embryos, one (or more) of the best-formed is transferred to the woman’s uterus. Each stage of the procedure is critical. Also, the number and quality of the embryos transferred greatly impacts the success rate. The greatest challenge in this stage is embryo formation, which depends on factors like the number of oocytes collected and the immunological characteristic of sperm, as well as its morphology and motility. Predicting the probability of ART success prior to the procedure can help perfect features that affect fertility. This is done by identifying how significant the features are, thus increasing the probability of success.

Studies on ART outcome prediction used datasets with limited data from regional fertility centers, as well as static data. However, algorithms need to be trained using an adequate volume of data to produce a model with the ability to predict the correct influencing attribute on the outcome. The classifiers used in most studies worked with static data, though ART data is likely to be dynamic, with the influencing variables changing with respect to environmental characteristics. Consequently, a dynamic predictive model that describes all the influencing attributes is lacking. Furthermore, from the point of view of the couple/s in question, finding influencing attributes reduces the cost, pain and time taken for a number of ART-related tests needed for treatment Predicting the probability of success prior to commencing treatment strengthens, at a psychological level, the couple/s involved [2323 Ramasamy N. Feature Reduction by Improvised Hybrid Algorithm for Predicting the IVF Success Rate. Int. J. Adv. Res. Comput. Sci. 2017 Jan 1;8(1).]. The problems listed above have motivated the building of a dynamic predictive model for ART outcome prediction. Dynamic models can be built using incremental classifiers in machine learning. The literature review makes it plain that using a combination of classifiers, rather than a single one, offers accurate predictions. So it is justified that this research work will use Ensemble of Heterogeneous Incremental Classifier (EHIC) which was implemented in [88 Ranjini K, Suruliandi A, Raja SP. An Ensemble of Heterogeneous Incremental Classifiers for Assisted Reproductive Technology Outcome Prediction. IEEE Trans. Comput. Soc. Syst. 2020 Nov 3;8(3):557-67.] and propose a new feature ranking algorithm based on EHIC which has the capability to list the influencing feature which in turn may improve the performance of EHIC.

Outline of the work done

Figure 1
Outline of the work done

The dynamic outcome prediction model for ART is built by using the new classifier EHIC and the proposed new feature selection algorithm. The fertility dataset used for the experiment is balanced by SMOTE. The outline of the proposed work is shown in Figure 1.The rest of the paper is organized as follows. Section 2 presents the methodology of the proposed work. Section 3 presents the experimental results and discussions. Section 4 concludes the paper.

MATERIAL AND METHODS

The methodology of the proposed dynamic outcome prediction model for ART commences with pre-processing. Next, the proposed feature ranking algorithm finds the influencing feature. Finally, the outcome prediction model for ART is built, using the EHIC.

Pre-processing

In order to build the ART outcome prediction model, this research uses the ART (fertility) dataset maintained by the Human Fertilization and Embryo Authority (HFEA). The dataset comprises a more records for negative results and a minimal number for positive results, rendering it imbalanced. An imbalanced dataset impacts the performance of the classifiers. Experiments conducted [2424 Suruliandi A, Ranjini K, Raja SP. Balancing assisted reproductive technology dataset for improving the efficiency of incremental classifiers and feature selection techniques.J. Circuits, Syst. Comput. 2021 May 24;30(06):2130007.] and corroborated by findings from the literature show that the SMOTE is best for balancing data [20 20 Branco P, Torgo L, Ribeiro RP. A survey of predictive modeling on imbalanced domains. ACM Computing Surveys (CSUR). 2016 Aug 13;49(2):1-50.- 2121 Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. J Artif Intell Res. 2002 Jun 1;16:321-57.][2424 Suruliandi A, Ranjini K, Raja SP. Balancing assisted reproductive technology dataset for improving the efficiency of incremental classifiers and feature selection techniques.J. Circuits, Syst. Comput. 2021 May 24;30(06):2130007.]. As a result, the ART dataset is balanced by the SMOTE.

Feature Selection: Proposed Voted Information Gain Attribute Rank Estimation Algorithm (VIGAREA)

The objective of this work is to propose a hybrid feature ranking algorithm with the ability to extract the influencing attribute by assigning weights to features (attributes). The novelty of the proposed algorithm lies in the calculation of weights for the features. In the real world dataset, especially in medical dataset, most of the features will be interrelated and adopting the existing filter type feature ranking methods which mostly depend on Univariate measure for ranking become inappropriate. Wrapper based feature selection methods may also ignore some important features since the selection of influential features depend on the performance of classifiers. Hence, instead of depending on a single feature selection method, the proposed work tries to take advantage of several feature selection methods. This research uses a number of feature selection methods, with the top vote going to features selected by the maximum number of feature selection methods so the best is selected. The votes for a particular feature are calculated by counting the number of feature selection methods choosing the said feature. To preserve the importance of individual features, the Information Gain Ratio (IGR) of each feature is calculated simultaneously. The weights for each featuresare calculated by multiplying the information gain ratio of each feature with the votes calculated for each. The weights for each feature are converted into normalized weights and checked with a certain threshold. Finally, features with a weight greater than or equal to the threshold value are selected as influencing features.

Figure 2
Block Diagram for the Proposed Voted Information Gain Attribute Rank Estimation Algorithm.

The schematic representation of the proposed VIGAREA is shown in Figure. 2, where a1an represents the list of attributes (features) which is given as input to the various feature selection algorithm and ai represents a single attribute and the value of i varies from 1ton. FS1FSn represents the list of feature selection algorithms that are selected for vote calculation. Voteforai represents the vote of each feature which is calculated by counting the number of feature selection methods that choose the particular feature. IGRP represents the procedure for calculating IGR for each feature. IGR for each feature is represented by IGRforai. ∏ calculates the weight Wai for each feature by multiplying vote and IGR of each feature. The calculated weight Waifor each feature is converted into normalized weight NWai by the normalization function N. The threshold limit thfor normalized weight will be set by passing the set of features satisfying varying criteria to the classifier. The attributes ai having the normalized weight NWai above and equal to threshold th will be added to the Influencing Feature Subset{IF}. Finally, the attributes will be ranked based on normalized weight.

List of Feature Selection Methods Used

This research uses various feature selection methods both from Filter Methods and Wrapper Methods. For filter approach, Correlation Feature Selection (CFS)filter methodis used in combination with the three search methods namely Best First Search (BFS), Greedy and Particle Swarm Optimization (PSO). For wrapper method four search methods namely Greedy, BFS, Incremental Wrapper Subset Selection (IWSS) and IWSS with embedded NB is used [25 25 Hall MA, Correlation-based feature subset selection for machine learning. Ph.D Thesis (University of Waikato, 1998).

26 Weka Attribute Selection -Class Best First, [cited 2019 Nov 20] Available from: http://weka. source forge.io/doc.dev/weka/attributeSelection/BestFirst.html,
http://weka. source forge.io/doc.dev/wek...

27 Weka Attribute Selection-Class Greedy Stepwise, [cited 2019 Nov 20]. Available from: http://weka.sourceforge.io/doc.dev/weka/attributeSelection/GreedyStepwise.html,
http://weka.sourceforge.io/doc.dev/weka/...

28 Xin-She Yang, Particle swarm optimization; 2014 [cited 2019 Dec 23]. Available from :https://www.sciencedirect.com/topics/engineering/particle-swarm-optimization.
https://www.sciencedirect.com/topics/eng...

29 Moraglio A, Di Chio C, Poli R. Geometric particle swarm optimisation. In European conference on genetic programming 2007 Apr 11. Springer, Berlin, Heidelberg. pp. 125-136.

30 Bermejo P, Gamez JA, Puerta JM. Improving incremental wrapper-based subset selection via replacement and early stopping.Int. J Pattern Recognit Artif Intell. 2011 Aug;25(05):605-25.
- 3131 Bermejo P, Gámez JA, Puerta JM. Speeding up incremental wrapper feature subset selection with Naive Bayes classifier. Knowledge-Based Systems. 2014 Jan 1;55:140-7.]. The classifiers used for wrapper method are EHIC, A1DE Updatable and IB1. Totally fifteen combinations of methods are used. Hence in this research the total number votes calculated for each attribute will be 15.

Information Gain Ratio (IGR)

Information Gain (IG) represents the average amount of information about the class value contained in a feature or attribute value [3232 Uyar A, "Assessing And Enhancing Machine Learning Methods in IVF Process: Predictive Modeling Of Implantation And Blastocyst Development", Ph.D Thesis(Bo˘gazi¸ci University, 2011).]. IG gives mutual information between the class and attribute. It is a symmetrical measure of dependency between two attributes and selects candidate features with more information which is identified with the help of entropy[3333 Hall MA, Smith LA. Feature selection for machine learning: comparing a correlation-based filter approach to the wrapper. InFLAIRS conference 1999 May 1 (Vol. 1999, pp. 235-239).].The IG values of features provide reasonable knowledge to reduce the search space for feature subset selection but it favours features with more values even though they are not informative. IG is defined as

I G ( A t t r i b u t e ) = I G ( C l a s s , A t t r i b u t e ) = H ( C l a s s ) H ( C l a s s | A t t r i b u t e )

where, H(Class) is called as the Entropy of the Class and is calculates as

H ( C l a s s ) = i = 1 N o . o f C l a s s P ( C l a s s i ) . log 2 P ( C l a s s i )

is the Shannon’s entropy and

H ( C l a s s | A t t r i b u t e ) = j P ( A t t r i b u t e s j ) i P ( C l a s s i | A t t r i b u t e j ) log 2 P ( C l a s s i | A t t r i b u t e j )

IGRevaluates the worth of an attribute by measuring the gain ratio with respect to the class. It is a non-symmetrical measure that is introduced to balance the bias of the IG (5). IGR is given as

I G R ( C l a s s , A t t r i b u t e ) = ( H ( C l a s s ) H ( C l a s s | A t t r i b u t e ) ) H ( A t t r i b u t e )

where,

H ( A t t r i b u t e ) = j = 1 P ( A t t r i b u t e j ) . log 2 P ( A t t r i b u t e j )

wherej is the number of different values of the attributes.

In IGR, when a class has to be predicted, it normalizes the IG by dividing by the entropy of that attribute, and vice versa and hence, the IGR values always fall in the range [0, 1]. A value of IGR = 1 indicates that the information of the particular attribute completely predicts the class, and A value of IGR = 0 indicates that there is no relation between that attribute and the class.

Algorithm: Proposed Voted Information Gain Attribute Rank Estimation Algorithm

Input:Attrinutesset{a1,a2,.an}TrainingSetD={xi,yi}i=1m:

L i s t o f F S a lg o r i t h m s F S = { f 1 , f 2 , , f n }

O u t p u t : R a n k e d I n f l u e n c e d A t t r i b u t e s e t I F

1. S e l e c t t h e l i s t o f F e a t u r e S e l e c t i o n A lg o r i t h m F S = { f 1 , f 2 , f 3 . f n }

2. Foreachattributea1,a2,.ando

a . F o r e a c h F S d o

i f a i i s s e l e c t e d b y f i t h e n

v o t e a i = v o t e a i + 1

E n d F o r

b . C l a c u l a t e I n f o r m a t i o n G a i n R a t i o f o r e a c h a t t r i b u t e

I G R i = I n f o r m a t i o n G a i n R a t i o ( a i )

c . C a l c u l a t e W e i g h t f o r e a c h a t t r i b u t e

w e i g h t a i = I G R i × v o t e a i

E n d F o r

3. C a l c u l a t e N o r m a l i z e d W e i g h t f o r e a c h a t t r i b u t e

N o r m a l i z e d W e i g h t a i = w e i g h t a i w e i g h t a i

4. O r d e r A t t r i b u t e s a i b a s e d o n N o r m a l i z e d W e i g h tan d a s s i g n R a n k

5. F i n d t h e t h r e s h o l d lim i t b y P a s sin g a t t r i b u t e s t o t h e s e l e c t e d c l a s s i f i e r

a . E v a l u a t e t h e p e r f o r m a n c e o f t h e c l a s s i f i e r .

b . W h e n t h e p e r f o r m a n c e o f t h e c l a s s i f i e r s t a r t s t o deg r a d e , s e t t h e t h r e s h o l d v a l u e .

6. C h o o s e t h e a t t r i b u t e s h a v i n g t h e N o r m a l i z e d w e i g h t a b o v e a n d e q u a l t o t h r e s h o l d

asinfluencingattribute {IF}

Classifier: Ensemble of Incremental Classifier for ART Outcome Prediction

The ensemble of heterogeneous incremental classifier (EHIC) which was proposed by the authors in [88 Ranjini K, Suruliandi A, Raja SP. An Ensemble of Heterogeneous Incremental Classifiers for Assisted Reproductive Technology Outcome Prediction. IEEE Trans. Comput. Soc. Syst. 2020 Nov 3;8(3):557-67.] is used for classification. The performance of the EHIC may be enhanced by supplying influenced feature subset. The block diagram for the proposed model for ART outcome prediction is shown in Figure. 3.

Figure 3
Block Diagram of Dynamic Outcome PredictionModel for ART.

RESULTS

This section explains the working of the proposed feature ranking algorithm, VIGAREA, in finding influencing features, and undertakes an evaluation of its performance. To evaluate the performance of the classifiers, the ROC is taken as a high-priority metric throughout the experiment, since it produces unbiased reports even for imbalanced datasets. The HFEA maintains ART treatment data which is available at https://www.hfea.gov.uk. Total of 16383 records with 52 ART-related attributes are used for this research [4040 Uyar A, Bener A, Ciray HN. Predictive modeling of implantation outcome in an in vitro fertilization setting: an application of machine learning methods. Med. Decis. Making. 2015 Aug;35(6):714-25.].

Working of Proposed Feature Ranking Algorithm

The performance of the machine learning algorithm depends on the quality of the data operated on, and that of a model may be enhanced by providing an adequate quantum of relevant information. The HFEA dataset used contains an adequate number of instances and is to be checked to ascertain if its contents comprise relevant or extraneous and irrelevant information. This task is executed using a number of feature selection methods that are chosen and applied to the HFEA dataset to find the influencing features. The resultant influencing feature for each method is shown as tick mark (✓) in Table 1. The last column of the Table 1 shows the number of feature selection method selects a particular feature, which is represented as a vote for the attribute. Each tick in Table 1 indicates the selection of a particular attribute by a particular feature selection method.The number of selected features by each method is shown within the parenthesis in the header row in Table 1.

Table 1
Attribute Selected By Various Feature Selection Method

Once the vote is calculated, the next step is to calculate the weight for each feature which is done by multiplying the calculated vote of each feature (Table 1) with the information gain ratio of each feature.Next, the calculated weight is converted into a normalized weight, based on which ranks are assigned to the attributes. Table 2 shows how the proposed feature selection method calculates the weight, how the calculated weight is converted to a normalized weight, and how a rank is assigned to each attribute in descending order of the calculated normalized weight.

Table 2
Attribute Ranking by the Proposed Feature Ranking Method

The results from Table 2 show that the proposed feature selection method VIGAREA takes advantage of both the filter and wrapper methods. Given that the IGR calculation does not depend on the classifier used, the working principle in operation is the filter approach, whereas both the filter and wrapper methods are used for vote calculation. Thus, the proposed feature selection method works well as a hybrid.

Despite the efficacy of the new feature ranking algorithm proposed, the goal of the experiment is to find the influencing features of the ART dataset for EHIC. To this end, a new experiment is conducted by setting several criteria to find the threshold limit for the attribute. Attributes satisfying the specified criteria, based on the normalized weight, are passed through the EHIC and the performance of the EHIC is evaluated. The results of the experiments are shown in Table 3.

Table 3 shows that the EHIC performs well for criteria where the normalized weight for an attribute less than 0.01 is removed. Hence the threshold limit is set to 0.01, and attributes with a normalized weight above and equal to 0.01 are chosen as influencing features. According to the proposed feature selection method, attributes with a rank from 1 to 22 are selected as influencing features. It is also evident that the ROC value for the proposed EHIC has risen to 95.5.

Table 3
Finding the Threshold Value for Normalized Weight (NW) to Choose Influencing Features

Evaluation of the Proposed Model for each Classes of ART Dataset

The following experiment checks the applicability of the model for individual classes in the ART dataset. The class variable of the HFEA ART dataset is Live Birth Occurrence which is a binary class variable (1 - Live Birth; 0 - Not a Live Birth). The predictive model must predict all the classes equally; and not skew towards a particular class. To check the stability of predicting both classes equally, the evaluation of the dynamic model for each class isshown in Table 4.

Table 4
Evaluation of the Proposed Model for each Classes of ART Dataset

From Table 4, it is clear that the proposed model is not skewed towards a particular class, and so it is recommended that the EHIC based on the hybrid feature ranking algorithm be used for ART outcome prediction.

A Performance Evaluation of the VIGAREA with other Feature Selection Methods

Without relying on the existing feature selection method for the EHIC, this research proposes a new feature ranking method for choosing influential attributes. There is, however, a need to validate the performance of the proposed VIGAREA by comparing it with other feature selection algorithms, as well as those without feature selection, for both the base learners IB1 and A1DE Updateable and the EHIC, as listed in Table 5. The last row of Table 5 shows the number of Influencing Features (IF) selected by each method. It is apparent from Table 5 that the proposed feature ranking algorithm performs best with only 22 selected features. It is followed by the Incremental Wrapper Subset Selection (IWSS) method with good results.

Further, Table 5 shows that the EHIC and base learners produce the highest results for the proposed feature ranking method, proving that it is best suited to ART outcome prediction.

Table 5
Comparison of the Proposed Feature Ranking Method with Various Feature Selection Methods.

Performance Evaluation of the Proposed VIGAREA With Other Classifiers

With the motivation to check how the proposed VIGAREA is functioning with various classifiers, the performance of various classifiers before and after the application proposed VIGAREA is compared. For comparison most of the classifiers which are already used in the literature for ART related studies are chosen. The same ART dataset maintained by HFEA is taken for the experiment and the results are shown in Table 6.

Table 6
Performance of Various Classifiers Before and After the Application of Proposed VIGAREA.

The results of Table 6 revealed that most of the classifiers performance is improved after the application of proposed feature ranking algorithm for ROC metric. For some classifiers the performance is slightly reduced, but it is promising that the classifiers achieves this performance with the reduced feature of 22 whereas the original dataset contains 52 features.

When analyzing the results deeply the interesting facts that are identified is actually the threshold limit for the proposed hybrid feature ranking algorithm is set by evaluating its performance with the proposed EHIC. Hence the performance of the ensemble and the performance of the base learner IB1 and A1DE updatable are greatly improved for all the metrics. Whereas for the tree based classifiers the performance improved for ROC metric and reduced slightly for accuracy.

Even though the proposed feature ranking is showing promising performance, its performance can be improved by choosing appropriate feature selection methods for combination. Hence it is recommended that the performance of the classifier can be improved by selecting suitable feature selection methods for that particular classifier and combine those methods using VIGAREA and sets the threshold based on its evaluation.

Comparing the performance of the Proposed EHIC with and without the Proposed Feature Ranking Algorithm

The performance of the EHIC model, before and after applying the proposed feature ranking algorithm, is shown in Figure 4.

Figure 4
Comparison of Proposed EHIC with and without Proposed Feature Ranking Algorithm.

Figure 4 shows that the performance of the EHIC model has improved after applying the proposed feature ranking method. It underscores the need to find influencing features,and also demonstrates that supplying relevant features enhances the performance of the model. Thus, the experiment has justified the use of the proposed feature ranking algorithm.

Performance Evaluation of the Proposed Model with the Existing State-of-Art Models Used for ART Outcome Prediction

The performance of the proposed model is checked with the existing models available for ART related works. In order to ensure appropriate comparison, all the models are provided with the same ART dataset balanced by SMOTE. The results of the comparison are shown in Table 7.

Table 7
Performance Evaluation of the Proposed Model with the Existing State-of-Art Models Used for ART Outcome Prediction

The results of Table 7 make it clear that the performance of the proposed model out beats the other model with the ROC of 95.5 with the selected feature of 22. Next to the proposed model RF with chi-square ranking proposed byHafiz.Pand coauthors (2017) (39) achieves 93.7 ROC value with 26 influencing attributes. The performances of the other models are comparatively low. This experiment makes it plain that the proposed model outperforms the other model in all the metrics and ensures its applicability for ART outcome prediction.

Calculating the Computational Complexity of the Proposed Methods

The computational complexity of the EHIC is calculated based on the time and space complexity of the base learners. The training time taken for the proposed algorithms depends on the number of training example taken and the number of attributes. The time taken for classification depends on the number of classes and the number of attributes. The space complexity of the EHIC depends on the number of training examples, the number of distinct values for each attribute and the number of available classes. The time complexity of the proposed hybrid feature ranking algorithm VIGAREA, depends on the number of FS algorithm chosen for combination as well the running time needed for those methods. The results are shown in Table 8

Table 8
Computational Complexity of the Algorithms

CONCLUSION

In this research work, a novelhybrid feature ranking algorithm based on EHIC for ART outcome prediction is proposed. The feature ranking algorithm is proposed by combining the results of existing feature selection algorithms using voting, and taking into consideration the information gain ratio of individual features. The dynamic ensemble model with the proposed feature selection method achieves the highest prediction with an Area Under the ROC of 95.5 for the HFEA ART dataset. The efficiency of the proposed method is checked with various feature selection methods as well as with various classifiers and it is identified that the proposed feature ranking method performs good. The performance of the proposed model is checked with other models and found that the former is performing better. So it is concluded that the proposed model may be useful for the physicians and biologists worldwide, for ART outcome prediction.

The strength of the proposed hybrid feature ranking algorithm VIGAREA is it can combine any number and type of feature selection methods along with information gain ratio. It has the capability to work in a versatile manner and preserves the importance of a single feature as well as the interdependent performance. The limitation is that the performance of the proposed VIGARE algorithm depends on the feature selection methods that are selected for combination.

Further research can be carried out in the following directions. One of the important factors that influence the success of ART treatment is the embryos transfer. Separate research can be carried out to choose the best embryos and also the number of embryos that can be transferred in a single cycle. This research work takes physical and reproductive fitness factors into consideration. But environmental and social factors may have impact on infertility. Hence environmental and social factors that affect fertility may also be taken into account and can apply associative rule mining to find any common pattern in the data. Further research can be done to predict which fertility treatments is better suited to couples having specific characteristics and to understand how the treatment effect vary with respect to couples’ characteristics.

Acknowledgments

We thank Dept. of Computer Science and Engineering of MSU.

REFERENCES

  • 1
    Guh RS, Wu TC, Weng SP. Integrating genetic algorithm and decision tree learning for assistance in predicting in vitro fertilization outcomes. Expert Syst. Appl. 2011 Apr 1;38(4):4437-49.
  • 2
    Zarinara A, Zeraati H, Kamali K, Mohammad K, Shahnazari P, Akhondi MM. Models predicting success of infertility treatment: a systematic review. J. Reprod. Infertil. 2016 Apr;17(2):68.
  • 3
    Harpreet Arora. Medically Reviewed by Review Team of Medindia Dec 2018 [cited 2019 Feb] Available from: https://www.medindia.net/patientinfo/assisted-reproductive-technologies.htm#,
    » https://www.medindia.net/patientinfo/assisted-reproductive-technologies.htm
  • 4
    Uyar A, Ciray HN, Bener A, Bahceci M. 3P: Personalized pregnancy prediction in IVF treatment process. InInternational Conference on Electronic Healthcare 2008 Sep 8 (pp. 58-65). Springer, Berlin, Heidelberg.
  • 5
    Mwadulo MW, A review on feature selection methods for classification tasks, I. J. Comput. Appl. Technol. Res.,2016; 5(6): 395- 402.
  • 6
    Polikar R, Upda L, Upda SS, Honavar V. Learn++: An incremental learning algorithm for supervised neural networks. IEEE Trans. Syst. Man Cybern., part C (applications and reviews). 2001 Nov;31(4):497-508.
  • 7
    Srivastava T, Basics of Ensemble Learning Explained in Simple English, 2015 Aug [cited 2020 Apr 19] Available from :https://www.analyticsvidhya.com/blog/2015/08/introduction-ensemble-learning/
  • 8
    Ranjini K, Suruliandi A, Raja SP. An Ensemble of Heterogeneous Incremental Classifiers for Assisted Reproductive Technology Outcome Prediction. IEEE Trans. Comput. Soc. Syst. 2020 Nov 3;8(3):557-67.
  • 9
    Lu H, Chen J, Yan K, Jin Q, Xue Y, Gao Z. A hybrid feature selection algorithm for gene expression data classification. Neurocomputing. 2017 Sep 20;256:56-62.
  • 10
    Ahmad J, Hayat M. MFSC: Multi-voting based feature selection for classification of Golgi proteins by adopting the general form of Chou's PseAAC components. J. Theor. Biol. 2019 Feb 21;463:99-109.
  • 11
    Ahmed N, Rafiq JI, Islam MR. Enhanced human activity recognition based on smartphone sensor data using hybrid feature selection model. Sensors. 2020 Jan;20(1):317.
  • 12
    Shaban WM, Rabie AH, Saleh AI, Abo-Elsoud MA. A new COVID-19 Patients Detection Strategy (CPDS) based on hybrid feature selection and enhanced KNN classifier. Knowledge-Based Systems. 2020 Oct 12;205:106270.
  • 13
    Ranjini K, Suruliandi A, Raja SP. Machine learning techniques for assisted reproductive technology: A review. J. Circuits, Syst. Comput. 2020 Sep 15;29(11):2030010.
  • 14
    Aha DW, Kibler D, Albert MK. Instance-based learning algorithms. Machine learning. 1991 Jan;6(1):37-66.
  • 15
    Cohen WW. Fast effective rule induction. In Machine learning proceedings 1995. 1995 Jan 1; Morgan Kaufmann pp. 115-123.
  • 16
    Kevin_Jorda, Bayesian-machine-learning: Mid Term, Dec 2016, [cited 2019 Dec 20] Available from: https://quizlet.com/112857387/bayesian-machine-learning-flash-cards/
    » https://quizlet.com/112857387/bayesian-machine-learning-flash-cards
  • 17
    Webb GI, Boughton JR, Wang Z. Not so naive Bayes: aggregating one-dependence estimators. Machine learning. 2005 Jan;58(1):5-24.
  • 18
    Wang L, Wang S, Li X, Chi B. Extracting credible dependencies for averaged one-dependence estimator analysis. Math. Prob. in Eng. 2014 Jun 17;2014.
  • 19
    Picek S, Heuser A, Guilley S. Template Attack vs Bayes Classifier. IACR Cryptol. ePrint Arch. 2017;2017:531.
  • 20
    Branco P, Torgo L, Ribeiro RP. A survey of predictive modeling on imbalanced domains. ACM Computing Surveys (CSUR). 2016 Aug 13;49(2):1-50.
  • 21
    Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: synthetic minority over-sampling technique. J Artif Intell Res. 2002 Jun 1;16:321-57.
  • 22
    Bradley AP. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern recognition. 1997 Jul 1;30(7):1145-59.
  • 23
    Ramasamy N. Feature Reduction by Improvised Hybrid Algorithm for Predicting the IVF Success Rate. Int. J. Adv. Res. Comput. Sci. 2017 Jan 1;8(1).
  • 24
    Suruliandi A, Ranjini K, Raja SP. Balancing assisted reproductive technology dataset for improving the efficiency of incremental classifiers and feature selection techniques.J. Circuits, Syst. Comput. 2021 May 24;30(06):2130007.
  • 25
    Hall MA, Correlation-based feature subset selection for machine learning. Ph.D Thesis (University of Waikato, 1998).
  • 26
    Weka Attribute Selection -Class Best First, [cited 2019 Nov 20] Available from: http://weka. source forge.io/doc.dev/weka/attributeSelection/BestFirst.html,
    » http://weka. source forge.io/doc.dev/weka/attributeSelection/BestFirst.html
  • 27
    Weka Attribute Selection-Class Greedy Stepwise, [cited 2019 Nov 20]. Available from: http://weka.sourceforge.io/doc.dev/weka/attributeSelection/GreedyStepwise.html,
    » http://weka.sourceforge.io/doc.dev/weka/attributeSelection/GreedyStepwise.html
  • 28
    Xin-She Yang, Particle swarm optimization; 2014 [cited 2019 Dec 23]. Available from :https://www.sciencedirect.com/topics/engineering/particle-swarm-optimization
    » https://www.sciencedirect.com/topics/engineering/particle-swarm-optimization
  • 29
    Moraglio A, Di Chio C, Poli R. Geometric particle swarm optimisation. In European conference on genetic programming 2007 Apr 11. Springer, Berlin, Heidelberg. pp. 125-136.
  • 30
    Bermejo P, Gamez JA, Puerta JM. Improving incremental wrapper-based subset selection via replacement and early stopping.Int. J Pattern Recognit Artif Intell. 2011 Aug;25(05):605-25.
  • 31
    Bermejo P, Gámez JA, Puerta JM. Speeding up incremental wrapper feature subset selection with Naive Bayes classifier. Knowledge-Based Systems. 2014 Jan 1;55:140-7.
  • 32
    Uyar A, "Assessing And Enhancing Machine Learning Methods in IVF Process: Predictive Modeling Of Implantation And Blastocyst Development", Ph.D Thesis(Bo˘gazi¸ci University, 2011).
  • 33
    Hall MA, Smith LA. Feature selection for machine learning: comparing a correlation-based filter approach to the wrapper. InFLAIRS conference 1999 May 1 (Vol. 1999, pp. 235-239).
  • 34
    Human Fertilization and Embryo Authority (HFEA), Anonymised Register- our data, [Cited 2018 Sep 23] Available from: https://www.hfea.gov.uk
    » https://www.hfea.gov.uk
  • 35
    An introduction to Machine Learning, 2019 Feb [Cited 2019 Dec 3] Available from: https://www.geeksforgeeks.org /introduction-machine-learning/.
    » https://www.geeksforgeeks.org /introduction-machine-learning
  • 36
    Brownlee. J, 'What is confusion matrix in machine learning', Dec 2016 [Cited 2018 Aug 2] Available from: https://machinelearningmastery.com/confusion-matrix-machine-learning/,
    » https://machinelearningmastery.com/confusion-matrix-machine-learning
  • 37
    Ramasamy N. Feature Reduction by Improvised Hybrid Algorithm for Predicting the IVF Success Rate. International J. Adv. Res. Comput. Sci. 2017 Jan 1;8(1).
  • 38
    Vijayalakshmi N, Uma Maheshwari M. Data mining to elicit predominant factors causing infertility in women. Interantional J. Comput. Sci. Mob. Comput. 2016;5(8):5-9.
  • 39
    Hafiz P, Nematollahi M, Boostani R, Jahromi BN. Predicting implantation outcome of in vitro fertilization and intracytoplasmic sperm injection using data mining techniques. Int J Fertil Steril. 2017 Oct;11(3):184.
  • 40
    Uyar A, Bener A, Ciray HN. Predictive modeling of implantation outcome in an in vitro fertilization setting: an application of machine learning methods. Med. Decis. Making. 2015 Aug;35(6):714-25.
  • 41
    Davidson L, Boland MR. Enabling pregnant women and their physicians to make informed medication decisions using artificial intelligence. J Pharmacokinet Pharm. 2020 Apr 11:1-4.
  • Funding:

    This research received no external funding

Edited by

Editor-in-Chief:

Alexandre Rasi Aoki

Associate Editor:

Fabio Alessandro Guerra

Publication Dates

  • Publication in this collection
    27 June 2022
  • Date of issue
    2022

History

  • Received
    29 Sept 2021
  • Accepted
    05 Nov 2021
Instituto de Tecnologia do Paraná - Tecpar Rua Prof. Algacyr Munhoz Mader, 3775 - CIC, 81350-010 Curitiba PR Brazil, Tel.: +55 41 3316-3052/3054, Fax: +55 41 3346-2872 - Curitiba - PR - Brazil
E-mail: babt@tecpar.br