So, here we have some missing values so it want to go through a cleaning process.

So here we have missing data rows in a random order so we have to drop them by 'dropna' function and again check the data that is it clean or not

When we create a model then we may need the data should be normalized so here we are using MinMaxScaler for preprocessing

From above clean data sets we can observe that some have non-numeric values we shouldn't use non-numeric value with a learning algorithm so here we should convert it with some method. Here we will use 'One-Hot-Coding' scheme which will convert non-numeric values to "dummy" variable which can be used with learning algorithm.

In addition, We also need to convert the target to numeric value but we needn't to use 'One-Hot-Coding', we can use simple encoding to binary from as we have only variables as 'Yes' and 'No'.

Now we have all non-numerical values converted into numerical value and all numerical values has been normalized form. So we will split the data into training and test data. We will use 80% of data as training data and 20% as test data.

SVM :

AUC Score:

F-beta Score:

K-Nearest-Neighbors (KNN) :

AUC Score:

F-beta Score:

Choosing K is tricky, so I can't discard KNN until we've tried different values of K. Hence we write a for loop to run KNN with K values ranging from 10 to 50 and see if K makes a substantial difference.

Logistic Regression :

AUC Score:

F-beta Score:

Decision Trees :

AUC Score:

F-beta Score: