My machine learning model dataset is cleaveland data base with 300 rows and 14 attributes--predicting whether a person has heart disease or not.. But aim is create a classification model on logistic regression... I preprocessed the data and ran the model with x_train,Y_train,X_test,Y_test.. and received avg of 82 % accuracy...
So to improve the accuracy I did remove features that are highly correlated to each other [as they would give the same inforamtion]
And I did RFE[recursive feature elimination]
followed by PCA[principle component analysis] for dimensionality reduction...
Still I didnt find the dataset to be be better in accuracy..
Why is that?
Also why does my model shows different accuracy each time? is it beacuse of taking different x_train,Y_train,X_test,Y_test each time?
Should i change my model for better accuracy? Is 80 % average good or bad accuracy?