HEART DISEASE PREDICTION Using MACHINE LEARNING ALGORITHM Presentation
HEART DISEASE PREDICTION Using MACHINE LEARNING ALGORITHM Presentation
HEART DISEASE PREDICTION Using MACHINE LEARNING ALGORITHM Presentation
BACHELOR OF TECHNOLOGY
IN COMPUTER SCIENCE ENGINEERING
Submitted by
Neeraj Yadav – 2100100109011
Aryan Srivastava -- 2100100109005
Vishal Kumar Yadav – 2000100100192
Amir Hamza Khan -- 2100100109003
Hardware requirements:
❖Processer : Any Update Processer
❖Ram : Min 4GB
❖Hard Disk : Min 100GB
Software requirements:
❖Operating System : Windows family
❖Technology : Python3.7
❖IDE : Jupiter notebook
WORKING OF SYSTEM .
1. SYSTEM ARCHITECTURE
The system architecture gives an overview of the working
of the system. The working of this system is described as
follows: Dataset collection is collecting data which contains
patient details.
Attributes selection process selects the useful attributes
for the prediction of heart disease. After identifying the
available data resources, they are further selected,
cleaned, made into the desired form. Different
classification techniques as stated will be applied on
preprocessed data to predict the accuracy of heart
disease. Accuracy measure compares the accuracy of
different classifiers.
2. MACHINE LEARNING
In machine learning, classification refers to a predictive
modeling problem where a class label is predicted for a
given example of input data.
● Unsupervised learning
Unsupervised learning cannot be directly applied to a
regression or classification problem because, unlike
supervised learning, we have the input data but no
corresponding output data.
The goal of unsupervised learning is to find the underlying
structure of the dataset, group that data according to
similarities, and represent that dataset in a compressed
format.
3. ALGORITHMS
• SUPPORT VECTOR MACHINE (SVM):
Support Vector Machine or SVM is one of the most popular
Supervised Learning algorithms, which is used for
Classification as well as Regression problems. However,
primarily, it is used for Classification problems in Machine
Learning.
The goal of the SVM algorithm is to create the best line or
decision boundary that can segregate n-dimensional space
into classes so that we can easily put the new data point in
the correct category in the future. This best decision
boundary is called a hyperplane.SVM chooses the extreme
points/vectors that help in creating the hyperplane. These
extreme cases are called support vectors, and hence the
algorithm is termed as Support Vector Machine.
• ADABOOST ALGORITHM:
Adaboost was the first really successful boosting algorithm
developed for the purpose of binary classification.
Adaboost is short for Adaptive Boosting and is a very
popular boosting technique which combines multiple
“weak classifiers” into a single “strong classifier”
Algorithm:
1. Initially, Adaboost selects a training subset randomly.
2. It iteratively trains the Adaboost machine learning
model by selecting the training set based on the accurate
prediction of the last training.
3. It assigns the higher weight to wrong classified
observations so that in the next iteration these
observations will get the high probability for classification.
4. Also, it assigns the weight to the trained classifier in
each iteration according to the accuracy of the classifier.
The more accurate classifier will get high weight.
5. This process iterates until the complete training data fits
without any error or until reached to the specified
maximum number of estimators.
• XGBOOST ALGORITHM:
XG-boost is an implementation of Gradient Boosted
decision trees. It is a type of Software library that was
designed basically to improve speed and model
performance. In this algorithm, decision trees are created
in sequential form. Weights play an important role in XG-
boost. Weights are assigned to all the independent
variables which are then fed into the decision tree which
predicts results. Weight of variables predicted wrong by
the tree is increased and these the variables are then fed
to the second decision tree. These individual
classifiers/predictors then assemble to give a strong and
more precise model.
It can work on regression, classification, ranking, and user-
defined predict. Regularization: XG-boost has in-built L1
(Lasso Regression) and L2 (Ridge Regression) regularization
which prevents the model from overfitting. That is why,
XG-boost is also called regularized form of GBM (Gradient
Boosting Machine). While using Scikit Learn libarary, we
pass two hyper-parameters (alpha and lambda) to XG-
boost related to regularization. alpha is used for L1
regularization and lambda is used for L2 regularization.
References:-