Ism Research Assessment 3
Ism Research Assessment 3
MLA Citation:
Artificial Intelligence Algorithms For Beginners. (2020, May 14). Retrieved October 17,
Assessment:
which are supervised, unsupervised, reinforcement, and ensemble learning, and also
the types of problems solved using each type of algorithm. These types of problems are
classification. regression, and clustering. The article then gives examples of specific
algorithms, which type of algorithm they are, and what category the problems they
This information could be useful in the future because I can now understand the
differences between different types of artificial intelligence algorithms and I know what
each type of algorithm is used for. For example, I now know that classification types
problems such as sorting emails into Spam folders and non-spam folders and uses
supervised learning algorithms so that the algorithm can check to see if it is classifying
the emails correctly. However, I still have a question about the difference between
algorithms that solve classification and clustering type problems. From what I can tell, it
seems to me that the only difference between classification and clustering is that
algorithm would have supervision so that the folders can be labeled as separate and
that in a clustering algorithm the program would separate the emails by how it thinks
clustering algorithms is correct, then I disagree with the article’s claim that they are
algorithms and classification algorithms are very similar in the way that both separate
data into groups or categories. The only difference is that classification algorithms are
supervised while clustering algorithms are unsupervised, and I believe that this is not a
big enough difference to separate them into separate types of problems solved by
algorithms.
The article also discusses a couple specific algorithms, how they work, and their
different uses and applications. These algorithms include: Naive Bayes, Decision Tree,
information could be useful because I previously didn’t know any specific algorithms
or how they are used, so I believe that now I have a starting point to shoot off of when
learning more about how Artificial Intelligence programs are implemented. The
information about how the algorithms work is useful, however the article does not go
into much detail about how the algorithms are implemented specifically, and only a few
of the descriptions of the algorithms give real world examples of how they are currently
being used. Because of this, I think it would be a good idea in the future to go into more
detail about an algorithm and research how they are implemented in my next
assessment. Also, I didn’t completely understand how all of the algorithms worked like
the Support Vector Machine algorithm and the Random Forest algorithm, so I could
edureka
We can all agree that Artificial Intelligence has created a huge impact on the world’s
economy and will continue to do so since we’re aiding its growth by producing an
Algorithms we can deal with such humungous data. In this blog post, you will
understand the different Artificial Intelligence Algorithms and how they can be used
To get in-depth knowledge of Artificial Intelligence and Machine Learning, you can
enroll for live Machine Learning Engineer Master Program by Edureka with 24/7
a. Classification Algorithms
b. Regression Algorithms
c. Clustering Algorithms
Since the development of complex Artificial Intelligence Algorithms, it has been able
to accomplish this by creating machines and robots that are applied in a wide range
Before we move any further let’s try to understand what Machine Learning is and
combination of both – inputs and outputs simultaneously in order to “learn” the data
Machine Learning is a sub-field of Artificial Intelligence, where we try to bring AI into the
If you’re curious to learn more about Machine Learning, give the following blogs a
read:
Learning
Machines can follow different approaches to learn depending on the data set and
the problem that is being solved. In the below section we’ll understand the different
2. Unsupervised Learning
3. Reinforcement Learning
4. Ensemble Learning
Let’s briefly understand the idea behind each type of Machine Learning.
algorithm learn the data while providing the correct answers or the labels to the
data. This essentially means that the classes or the values to be predicted are
known and well defined for the algorithm from the very beginning.
methods the algorithm doesn’t have correct answers or any answers at all, it is up
to the algorithms discretion to bring together similar data and understand it.
principles by either rewarding them when doing the right thing or punishing upon
doing something wrong, in Reinforcement Learning, there are rewards given to the
algorithm upon every correct prediction thus driving the accuracy higher up.
Here’s a short video recorded by our Machine Learning experts. This will help you
learning.
still land into the issue of having to bump up the performance of our model. In such
cases it might make sense, to use ensemble methods (explained later) to get the
Now let’s understand how Artificial Intelligence algorithms can be used to solve
outputs given unknown inputs, however, here data is the key driver when it comes
1. Classification
2. Regression
3. Clustering
Here’s a table that effectively differentiates each of these categories of problems.
For each category of tasks, we can use specific algorithms. In the below section
complex problems.
a category of problems. In the below section we’ll see the different types of
Classification Algorithms
Classification, as the name suggests is the act of dividing the dependent variable
(the one we try to predict) into classes and then predict a class for a given input. It
falls into the category of Supervised Machine Learning, where the data set needs to
1. Naive Bayes
2. Decision Tree
3. Random Forest
4. Logistic Regression
6. K Nearest Neighbours
Let us break them down and see where they fit in when it comes to application.
Naive Bayes
Naive Bayes algorithm follows the Bayes theorem, which unlike all the other
algorithms in this list, follows a probabilistic approach. This essentially means, that
instead of jumping straight into the data, the algorithm has a set of prior
Explore Curriculum
Once you feed in the data, the algorithm updates these prior probabilities to form
Hence this can be extremely useful in cases where you need to predict whether
your input belongs to either a given list of n classes or does it not belong to any of
them. This can be possible using a probabilistic approach mainly because the
Let us try to understand this with an example, of a person playing golf, depending
● We first try to generate the frequencies with which certain events occur, in
this case, we try to find frequencies of the person playing golf if it’s sunny,
0.64)
questions like “what would be the probability of it being sunny outside and
Here we have P (Sunny |Yes) = 3/9 = 0.33, P(Sunny) = 5/14 = 0.36, P( Yes)= 9/14 = 0.64
You can go through this A Comprehensive Guide To Naive Bayes blog to help you
Decision Tree
The Decision Tree can essentially be summarized as a flowchart-like tree structure
where each external node denotes a test on an attribute and each branch
represents the outcome of that test. The leaf nodes contain the actual predicted
labels. We start from the root of the tree and keep comparing attribute values until
We use this classifier when handling high dimensional data and when little time has
been spent behind data preparation. However, a word of caution – they tend to
overfit and are prone to change drastically even with slight nuances in the training
data.
You can through these blogs to learn more about Decision Trees:
Random Forest
Think of this as a committee of Decision Trees, where each decision tree has been
fed a subset of the attributes of data and predicts on the basis of that subset. The
average of the votes of all decision trees are taken into account and the answer is
given.
which was present in a standalone decision tree, leading to a much more robust
As we can see in the above image, we have 5 decision trees trying to classify a color.
Here 3 of these 5 decision trees predict blue and two have different outputs,
namely green and red. In this case, we take the average of all the outputs, which
working of Random forest algorithm and how it can be used to solve real-world
problems.
Logistic Regression
It’s a go-to method mainly for binary classification tasks. The term ‘logistic’ comes
from the logit function that is used in this method of classification. The logistic
function, also called as the sigmoid function is an S-shaped curve that can take any
real-valued number and map it between 0 and 1 but never exactly at those limits.
Let’s assume that your little brother is trying to get into grad school, and you want
to predict whether he’ll get admitted in his dream establishment. So, based on his
CGPA and the past data, you can use Logistic Regression to foresee the outcome.
Logistic Regression allows you to analyze a set of variables and predict a categorical
outcome. Since here we need to predict whether he will get into the school or not,
Logistic Regression is used to predict house values, customer lifetime value in the
between two classes as far apart as possible. This is called maximum margin
separation.
Another thing to take note of here is the fact that SVM’s take into account only the
support vectors while plotting the hyperplane, unlike linear regression which uses
the entire dataset for that purpose. This makes SVM’s quite useful in situations
Let’s try to understand this with an example. In the below figure we have to classify
So, you start off by drawing a random hyperplane and then you check the distance
between the hyperplane and the closest data points from each class. These closest
data points to the hyperplane are known as Support vectors. And that’s where the
hyperplane will have a maximum distance from each of the support vectors. And this
distance between the hyperplane and the support vectors is known as the margin.
5(2279)
Reviews
5(18299)
Reviews
5(1296)
REINFORCEMENT LEARNING
Reinforcement Learning
Reviews
5(1338)
Reviews
5(8807)
Next
To sum it up, SVM is used to classify data by using a hyperplane, such that the
To learn more about SVM, you can go through this, Using SVM To Predict Heart
Diseases blog.
K Nearest Neighbors
KNN is a non-parametric (here non-parametric is just a fancy term which essentially
means that KNN does not make any assumptions on the underlying data
distribution), lazy learning algorithm (here lazy means that the “training” phase is
fairly short).
Its purpose is to use a whole bunch of data points separated into several classes to
The following points serve as an overview of the general working of the algorithm:
● A positive integer N is specified, along with a new sample
● We select the N entries in our database which are closest to the new sample
However, there are some downsides to using KNN. These downsides mainly revolve
around the fact that KNN works on storing the entire dataset and comparing new
points to existing ones. This means that the storage space increases as our training
set increases. This also means that the estimation time increases in proportion to
The following blogs will help you understand how the KNN algorithm works in
depth:
Now let’s understand how regression problems can be solved by using regression
algorithms.
Regression Algorithms
In the case of regression problems, the output is a continuous quantity. Meaning
that we can use regression algorithms in cases where the target variable is a
where the data set needs to have the labels, to begin with.
Linear Regression
Linear Regression is the most simple and effective regression algorithm. It is
utilized to gauge genuine qualities (cost of houses, number of calls, all out deals
between free and ward factors by fitting the best line. This best fit line is known as
Consider that you are given the challenge to estimate an unknown person’s weight
by just looking at them. With no other values in hand, this might look like a fairly
difficult task, however using your past experience you know that generally speaking
the taller someone is, the heavier they are compared to a shorter person of the
Some of the most popular applications of Linear regression are in financial portfolio
ETAs
Now let’s discuss how clustering problems can be solved by using the K-means
Clustering Algorithms
The basic idea behind clustering is to assign the input into two or more clusters
Learning, where the algorithm learns the patterns and useful insights from data
For example, clustering viewers into similar groups based on their interests, age,
Means Clustering.
K-Means Clustering
K-means is probably the simplest unsupervised learning approach. The idea here is
to gather similar data points together and bind them together in the form of a
cluster. It does this by calculating the centroid of the group of data points.
To carry out effective clustering, k-means evaluates the distance between each
point from the centroid of the cluster. Depending on the distance between the data
point and the centroid, the data is assigned to the closest cluster. The goal of
The ‘K’ in K-means stands for the number of clusters formed. The number of
clusters (basically the number of classes in which your new instances of data can
K-means is used majorly in cases where the data set has points which are distinct
and well separated from each other, otherwise, the clusters won’t be far apart,
rendering them inaccurate. Also, K-means should be avoided in cases where the
data set contains a high amount of outliers or the data set is non-linear.
So that was a brief about K-means algorithm, to learn more you can go through this
using python.
Consider the scenario, you have a decision tree trained on a data set along with a
accuracy is still slightly off than you’d like. In this case, while it might seem that you
have run out of possible things to try, ensemble learning comes to the rescue.
Ensemble Learning – Artificial Intelligence Algorithms – Edureka
You have two different ways in which you can use ensemble learning, in this case,
to bump up your accuracy. Let us say your decision tree was failing on a set of input
test values, what you do now is, to train a new decision tree model and give a
higher weighting to those input test values that your previous model struggled with.
This is also called as Boosting, where our initial tree can be formally stated as a
weak learner, and the mistakes caused by that model pave way for a better and
stronger model.
Another way to approach this is by simply training a whole bunch of trees at once
(this can be done fairly quickly and in a parallel fashion) and then taking outputs
from each tree and averaging them out. So this way, if after training 10 trees, let’s
say 6 trees reply positive to input and 4 trees reply negative, the output you
They are used to reduce the bias and variance in supervised learning techniques.
There are a host of boosting algorithms available, a few of them discussed below:
Gradient Boosting
Gradient Boosting is a boosting algorithm used when we deal with plenty of data to
average predictors to build strong predictor. These boosting algorithms are heavily
distance from that “optimal” model. What we now do is, use gradient mathematics
XGBoost
Pertaining to its extremely high predictive power, XGBoost is one of the go-to
algorithms when it comes to increasing accuracy as it contains both linear & tree
used it in the model for classification of signals from the Large Hadron Collider.
If you want to learn more about Boosting Machine Learning, you can go through
So with this, we come to an end of this Artificial Intelligence Algorithms blog. If you
wish to learn more about Artificial Intelligence, you can give these blogs a read:
If you wish to enroll for a complete course on Artificial Intelligence and Machine
Learning, Edureka has a specially curated Machine Learning Engineer Master Program
that will make you proficient in techniques like Supervised Learning, Unsupervised