0% found this document useful (0 votes)
115 views27 pages

Ism Research Assessment 3

The document summarizes different types of artificial intelligence algorithms and the problems they solve. It discusses supervised, unsupervised, reinforcement, and ensemble learning techniques. Supervised learning algorithms like naive Bayes, decision trees, and support vector machines are used for classification problems to organize data into categories. Unsupervised algorithms like K-means clustering group unlabeled data without supervision. Regression algorithms like linear regression predict continuous valued outputs. The document provides examples of how these algorithms work and are applied to real-world problems.

Uploaded by

api-549257349
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
115 views27 pages

Ism Research Assessment 3

The document summarizes different types of artificial intelligence algorithms and the problems they solve. It discusses supervised, unsupervised, reinforcement, and ensemble learning techniques. Supervised learning algorithms like naive Bayes, decision trees, and support vector machines are used for classification problems to organize data into categories. Unsupervised algorithms like K-means clustering group unlabeled data without supervision. Regression algorithms like linear regression predict continuous valued outputs. The document provides examples of how these algorithms work and are applied to real-world problems.

Uploaded by

api-549257349
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 27

Research Assessment #3

Date: October 16th, 2020

Subject: Artificial Intelligence

MLA Citation:

Artificial Intelligence Algorithms For Beginners. (2020, May 14). Retrieved October 17,

2020, from https://www.edureka.co/blog/artificial-intelligence-algorithms/

Assessment:

The article discusses the major techniques artificial intelligence algorithms,

which are supervised, unsupervised, reinforcement, and ensemble learning, and also

the types of problems solved using each type of algorithm. These types of problems are

classification. regression, and clustering. The article then gives examples of specific

algorithms, which type of algorithm they are, and what category the problems they

solve fall under.

This information could be useful in the future because I can now understand the

differences between different types of artificial intelligence algorithms and I know what

each type of algorithm is used for. For example, I now know that classification types

problems such as sorting emails into Spam folders and non-spam folders and uses

supervised learning algorithms so that the algorithm can check to see if it is classifying

the emails correctly. However, I still have a question about the difference between

algorithms that solve classification and clustering type problems. From what I can tell, it

seems to me that the only difference between classification and clustering is that

classification is supervised while clustering is unsupervised. This would mean that in


the previous example of spam email folders and non spam email folders, a classification

algorithm would have supervision so that the folders can be labeled as separate and

that in a clustering algorithm the program would separate the emails by how it thinks

they are most different without supervision. If my understanding of classification and

clustering algorithms is correct, then I disagree with the article’s claim that they are

separate types of problems solved by algorithms. This is because both clustering

algorithms and classification algorithms are very similar in the way that both separate

data into groups or categories. The only difference is that classification algorithms are

supervised while clustering algorithms are unsupervised, and I believe that this is not a

big enough difference to separate them into separate types of problems solved by

algorithms.

The article also discusses a couple specific algorithms, how they work, and their

different uses and applications. These algorithms include: Naive Bayes, Decision Tree,

Random Forest, Logistic Regression, Support Vector Machines, K Nearest Neighbors,

Linear Regression, K-Means Clustering, Gradient Boosting, and XG Boost. This

information could be useful because I previously didn’t know any specific algorithms

or how they are used, so I believe that now I have a starting point to shoot off of when

learning more about how Artificial Intelligence programs are implemented. The

information about how the algorithms work is useful, however the article does not go

into much detail about how the algorithms are implemented specifically, and only a few

of the descriptions of the algorithms give real world examples of how they are currently

being used. Because of this, I think it would be a good idea in the future to go into more
detail about an algorithm and research how they are implemented in my next

assessment. Also, I didn’t completely understand how all of the algorithms worked like

the Support Vector Machine algorithm and the Random Forest algorithm, so I could

also look into this in my next research assessment.


Artificial Intelligence Algorithms: All
you need to know
Last updated on May 14,202021.3K Views

edureka

3 / 12 Blog from Introduction to Artificial Intelligence

We can all agree that Artificial Intelligence has created a huge impact on the world’s

economy and will continue to do so since we’re aiding its growth by producing an

immeasurable amount of data. Thanks to the advancement in Artificial Intelligence

Algorithms we can deal with such humungous data. In this blog post, you will

understand the different Artificial Intelligence Algorithms and how they can be used

to solve real-world problems.

To get in-depth knowledge of Artificial Intelligence and Machine Learning, you can

enroll for live Machine Learning Engineer Master Program by Edureka with 24/7

support and lifetime access.

Here’s a list of topics that will be covered in this post:

1. What Is Artificial Intelligence?

2. What Is Machine Learning?

3. Types Of Machine Learning


4. Types Of Problems Solved Using Artificial Intelligence Algorithms

5. Artificial Intelligence Algorithms

a. Classification Algorithms

b. Regression Algorithms

c. Clustering Algorithms

d. Ensemble Learning Algorithms

What Is Artificial Intelligence?


To simply put it, Artificial Intelligence is the science of getting machines to think and

make decisions like human beings do.

Since the development of complex Artificial Intelligence Algorithms, it has been able

to accomplish this by creating machines and robots that are applied in a wide range

of fields including agriculture, healthcare, robotics, marketing, business analytics

and many more.

Before we move any further let’s try to understand what Machine Learning is and

how does it is related to AI.

What Is Machine Learning?


Generally, an algorithm takes some input and uses mathematics and logic to

produce the output. In stark contrast, an Artificial Intelligence Algorithm takes a

combination of both – inputs and outputs simultaneously in order to “learn” the data

and produce outputs when given new inputs.


This process of making machines learn from data is what we call Machine Learning.

Artificial Intelligence Algorithm – Artificial Intelligence Algorithms – Edureka

Machine Learning is a sub-field of Artificial Intelligence, where we try to bring AI into the

equation by learning the input data.

If you’re curious to learn more about Machine Learning, give the following blogs a

read:

1. Introduction To Machine Learning: All You Need To Know About Machine

Learning

2. Machine Learning Tutorial for Beginners

3. Machine Learning Algorithms

Machines can follow different approaches to learn depending on the data set and

the problem that is being solved. In the below section we’ll understand the different

ways in which machines can learn.

Types Of Machine Learning


Machine Learning can be done in the following ways:
1. Supervised Learning

2. Unsupervised Learning

3. Reinforcement Learning

4. Ensemble Learning

Let’s briefly understand the idea behind each type of Machine Learning.

What Is Supervised Learning?


In Supervised Learning, as the name rightly suggests, it involves making the

algorithm learn the data while providing the correct answers or the labels to the

data. This essentially means that the classes or the values to be predicted are

known and well defined for the algorithm from the very beginning.

What Is Unsupervised Learning?


The other class falls under Unsupervised Learning, where, unlike supervised

methods the algorithm doesn’t have correct answers or any answers at all, it is up

to the algorithms discretion to bring together similar data and understand it.

What Is Reinforcement Learning?


Along with these two prominent classes, we also have a third class, called

Reinforcement Learning. Just as children are generally “reinforced” certain ideas,

principles by either rewarding them when doing the right thing or punishing upon

doing something wrong, in Reinforcement Learning, there are rewards given to the

algorithm upon every correct prediction thus driving the accuracy higher up.
Here’s a short video recorded by our Machine Learning experts. This will help you

understand the difference between Supervised, Unsupervised and Reinforcement

learning.

What Is Ensemble Learning?


While the above three classes cover most fields comprehensively, we sometimes

still land into the issue of having to bump up the performance of our model. In such

cases it might make sense, to use ensemble methods (explained later) to get the

accuracy higher up.

Now let’s understand how Artificial Intelligence algorithms can be used to solve

different types of problems.

Types Of Problems Solved Using Artificial Intelligence


Algorithms
Algorithms in each category, in essence, perform the same task of predicting

outputs given unknown inputs, however, here data is the key driver when it comes

to picking the right algorithm.

What follows is an outline of categories of Machine Learning problems with a brief

overview of the same:

1. Classification

2. Regression

3. Clustering
Here’s a table that effectively differentiates each of these categories of problems.

Type Of Problems Solved Using AI – Artificial Intelligence Algorithms – Edureka

For each category of tasks, we can use specific algorithms. In the below section

you’ll understand how a category of algorithms can be used as a solution to

complex problems.

Artificial Intelligence Algorithms


As mentioned above, different Artificial Intelligence algorithms can be used to solve

a category of problems. In the below section we’ll see the different types of

algorithms that fall under Classification, Regression and Clustering problems.

Classification Algorithms
Classification, as the name suggests is the act of dividing the dependent variable

(the one we try to predict) into classes and then predict a class for a given input. It

falls into the category of Supervised Machine Learning, where the data set needs to

have the classes, to begin with.


Thus, classification comes into play at any place where we need to predict an

outcome, from a set number of fixed, predefined outcomes.

Classification uses an array of algorithms, a few of them listed below

1. Naive Bayes

2. Decision Tree

3. Random Forest

4. Logistic Regression

5. Support Vector Machines

6. K Nearest Neighbours

Let us break them down and see where they fit in when it comes to application.

Naive Bayes
Naive Bayes algorithm follows the Bayes theorem, which unlike all the other

algorithms in this list, follows a probabilistic approach. This essentially means, that

instead of jumping straight into the data, the algorithm has a set of prior

probabilities set for each of the classes for your target.


Machine Learning Engineer Masters Program

Explore Curriculum

Once you feed in the data, the algorithm updates these prior probabilities to form

something known as the posterior probability.

Hence this can be extremely useful in cases where you need to predict whether

your input belongs to either a given list of n classes or does it not belong to any of

them. This can be possible using a probabilistic approach mainly because the

probabilities thrown for all the n classes will be quite low.

Let us try to understand this with an example, of a person playing golf, depending

on factors like the weather outside.

● We first try to generate the frequencies with which certain events occur, in

this case, we try to find frequencies of the person playing golf if it’s sunny,

rainy, etc outside.

Naive Bayes – Artificial Intelligence Algorithms – Edureka


● Using these frequencies we generate our apriori or initial probabilities (eg,

the probability of overcast is 0.29 while the general probability of playing is

0.64)

● Next up, we generate the posterior probabilities, where we try to answer

questions like “what would be the probability of it being sunny outside and

the person would play golf?”

We use the Bayesian formula here,

P(Yes | Sunny) = P( Sunny | Yes) * P(Yes) / P (Sunny)

Here we have P (Sunny |Yes) = 3/9 = 0.33, P(Sunny) = 5/14 = 0.36, P( Yes)= 9/14 = 0.64

You can go through this A Comprehensive Guide To Naive Bayes blog to help you

understand the math behind Naive Bayes.

Decision Tree
The Decision Tree can essentially be summarized as a flowchart-like tree structure

where each external node denotes a test on an attribute and each branch

represents the outcome of that test. The leaf nodes contain the actual predicted

labels. We start from the root of the tree and keep comparing attribute values until

we reach a leaf node.


Decision Trees – Artificial Intelligence Algorithms – Edureka

We use this classifier when handling high dimensional data and when little time has

been spent behind data preparation. However, a word of caution – they tend to

overfit and are prone to change drastically even with slight nuances in the training

data.

You can through these blogs to learn more about Decision Trees:

1. A Complete Guide On Decision Tree Algorithm

2. Decision Tree: How To Create A Perfect Decision Tree?

Random Forest
Think of this as a committee of Decision Trees, where each decision tree has been

fed a subset of the attributes of data and predicts on the basis of that subset. The

average of the votes of all decision trees are taken into account and the answer is

given.

An advantage of using Random Forest is that it alleviates the problem of overfitting

which was present in a standalone decision tree, leading to a much more robust

and accurate classifier.

Random Forest – Artificial Intelligence Algorithms – Edureka

As we can see in the above image, we have 5 decision trees trying to classify a color.

Here 3 of these 5 decision trees predict blue and two have different outputs,

namely green and red. In this case, we take the average of all the outputs, which

gives blue as the highest weightage.


Here’s a blog on Random Forest Classifier that will help you understand the

working of Random forest algorithm and how it can be used to solve real-world

problems.

Logistic Regression
It’s a go-to method mainly for binary classification tasks. The term ‘logistic’ comes

from the logit function that is used in this method of classification. The logistic

function, also called as the sigmoid function is an S-shaped curve that can take any

real-valued number and map it between 0 and 1 but never exactly at those limits.

Logistic Regression – Artificial Intelligence Algorithms – Edureka

Let’s assume that your little brother is trying to get into grad school, and you want

to predict whether he’ll get admitted in his dream establishment. So, based on his

CGPA and the past data, you can use Logistic Regression to foresee the outcome.
Logistic Regression allows you to analyze a set of variables and predict a categorical

outcome. Since here we need to predict whether he will get into the school or not,

which is a classification problem, logistic regression would be ideal.

Logistic Regression is used to predict house values, customer lifetime value in the

insurance sector, etc.

Support Vector Machine


An SVM is unique, in the sense that it tries to sort the data with the margins

between two classes as far apart as possible. This is called maximum margin

separation.

Another thing to take note of here is the fact that SVM’s take into account only the

support vectors while plotting the hyperplane, unlike linear regression which uses

the entire dataset for that purpose. This makes SVM’s quite useful in situations

when data is in high dimensions.

Let’s try to understand this with an example. In the below figure we have to classify

data points into two different classes (squares and triangles).


Support Vector Machine – Artificial Intelligence Algorithms – Edureka

So, you start off by drawing a random hyperplane and then you check the distance

between the hyperplane and the closest data points from each class. These closest

data points to the hyperplane are known as Support vectors. And that’s where the

name comes from, Support Vector Machine.

The hyperplane is drawn based on these support vectors and an optimum

hyperplane will have a maximum distance from each of the support vectors. And this

distance between the hyperplane and the support vectors is known as the margin.

Artificial Intelligence Training

NATURAL LANGUAGE PROCESSING WITH PYTHON CERTIFICATION COURSE

Natural Language Processing with Python Certification Course


Reviews

5(2279)

DEEP LEARNING WITH TENSORFLOW 2.0 CERTIFICATION TRAINING

Deep Learning with TensorFlow 2.0 Certification Training

Reviews

5(18299)

GRAPHICAL MODELS CERTIFICATION TRAINING

Graphical Models Certification Training

Reviews

5(1296)

REINFORCEMENT LEARNING

Reinforcement Learning

Reviews
5(1338)

MACHINE LEARNING WITH MAHOUT CERTIFICATION TRAINING

Machine Learning with Mahout Certification Training

Reviews

5(8807)

Next

To sum it up, SVM is used to classify data by using a hyperplane, such that the

distance between the hyperplane and the support vectors is maximum.

To learn more about SVM, you can go through this, Using SVM To Predict Heart

Diseases blog.

K Nearest Neighbors
KNN is a non-parametric (here non-parametric is just a fancy term which essentially

means that KNN does not make any assumptions on the underlying data

distribution), lazy learning algorithm (here lazy means that the “training” phase is

fairly short).

Its purpose is to use a whole bunch of data points separated into several classes to

predict the classification of a new sample point.

The following points serve as an overview of the general working of the algorithm:
● A positive integer N is specified, along with a new sample

● We select the N entries in our database which are closest to the new sample

● We find the most common classification of these entries

● This is the classification we give to the new sample

However, there are some downsides to using KNN. These downsides mainly revolve

around the fact that KNN works on storing the entire dataset and comparing new

points to existing ones. This means that the storage space increases as our training

set increases. This also means that the estimation time increases in proportion to

the number of training points.

The following blogs will help you understand how the KNN algorithm works in

depth:

1. A Practical Implementation Of KNN Algorithm In R

2. K-Nearest Neighbors Algorithm Using Python

Now let’s understand how regression problems can be solved by using regression

algorithms.

Regression Algorithms
In the case of regression problems, the output is a continuous quantity. Meaning

that we can use regression algorithms in cases where the target variable is a

continuous variable. It falls into the category of Supervised Machine Learning,

where the data set needs to have the labels, to begin with.
Linear Regression
Linear Regression is the most simple and effective regression algorithm. It is

utilized to gauge genuine qualities (cost of houses, number of calls, all out deals

and so forth.) in view of the consistent variable(s). Here, we build up a connection

between free and ward factors by fitting the best line. This best fit line is known as

regression line and spoken to by a direct condition Y= a *X + b.

Linear Regression – Artificial Intelligence Algorithms – Edureka

Let us take a simple example here to understand linear regression.

Consider that you are given the challenge to estimate an unknown person’s weight

by just looking at them. With no other values in hand, this might look like a fairly

difficult task, however using your past experience you know that generally speaking

the taller someone is, the heavier they are compared to a shorter person of the

same build. This is linear regression, in actuality!


However, linear regression is best used in approaches involving a low number of

dimensions. Also, not every problem is linearly separable.

Some of the most popular applications of Linear regression are in financial portfolio

prediction, salary forecasting, real estate predictions and in traffic in arriving at

ETAs

Now let’s discuss how clustering problems can be solved by using the K-means

algorithm. Before that, let’s understand what clustering is.

Clustering Algorithms
The basic idea behind clustering is to assign the input into two or more clusters

based on feature similarity. It falls into the category of Unsupervised Machine

Learning, where the algorithm learns the patterns and useful insights from data

without any guidance (labeled data set).

For example, clustering viewers into similar groups based on their interests, age,

geography, etc can be done by using Unsupervised Learning algorithms like K-

Means Clustering.

K-Means Clustering
K-means is probably the simplest unsupervised learning approach. The idea here is

to gather similar data points together and bind them together in the form of a

cluster. It does this by calculating the centroid of the group of data points.

To carry out effective clustering, k-means evaluates the distance between each

point from the centroid of the cluster. Depending on the distance between the data
point and the centroid, the data is assigned to the closest cluster. The goal of

clustering is to determine the intrinsic grouping in a set of unlabelled data.

K-means – Artificial Intelligence Algorithms – Edureka

The ‘K’ in K-means stands for the number of clusters formed. The number of

clusters (basically the number of classes in which your new instances of data can

fall into) is determined by the user.

K-means is used majorly in cases where the data set has points which are distinct

and well separated from each other, otherwise, the clusters won’t be far apart,

rendering them inaccurate. Also, K-means should be avoided in cases where the

data set contains a high amount of outliers or the data set is non-linear.

So that was a brief about K-means algorithm, to learn more you can go through this

content recorded by our Machine Learning experts.

K Means Clustering Algorithm | Edureka


In this video, you learn the concepts of K-Means clustering and its implementation

using python.

Ensemble Learning Algorithms


In cases where data is of abundance and prediction precision is of high value,

boosting algorithms come into the picture.

Consider the scenario, you have a decision tree trained on a data set along with a

whole bunch of hyperparameter tuning already performed, however, the final

accuracy is still slightly off than you’d like. In this case, while it might seem that you

have run out of possible things to try, ensemble learning comes to the rescue.
Ensemble Learning – Artificial Intelligence Algorithms – Edureka

Machine Learning Engineer Masters Program

Weekday / Weekend Batches

See Batch Details

You have two different ways in which you can use ensemble learning, in this case,

to bump up your accuracy. Let us say your decision tree was failing on a set of input

test values, what you do now is, to train a new decision tree model and give a

higher weighting to those input test values that your previous model struggled with.

This is also called as Boosting, where our initial tree can be formally stated as a

weak learner, and the mistakes caused by that model pave way for a better and

stronger model.
Another way to approach this is by simply training a whole bunch of trees at once

(this can be done fairly quickly and in a parallel fashion) and then taking outputs

from each tree and averaging them out. So this way, if after training 10 trees, let’s

say 6 trees reply positive to input and 4 trees reply negative, the output you

consider is positive. This is formally known as Bagging.

They are used to reduce the bias and variance in supervised learning techniques.

There are a host of boosting algorithms available, a few of them discussed below:

Gradient Boosting
Gradient Boosting is a boosting algorithm used when we deal with plenty of data to

make a prediction with high prediction power. It combines multiple weak or

average predictors to build strong predictor. These boosting algorithms are heavily

used to refine the models in data science competitions.

Here, we consider an “optimal” or best model, so essentially our model is at some

distance from that “optimal” model. What we now do is, use gradient mathematics

and try to get our model closer to the optimal space.

XGBoost
Pertaining to its extremely high predictive power, XGBoost is one of the go-to

algorithms when it comes to increasing accuracy as it contains both linear & tree

learning algorithms making it 10 times faster than most boosting techniques.

It is the holy grail algorithm when it comes to hackathons, it is no wonder CERN

used it in the model for classification of signals from the Large Hadron Collider.
If you want to learn more about Boosting Machine Learning, you can go through

this, Comprehensive Guide To Boosting Machine Learning Algorithms blog.

So with this, we come to an end of this Artificial Intelligence Algorithms blog. If you

wish to learn more about Artificial Intelligence, you can give these blogs a read:

1. Artificial Intelligence – What It Is And How Is It Useful?

2. A Comprehensive Guide To Artificial Intelligence With Python

3. Artificial Intelligence Tutorial: All you need to know about AI

4. AI vs Machine Learning vs Deep Learning

5. AI Applications: Top 10 Real World Artificial Intelligence Applications

If you wish to enroll for a complete course on Artificial Intelligence and Machine

Learning, Edureka has a specially curated Machine Learning Engineer Master Program

that will make you proficient in techniques like Supervised Learning, Unsupervised

Learning, and Natural Language Processing. It includes training on the latest

advancements and technical approaches in Artificial Intelligence & Machine Learning

such as Deep Learning, Graphical Models and Reinforcement Learning.

You might also like