Chapter 2

Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

AI Project Cycle

Uncategorized
AI Project Cycle

The AI Project Cycle provides us with an appropriate framework that can lead us towards the goal. The AI Project Cycle
mainly has 5 stages:

AI Project Cycle

• You need to acquire data that will become the base of your project as it will help you in understanding the
parameters that are related to problem scoping.
• You go for data acquisition by collecting data from various reliable and authentic sources. Since the data you
collect would be in large quantities, you can try to give it a visual image of different types of representations like
graphs, databases, flow charts, maps, etc. This makes it easier for you to interpret the patterns which your acquired
data follows.
• After exploring the patterns, you can decide upon the type of model you would build to achieve the goal. For this,
you can research online and select various models which give a suitable output.
• You can test the selected models and figure out which is the most efficient one.
• The most efficient model is now the base of your AI project and you can develop your algorithm around it.
• Once the modelling is complete, you now need to test your model on some newly fetched data. The results will
help you in evaluating your model and improve it.
• Finally, after evaluation, the project cycle is now complete and what you get is your AI project.

Problem Scoping

Scoping a problem is not that easy as we need to have a deeper understanding around it so that the picture becomes
clearer while we are working to solve it. Hence, we use the 4Ws Problem Canvas to help us out.

Who?

The “Who” block helps in analysing the people getting affected directly or indirectly due to it. Under this, we find out
who the ‘Stakeholders’ of this problem are and what we know about them. Stakeholders are the people who face this
problem and would be benefitted from the solution.

What?

Under the “What” block, you need to look into what you have on hand. At this stage, you need to determine the nature of
the problem. What is the problem and how do you know that it is a problem? Under this block, you also gather evidence
to prove that the problem you have selected actually exists. Newspaper articles, Media, announcements, etc are some
examples.

Where?

Now that you know who is associated with the problem and what the problem actually is; you need to focus on the
context/situation/location of the problem. This block will help you look into the situation in which the problem arises, its
context of it, and the locations where it is prominent.
Why?

You have finally listed down all the major elements that affect the problem directly. Now it is convenient to understand
who the people that would be benefitted by the solution are; what is to be solved; and where will the solution be
deployed.

Problem Statement

The Problem Statement Template helps us to summarise all the key points into one single Template so that in the future,
whenever there is a need to look back at the basis of the problem, we can take a look at the Problem Statement Template
and understand the key elements of it.

Data Acquisition

As the term clearly mentions, this stage is about acquiring data for the project. Data can be a piece of information or facts
and statistics collected together for reference or analysis. Whenever we want an AI project to be able to predict an output,
we need to train it first using data.

For any AI project to be efficient, the training data should be authentic and relevant to the problem statement scoped.

Data Features

Data features refer to the type of data you want to collect. There can be various ways in which you can collect data. Some
of them are:

• Surveys
• Web Scraping
• Sensors
• Cameras
• Observations
• API (Application Program Interface)

Data Exploration

To analyse the data, you need to visualise it in some user-friendly format so that you can:

• Quickly get a sense of the trends, relationships and patterns contained within the data.
• Define the strategy for which model to use at a later stage.
• Communicate the same to others effectively. To visualise data, we can use various types of visual representations.

Modelling

The ability to mathematically describe the relationship between parameters is the heart of every AI model. Thus,
whenever we talk about developing AI models, it is the mathematical approach to analyzing data that we refer to.
Generally, AI models can be classified as follows:

Rule Based Approach

It refers to the AI modelling where the rules are defined by the developer. The machine follows the rules or instructions
mentioned by the developer and performs its task accordingly. For example, we have a dataset that tells us about the
conditions on the basis of which we can decide if an elephant may be spotted or not while on safari. The parameters are
Outlook, Temperature, Humidity and Wind.

This is known as a rule-based approach because we fed the data along with rules to the machine and the machine after
getting trained on them is now able to predict answers for the same.

A drawback/feature of the rule-based approach is that the learning is static. The machine once trained, does not take
into consideration any changes made in the original training dataset. That is, if you try testing the machine on a dataset
that is different from the rules and data you fed it at the training stage, the machine will fail and will not learn from its
mistake. Once trained, the model cannot improvise itself on the basis of feedback.

Learning Based Approach

It refers to AI modelling where the machine learns by itself. Under the Learning Based approach, the AI model gets
trained on the data fed to it and then is able to design a model which is adaptive to the change in data.

If the machine learning based model is trained with X type of data and the machine designs the algorithm around it, the
model would modify itself according to the changes which occur in the data so that all the exceptions are handled in this
case.

The learning-based approach can further be divided into three parts:

• Supervised Learning
• Unsupervised Learning
• Reinforcement Learning

Supervised Learning

In a supervised learning model, the dataset which is fed to the machine is labelled. In other words, we can say that the
dataset is known to the person who is training the machine only then he/she is able to label the data.

There are two types of Supervised Learning models:

Classification: Where the data is classified according to the labels. For example, in the grading system, students are
classified on the basis of the grades they obtain with respect to their marks in the examination. This model works on a
discrete dataset which means the data need not be continuous.

Regression: Such models work on continuous data. For example, if you wish to predict your next salary, then you would
put in the data of your previous salary, any increments, etc., and would train the model. Here, the data which has been fed
to the machine is continuous.

Unsupervised Learning

An unsupervised learning model works on unlabelled dataset. This means that the data which is fed to the machine is
random and there is a possibility that the person who is training the model does not have any information regarding it.
The unsupervised learning models are used to identify relationships, patterns and trends out of the data which is fed into
it. It helps the user in understanding what the data is about and what are the major features identified by the machine in it.

Unsupervised learning models can be further divided into two categories:

Clustering: Refers to the unsupervised learning algorithm which can cluster the unknown data according to the patterns
or trends identified out of it. The patterns observed might be the ones that are known to the developer or it might even
come up with some unique patterns out of it.

Dimensionality Reduction: We humans are able to visualise upto 3-Dimensions only but according to a lot of theories
and algorithms, there are various entities that exist beyond 3-Dimensions. For example, in Natural language Processing,
the words are considered to be N-Dimensional entities. Which means that we cannot visualise them as they exist beyond
our visualisation ability. Hence, to make sense out of it, we need to reduce their dimensions. Here, a dimensionality
reduction algorithm is used.

Evaluation

Once a model has been made and trained, it needs to go through proper testing so that one can calculate the efficiency
and performance of the model. Hence, the model is tested with the help of Testing Data (which was separated out of the
acquired dataset at the Data Acquisition stage) and the efficiency of the model is calculated on the basis of the parameters
mentioned below:
Neural Networks

Neural networks are loosely modelled by how neurons in the human brain behave. The key advantage of neural networks
are that they are able to extract data features automatically without needing the input of the programmer. A neural
network is essentially a system of organizing machine learning algorithms to perform certain tasks. It is a fast and
efficient way to solve problems for which the dataset is very large, such as in images.

Some of the features of a Neural Network are listed below:

You might also like