Project Report Big Data
Project Report Big Data
Project Report Big Data
Working on CNN
The basic building block of a CNN is the convolutional
layer. In this layer, a set of filters (also known as kernels)
are applied to the input image, with each filter learning to
detect a specific feature or pattern. The filters are typically
small in size (e.g. 3x3 or 5x5), and they slide over the input
image one pixel at a time, producing a 2D activation map
(also known as a feature map) that highlights the regions of
the input image that correspond to the learned feature.
The output of a convolutional layer is then passed
D. Data Augnmentation through a non-linear activation function, such as ReLU
Data augmentation refers to the technique of artificially (Rectified Linear Unit), which introduces non-linearity into
increasing the size of a training dataset by applying vari- the network and allows it to capture more complex patterns.
ous transformations to the existing data. The goal of data The resulting feature maps are then fed into a pooling layer,
augmentation is to improve the performance of a machine which reduces the spatial size of the feature maps by taking
learning model by increasing the diversity of the training the maximum or average value over small regions (e.g. 2x2)
data, which helps the model to generalize better to new, in each feature map. This pooling operation helps to make
unseen data. Some common features are: the network more robust to small variations in the input
In addition to image data, data augmentation can also data, and reduces the number of parameters in the network,
be applied to other types of data, such as text or audio making it computationally more efficient.
data. For example, in natural language processing (NLP), The output of the pooling layer is then passed through
data augmentation techniques can include paraphrasing or another set of convolutional and pooling layers, creating a
altering the order of sentences in a document. By increasing hierarchy of feature maps that capture increasingly complex
the size and diversity of the training data, data augmentation patterns and structures in the input data. Finally, the output
can help to prevent overfitting and improve the robustness of the last convolutional layer is flattened into a 1D vector
and generalization ability of the model. In this model, we and passed through one or more fully connected layers,
have used ImageDataGenerator to apply augmentation with which perform a non-linear transformation of the features
few of the above mentioned features. and produce the final output of the network.
During training, the parameters (i.e. weights and bi-
ases) of the network are learned using backpropagation,
E. Model Building a gradient-based optimization algorithm that adjusts the
The model used for this project is one the most common parameters to minimize the difference between the network’s
models used for images recognition, CNN (Convolutional output and the desired output (e.g. a class label for an image).
Neural Network). The learning process involves iteratively feeding training
Convolutional Neural Networks (CNNs) are a type of examples through the network, computing the error between
artificial neural network that are commonly used for image the predicted output and the true output, and updating the
recognition and computer vision tasks. CNNs are designed parameters of the network using the gradients of the error
to automatically learn spatial hierarchies of features from with respect to the parameters.
2020BTECHCSE024
JK Lakshmipat University, Jaipur
Vikas Janu
2020BTECHCSE083
JK Lakshmipat University, Jaipur
F. Source Code
https://github.com/chitra-sl/Big-data-Project
G. IPR Certificate
We, Chitra S. Lakhani and Vikas Janu, with this certify
that the project work submitted by us, entitled Olympic
Sports Recognition, to our supervisors, Dr. Alok Kumar and
Dr. Utsav Upadhyay, in partial fulfilment of the requirements
for the course is a Bonafede work carried out by us and
has not been previously submitted to any other course. We
further certify that no part of this work shall be published,
reproduced, or distributed in any form without the prior
permission of our supervisors. We understand that any such
unauthorized use of the project work may be considered a
violation of academic ethics and result in severe penalties.
We also affirm that the project work has been carried out
under the ethical standards and guidelines set forth by our
supervisors. We acknowledge that our supervisor has the
right to make any modifications or revisions to the project
work that may be deemed necessary. We also agree to abide
by any additional terms and conditions as stipulated by our
supervisor.
Chitra Lakhani