Research Work 2023
Research Work 2023
Research Work 2023
PAPER NAME
Research_Work_2023.docx
12 Pages 554.1KB
Nov 29, 2023 3:25 PM GMT+5:30 Nov 29, 2023 3:26 PM GMT+5:30
Summary
Research Paper
35
Abstract: In the dynamic field of computer vision and image I.
classification, the choice of a deep learning model plays a crucial
role in achieving optimal performance across various applications.
14
INTRODUCTION
This study conducts a comparative analysis of three distinct
convolutional neural network (CNN) architectures: MobileNetV2, In the rapidly evolving digital era, image classification has
VGG19, and a simplified CNN model. The objective is to assess become a pivotal task across various applications, including
45
their effectiveness in image classification tasks, considering factors autonomous vehicles, healthcare, security systems, and social
like accuracy, computational efficiency, and model complexity. media platforms. The accurate classification of images is
fundamental in these sectors, serving as the foundation
6
for
To ensure a thorough evaluation, a diverse dataset covering subsequent analysis and decision-making processes. Convolutional
multiple categories is utilized, enabling a robust assessment of the Neural Networks (CNNs) have significantly enhanced the
models' generalization capabilities. The training process involves performance of image classification tasks by autonomously
hyperparameter optimization for fair comparisons, and the models learning spatial hierarchies of features. This research paper
7
undergo rigorous testing on a separate validation set to measure explores the effectiveness of three CNN architectures:
their performance under real-world conditions. MobileNetV2, VGG19, and a simple CNN model in image
classification.
4
The experimental results provide nuanced insights into the
strengths and weaknesses of each architecture. MobileNetV2, MobileNetV2, developed by Google, is an efficient and lightweight
recognized for its lightweight design, demonstrates notable model tailored for mobile and embedded vision applications. It
efficiency in computational resource usage, making it well-suited utilizes inverted residuals and linear bottlenecks to balance
for deployment on resource-constrained devices. VGG19, computational efficiency and model accuracy, crucial in resource-
57
characterized by its deep and intricate structure, exhibits a strong limited mobile and embedded scenarios. This research focuses on
ability to capture complex hierarchical features, albeit with investigating the effectiveness of MobileNetV2 in such
increased computational demands. The simplified CNN model, applications.
designed to strike a balance between complexity and performance,
emerges as a practical alternative in scenarios where a compromise In contrast, VGG19, originating from the Visual Graphics Group
between accuracy and resource efficiency is sought. at Oxford, is a deeper and more complex model renowned for its
outstanding performance on the ImageNet dataset. Its depth and
52
This research contributes to the ongoing discourse surrounding the complexity, coupled with high performance, make it a popular
selection of CNN architectures for image classification tasks. By choice for image classification. However, VGG19 comes with
offering empirical evidence, it equips practitioners and researchers significantly higher computational requirements compared to
with valuable insights to inform their decisions based on the MobileNetV2. This study aims to provide a detailed analysis of
specific requirements of their applications. VGG19's performance and computational demands, offering
comprehensive insights into its suitability for image classification
tasks.
CNN 0.9421106
MobileNetV2 0.7725232888888888
VGG19 0.9116561666666667
B. Accuracy Comparison
The evaluation of three distinct models—Simple CNN,
MobileNetV2, and VGG19—on the CIFAR-10 dataset holds
1
A. AUC-ROC Comparison pivotal significance in our study. CIFAR-10, comprising 60,000
32x32 color images distributed across 10 classes, presents a
3
The AUC-ROC (Area Under the Receiver 6 Operating formidable benchmark for image classification. This section
Characteristic) score stands as a widely employed metric for explores the comparative examination of accuracy attained by each
assessing the performance of models in binary classification model, providing insights into their1
respective capabilities in
navigating the intricacies embedded within the CIFAR-10 dataset.
18
dataset. The analysis of these loss metrics offers valuable insights
1. Model Descriptions into the convergence and generalization capabilities of each model.
Following graph presents the accuracy results for the three models
on the CIFAR-10 dataset.
C. Loss Comparison
The comprehensive evaluation of deep learning models is crucial
for discerning their efficacy in tackling specific tasks. This section
presents an in-depth comparison of the training and validation
losses incurred by three distinct models—Simple CNN,
11
MobileNetV2, and VGG19—when applied to the CIFAR-10
The loss comparison of individual function is also illustrated by
following figures
D. VGG19 Model
B. CNN - Our Model
C. Confusion Matrix
C. MobileNetV2 Model
To gauge the classification prowess13
of the proposed models,
confusion matrices are employed as a fundamental tool in the 13
evaluation of machine learning classifiers. These matrices offer a
detailed breakdown of the model's predictions, providing valuable
13
insights into its capacity to accurately classify instances across
diverse classes.
C.VGG19 Model
A. CNN - Our Model
The study findings reveal that among the three models assessed, a
simple CNN model attains the highest accuracy. This implies that
7
a streamlined and adaptable architecture
49
can be highly effective for
B. MobileNetV2 Model image classification tasks. The added advantage of the simple CNN
model lies in its lower computational cost, emphasizing its Conversely, the Simple CNN model proves to be adept at capturing
practical utility. Although the MobileNetV2 and VGG19 models intricate features within images, resulting in commendable
exhibit comparatively lower accuracies in this study, it's worth accuracy. However, its considerable computational demands may
54
noting that their pre-trained weights and transfer learning limit its practicality in resource-constrained environments. The
capabilities may confer advantages for particular image delicate balance between computational efficiency and accuracy
classification tasks. emerges as a pivotal factor in the decision-making process, 14
underscoring the importance of choosing models judiciously based
on the specific requirements of the image classification task.
REFERENCES
[1] Kaya, Yasin, and Ercan Gürsoy. "A MobileNet-based CNN model with a novel fine-tuning mechanism for COVID-19 infection detection." Soft Computing 27.9
(2023): 5521-5535.
[2] Jasil, SP Godlin, and V. Ulagamuthalvi. "Deep learning architecture using transfer learning for classification of skin lesions." Journal of Ambient Intelligence and
Humanized Computing (2021): 1-8.
[3] Bansal, Monika, et al. "Transfer learning for image classification using VGG19: Caltech-101 image data set." Journal of ambient intelligence and humanized computing
(2021): 1-12.
[4] Sharma, Atul, and Gurbakash Phonsa. "Image classification using CNN." Proceedings of the International Conference on Innovative Computing & Communication
(ICICC). 2021.
[5] Gulzar, Yonis. "Fruit image classification model based on MobileNetV2 with deep transfer learning technique." Sustainability 15.3 (2023): 1906.
[6] Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., & Chen, L. C. (2018). MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Simonyan, K., & Zisserman, A. (2014). Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556.
[8] Farabet, C., Couprie, C., Najman, L., & LeCun, Y. (2013). Learning Hierarchical Features for Scene Labeling. IEEE Transactions on Pattern Analysis and Machine
Intelligence (TPAMI), 35(8), 1915-1929.
[9] Krizhevsky, A., Sutskever, I., & Hinton, G. E. (2012). ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information
Processing Systems (NIPS).
[10] Tan, M., & Le, Q. V. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the International Conference on Machine
Learning (ICML).
[11] Verma, A., & Chandra, D. (2020). A Comprehensive Review on Convolutional Neural Network with Various Applications. Artificial Intelligence Review, 53(8),
5455-5505)
[12] Lane, N. D., Bhattacharya, S., & Georgiev, P. (2019). Practical Deep Learning for Cloud, Mobile, and Edge. O'Reilly Media.
[13] Goodfellow, I., Bengio, Y., Courville, A., & Bengio, Y. (2016). Deep Learning (Vol. 1). MIT press Cambridge.
[14] Stanford University. (Course materials and lecture notes on Convolutional Neural Networks for Visual Recognition. Available online: http://cs231n.stanford.edu/)
[15] Rosebrock, A. (2018). Deep Learning for Computer Vision with Python. PyImageSearch.
[16] Simonyan, Karen, and Andrew Zisserman. "Very deep convolutional networks for large-scale image recognition." arXiv preprint arXiv:1409.1556 (2014).
Similarity Report ID: oid:28506:47471634
TOP SOURCES
The sources with the highest number of matches within the submission. Overlapping sources will not be
displayed.
mdpi.com
5 <1%
Internet
Sources overview
Similarity Report ID: oid:28506:47471634
Omar Jilani Jidan, Susmoy Paul, Anirban Roy, Sharun Akter Khushbu, ...
14 <1%
Crossref
Sources overview
Similarity Report ID: oid:28506:47471634
ncbi.nlm.nih.gov
29 <1%
Internet
dspace.bracu.ac.bd
31 <1%
Internet
indico.desy.de
32 <1%
Internet
Sources overview
Similarity Report ID: oid:28506:47471634
digital.lib.washington.edu
34 <1%
Internet
iaeme.com
37 <1%
Internet
ijournalse.org
38 <1%
Internet
papers.ssrn.com
39 <1%
Internet
Intercollege on 2023-10-29
41 <1%
Submitted works
Xu, Chunyan, Canyi Lu, Xiaodan Liang, Junbin Gao, Wei Zheng, Tianjian...
43 <1%
Crossref
amitos.library.uop.gr
44 <1%
Internet
Sources overview
Similarity Report ID: oid:28506:47471634
data-science-ua.com
45 <1%
Internet
Sources overview
Similarity Report ID: oid:28506:47471634
doras.dcu.ie
57 <1%
Internet
ebin.pub
58 <1%
Internet
image-net.org
59 <1%
Internet
library.ndsu.edu
60 <1%
Internet
Qian Zhao, Shuzhi Sam Ge, Mao Ye, Sibang Liu, Wei He. "Learning Salie...
62 <1%
Crossref
Sources overview