Software Engineering Software Requirements Specification (SRS) Document
Software Engineering Software Requirements Specification (SRS) Document
Software Engineering Software Requirements Specification (SRS) Document
September 5, 2021
Version 1.0
By
Mrityunjay Mishra
Shaikh Azim Abdul Rahim
Allaudin Hussain Shaikh
Anand Kumar Rai
Revisions
1
Table of Contents
1.Introduction
1.1 Purpose
1.5 References
2.Overall Description
3. System Requirements
5.Non-functional Requirements
1.1 Purpose
Due to an advent rise in technology, it has almost become very difficult to distinguish between
deepfakes and real video. Deepfake videos are usually made by using StyleGAN’s(Generative
Adversial Network) framework and other deep learning technologies. The danger of this is that
technology can be used to make people believe something is real when it is not actually. Few
applications such as FaceApp and Fake App are built on this technology to be used on mobile
and desktop, which results in affecting a person’s integrity. Identifying, authenticating, and
categorizing these videos has become a necessity. It’s not just about these latest advancements in
creating fake images and video, it is the injection of these techniques into an ecosystem that is
already promoting fake news, sensational, and conspiracy theories. For finding a secured
solution for this we need to develop an anti-deepfake technology that will be implementing
deepfake detection models with different features using a deep learning approach to detect these
videos by isolating, analyzing, and verifying the content.
DB Database
ER Entity Relationship
AI Artificial Intelligence
ML Machine Learning
3
allows software developers and software engineers to use
This project is a prototype for implementing AI and Digital Forensics concepts collectively to
develop an Anti-Deepfake Technology a proven methods to detect deepfakes. This project has
been implemented on RnD Cell of our department under the guidance of Dr. Sunil Rathod. This
project is useful for the all the LEA’s, Security Agencies, Media Companies & Intelligence
Authorities.
1.4 Scope
The technology of image and video editing has made huge leaps in recent years and the
emergence of deepfake images has caused serious problems already and has the potential to
become more of an issue in the future as the technology improves.
Adding additional images to the dataset, as having more data will inherently make the
models more accurate.
Update the associated web App to be more user-friendly and have more features for
deepfake image classification.
Expand the scope of this project to include being able to scan video for deepfakes as this
is increasingly becoming problematic technology.
In order to detect deepfake videos, the first thing we need to do is that, we need to
understand the actual picture, and how actually these are created, which will help us in
understanding the gaps in it.
In the approach, the frame-level scene inconsistency is the first feature that is exploited,
if the encoder is not aware of skin or other scene information, there will be boundary
effect due to seamed fusion between the new face and the rest of the frame which is
another weak point.
The other major weakness that is exploited here is the source of multiple anomalies and
leads to a flickering phenomenon in the face region. This flickering is common to most
4
fake videos, even though this is hard to find with our naked eyes, it can be easily captured
by a pixel-level CNN feature extraction.
The method here will take an advantage of the fact that during deepfake video generation,
the deepfake algorithm could only generate fake faces with specific size and resolution.
In order to match and fit the arrangement of the source’s face on original videos, a further
blur function must be added to the synthesized faces.
The transformation causes exclusive blur inconsistency between the generated phase and
its background outcome deep fake videos, the method detects such inconsistency by
comparing the blurred synthesized areas ROI and surrounding context with the dedicated
Haar Wavelet Transform function.
Accordingly, the videos will be converted into frames, image extraction would be done
using MOVIEPY, then stepping ahead with preprocessing, mouth nodal points will be
captured from each frame then CNN will be used to classify real or fake amongst all.
1.5 References
[1]https://ayushbasral.medium.com/deepfake-detection-using-resnxt-and-lstm-bcc08c086f84.
[2]https://lab.irt.de/a-system-for-deepfake-detection-dfirt/ .
[3]FaceForensics++: Learning to Detect Manipulated Facial Images by Andreas Rossler ¨ 1
Davide Cozzolino2 Luisa Verdoliva2 Christian Riess3 Justus Thies1 Matthias Nießner1.
[4]Deepfake Video Detection Using Recurrent Neural Networks David Guera Edward J. Delp ¨
Video and Image Processing Laboratory (VIPER), Purdue Universit.
[5]Digital Forensics and Analysis of Deepfake Videos Mousa Tayseer Jafar Mohammad
Ababneh Mohammad Al-Zoube Ammar Elhassan
[7] Deepfake Detection using ResNxt and LSTM _ by Ayush Basral _ Medium
2. Overall Description
2.1 Product perspective:
Security:- The underlying Artificial Intelligence (AI) technologies are used to manipulate
data (e.g. video and audio), and enable attackers to accurately impersonate individuals. This has
initiated security challenges for organizations and individuals. However, the development of
deepfakes is progressing quickly, with the use of socio-engineering for AI-assisted vishing. The
5
advanced deepfakes of high-profile individuals or executives will threaten to undermine digital
communications, spreading highly credible fake news.
While deepfake technologies are evolving rapidly and opening-up new positive innovations,
they also raise concerns as they are not trust-worthy due to possible malicious users.
For example,
a. deepfakes can be leveraged to defame, impersonate, and spread disinformation;
b. audio deepfakes can be used for scams and voice phishing, which opens-up for security
concerns;
c. visual deepfakes can be targeted at affecting the reputation of specific individuals.
PII (Personal Identifiable Information) :- Perhaps there should be a bill to criminalize the
creation and distribution of deepfakes, or at least some realistic legislation around its creation
and use. So much of a person’s life is online now, from social media to blogging, renewing a
driver’s license, paying a parking fine, signing up for college courses, working from home — the
list goes on. Unless every entity a person conducts online business with over their lifetime has a
way to identify fake photographs and voice passwords in the next two to three years, databases
will quickly be filled with people who don’t actually exist.
Nearly every entity — federal, state, civil, county and corporate — requires some kind of
photo ID to sign up for a new account or service. However, only a handful of countries consider
a person’s facial geometry or voice signatures to be a form of personally identifiable information
(PII), and no country currently considers a person’s gait to be PII. In these three areas alone,
humanity is already behind artificial intelligence and its uses. It is time for the cybersecurity
community to apply a bit of rational thinking to catch up.
UI:-
Attribute:
UID, Name, Mobile, email,
Operation:
Registration()
Login()
Upload(Resource) [Resource – Image/Video]
ScanningVedio()
Splitting()
ImageFeatureExtraction()
PredictionAlgorithm()
2. Media Partners
Attribute:
UID, Name, Mobile, email,
Operation:
6
Registration()
Login()
Upload(Resource) [Resource – Image/Video]
ScanningVedio()
Splitting()
ImageFeatureExtraction()
PredictionAlgorithm()
3. Security Companies/Consultants
Attribute:
UID, Name, Mobile, email,
Operation:
Registration()
Login()
Upload(Resource) [Resource – Image/Video]
ScanningVedio()
Splitting()
ImageFeatureExtraction()
PredictionAlgorithm()
7
The major features of this technology is as shown in below entity–relationship model (ER
model)
8
2.5 Constraints: One of the most important challenges facing the researchers is the lack of
high-quality dataset. It is well-known that deep learning models often require large
dataset for the training step in order to produce good results, which, regrettably, are not
freely accessible or need permission from social media providers. The rapid development
of deepfake GAN models can also bring a new challenge where unseen types of
generated images and video may not be discovered by the current deep learning models.
Software will not work without any web browser or basic internet connection
3. System Requirements
3.1 Functional requirements
a. Completely based on Client-Server Architecture
b. Functioning should be completely on the basis of the registered user
i. Admin Users
ii. LEA’s Authorities
iii. Intelligence Agencies
iv. Media Partners
v. Independent Security Consultants
c. This system uses ML And AI algorithms for getting attributes from videos for
training and prediction
CLIENT/SERVER SYSTEM
Some sites are client sites and others are server sites. - All the data resides at the
server sites.
Windows.
A browser which supports CGI, HTML & Javascript.
8 MB RAM
9
128 Gb hard free drive space
Graphics processing unit /Cuda Capable system
Database MySQL
10
mainly focuses on image processing, video capture and
detection.
TensorFlow programming.
Visual Studio build the wheel for dlib and run Cmake.
5. Non-Functional Requirements
5.1 Performance requirements
The system should respond to a user query and return a results within a minute.
11
Feature extraction must be done in seconds.
Time taken by ML algorithms should be in seconds.
System should be able to determine whether a given image is either real or a deepfake.
If we face any extensive damage to avoid proportion of our services due to technical glitches and
failures we will be following a complete disaster recovery processes and policies.
Security system needs database storage just like many other applications, however, the special
requirements of the security market mean that vendor must choose their database partner
carefully. Privacy and data protection regulations that need to be adhered to while designing of
the product.
4. Portability:- This system must be portable to any operating system version and internet
connection
This system should not need any software installation.
12