Accurate Traffic Prediction 4.4
Accurate Traffic Prediction 4.4
Accurate Traffic Prediction 4.4
MODEL
Submitted in Partial fulfilment of the requirements for the award of the degree of
BACHELOR OF TECHNOLOGY
IN
2022-2023
Department Of Computer Science and Engineering
SRIDEVI WOMEN’S ENGINEERING COLLEGE
(Approved by AICTE and Govt. of TS | Affiliated to JNTUH | An ISO 9001:2015 Certified
Institution | An UGC Autonomous Institute | Accredited by MBA and NAAC with A++ Grade)
CERTIFICATE
This is to certify that the MINOR PROJECT report entitled “Accurate Traffic
Prediction using Ensemble Model” is being submitted by A.Harshita Krishnasree
(19D21A05H5), D.Lahari (19D21A05I3), I.Monika(19D21A05I9) in partial fulfilment for
the award of the Degree of Bachelor of Technology in Computer Science and Engineering
is a record of bonafied work carried out by them.
EXTERNAL EXAMINER
ii
DECLARATION
We hereby declare that the Mini Project entitled “Accurate Traffic Prediction using
Ensemble Model” is the work done during the period from 29-08-2022 to 15-12-2022 and
the same is submitted in partial fulfilment of the requirements for the award of degree of
Bachelor of Technology in Computer Science and Engineering from Jawaharlal Nehru
Technological University, Hyderabad.
iii
ACKNOWLEDGEMENT
We would like to express our sincere gratitude and indebtedness to our internal guide
Mr. K. RAJASEKHAR RAO, Assistant Professor, Department of Computer Science and
Engineering for his valuable guidance, suggestions, and keen personal interest throughout the
course of this project and for his tireless patience in hearing our seminar, minutely seeing all
the reports and giving appropriate guidance and suggestions.
We would like to express our sincere gratitude and indebtedness to our project co-
ordinator Dr. M. RAMASUBRAMANIAN, Professor, Department of Computer Science
and Engineering for his timely cooperation and valuable suggestions throughout the project.
We are indebted to him for the support given to us throughout the project work.
We are also extremely thankful to our principal Dr. B. L. MALLESWARI for her
precious guidance and valuable suggestions.
Finally, we would like to thank all our faculty and friends for their support throughout this
work. We are very much indebted to our parents for their moral support and encouragement
to achieve goals.
iv
ABSTRACT
The goal of this study is to create a mechanism for forecasting precise and timely
traffic flow data. Traffic Environment refers to everything that might have an impact on how
much traffic is moving down the road, including traffic signals, accidents, protests, and even
road repairs that might result in a backup. A motorist or rider can make an informed choice if
they have previous knowledge that is very close to approximate about all the above and many
more real-world circumstances that can affect traffic. Additionally, it aids in the development
of driverless vehicles. Traffic data have been multiplying tremendously in recent decades,
and big data concepts for transportation have been popular. The current approaches for
predicting traffic flow use some traffic prediction models, however they are still inadequate
to handle practical situations. This fact motivated us to focus on the problem of predicting
traffic flow based on traffic data and models.
v
TABLE OF CONTENTS
vii
LIST OF FIGURES
viii
LIST OF TABLES
ix
CHAPTER 1
INTRODUCTION
1.1PURPOSE
The network occasionally experienced a lot of issues, much like an urban region. The
enlargement of the roads and lanes is not possible on this piece of land. The second strategy
makes effective use of the current road network by utilising some control mechanisms. The
cost also decreases when these control tactics are used, making them cost-effective models
for the government or traffic controllers. The tactics used in this control point out probable
traffic jams and advise travellers to choose alternate routes to their destinations.
1.2 SCOPE
Traffic Environment refers to everything that might have an impact on how much
traffic is moving down the road, including traffic signals, accidents, protests, and even road
repairs that might result in a backup. A motorist or rider can make an informed decision if
they are prepared with prior knowledge that is very close to accurate about all the
aforementioned factors and many more real-world circumstances that can affect traffic.
1
CHAPTER 2
LITERATURE SURVEY
Parallel control and management have been proposed as a new mechanism for conducting
operations of complex systems, especially those that involved complexity issues of both
engineering and social dimensions, such as transportation systems. This paper presents an
overview of the background, concepts, basic methods, major issues, and current applications
of Parallel transportation Management Systems (PtMS). In essence, parallel control and
management is a data-driven approach for modeling, analysis, and decision-making that
considers both the engineering and social complexity in its processes. The developments and
applications described here clearly indicate that PtMS is effective for use in networked
complex traffic systems and is closely related to emerging technologies in cloud computing,
social computing, and cyberphysical-social systems. A description of PtMS system
architectures, processes, and components, including OTSt, Dyna CAS, aDAPTS, iTOP, and
TransWorld is presented and discussed. Finally, the experiments and examples of real-world
applications are illustrated and analyzed.
This study presents a framework for highway incident detection using vehicle kinetics,
such as speed profile and lane changing behaviour, as envisioned in the vehicle-infrastructure
integration (VII, also known as IntelliDrive) in which vehicles and infrastructure
communicate to improve mobility and safety. This framework uses an in-vehicle intelligent
module, based on a support vector machine (SVM), to determine the vehicle's travel
experiences with autonomously generated kinetics data.
2
Roadside infrastructure agents (also known as RSUs: roadside units) detect the
incident by compiling travel experiences from several vehicles and comparing the aggregated
results with the pre-selected threshold values. The authors developed this VII-SVM incident
detection system on a previously calibrated and validated simulation network in rural
Spartanburg, South Carolina and deployed it on an urban freeway network in Baltimore,
Maryland to evaluate its transportability. The study found no significant differences in the
detection performance between the original network and a new network that the VII-SVM
system has not seen before. This demonstrated the feasibility of developing a generic VII-
SVM system, applicable across transportation networks.
Advanced vehicle guidance systems use real-time traffic information to route traffic
and to avoid congestion. Unfortunately, these systems can only react upon the presence of
traffic jams and not to prevent the creation of unnecessary congestion. Anticipatory vehicle
routing is promising in that respect, because this approach allows directing vehicle routing by
accounting for traffic forecast information.
This paper presents a decentralized approach for anticipatory vehicle routing that is
particularly useful in large-scale dynamic environments. The approach is based on delegate
multiagent systems, i.e., an environment-centric coordination mechanism that is, in part,
inspired by ant behavior. Antlike agents explore the environment on behalf of vehicles and
detect a congestion forecast, allowing vehicles to reroute. The approach is explained in depth
and is evaluated by comparison with three alternative routing strategies. The experiments are
done in simulation of a real-world traffic environment. The experiments indicate a
considerable performance gain compared with the most advanced strategy under test, i.e., a
traffic-message-channel-based routing strategy.
3
4. Z. Zhao, W. Chen, X. Wu, P. C. Y. Chen, and J. Liu, Lstm network: a
deep learning approach for short-term traffic forecast.
4
CHAPTER 3
SYSTEM ANALYSIS
Traffic data have been growing dramatically in the recent decades, and we are moving
toward big data concepts for transportation. The current approaches for predicting traffic
flow use some traffic prediction models, however they are still inadequate to handle practical
situations. We began working on the traffic flow forecast problem using the traffic data and
models as a result of this fact. Since there is such a vast amount of data available for the
transportation system, it is difficult to anticipate the traffic flow accurately.
• The network occasionally experienced a lot of issues, much like an urban region.
• The expansion of the roads and lanes is not possible on this land facility.
• The primary benefit of it is to ensure the secure and efficient flow of road transportation.
• In terms of environmental friendliness, lowering carbon emissions is also beneficial.
• It offers the car industry numerous options to improve the safety and security of its
customers.
5
3.3 SRS SOFTWARE REQUIREMENT SPECIFICATIONS
Systems Development Life Cycle
A software cycle deals with various parts and phases from planning to testing and
deploying software. All these activities are carried out in different ways, as per the needs.
Each way is known as a Software Development Lifecycle Model (SDLC).
SDLC models * The Linear model (Waterfall) - Separate and distinct phases of specification
and development. - All activities in linear fashion. - Next phase starts only when first one is
complete. * Evolutionary development - Specification and development are interleaved
(Spiral, incremental, prototype based, Rapid Application development). - Incremental Model
(Waterfall in iteration), - RAD(Rapid Application Development) - Focus is on developing
quality product in less time, - Spiral Model - We start from smaller module and keeps on
building it like a spiral. It is also called Component based development. * Formal systems
development - A mathematical system model is formally transformed to an implementation.
* Agile Methods. - Inducing flexibility into development. * Reuse-based development - The
system is assembled from existing components.
SDLC Methodology:
Spiral Model
The spiral model is similar to the incremental model, with more emphases placed on
risk analysis. The spiral model has four phases: Planning, Risk Analysis, Engineering and
Evaluation. A software project repeatedly passes through these phases in iterations (called
Spirals in this model). The baseline spiral, starting in the planning phase, requirements is
gathered and risk is assessed. Each subsequent spirals builds on the baseline spiral.
Requirements are gathered during the planning phase. In the risk analysis phase, a process is
undertaken to identify risk and alternate solutions. A prototype is produced at the end of the
risk analysis phase. Software is produced in the engineering phase, along with testing at
the end of the phase. The evaluation phase allows the customer to evaluate the output of the
project to date before the project continues to the next spiral. In the spiral model, the angular
component represents progress, and the radius of the spiral represents cost. Spiral Life Cycle
Model.
6
CHAPTER 4
7
4.3 FUNCTIONAL REQUIREMENTS
1. Upload Traffic Dataset
2. Data Preprocessing
3. Build RF,DT & SVM Classifiers
4. Upload Test Data
5. Predict Traffic Result
• Usability
Usability is the main non-functional requirement for a Traffic Prediction for Intelligent
Transportation System using Machine Learning. The UI should be simple enough for
everyone to understand and get the relevant information without any special training.
Different languages can be provided based on the requirements.
• Accuracy
Accuracy is another important non-functional requirement for the Traffic Prediction for
Intelligent Transportation System using Machine Learning. The dataset is used to Train and
Test Model in python .Prediction should be correct, consistent, and reliable.
• Availability
The System should be available for the duration when the user operates and must be
recovered within an hour or less if it fails. The system should respond to the requests within
two seconds or less.
• Maintainability
The software should be easily maintainable and adding new features and making
changes to the software must be as simple as possible. In addition to this, the software must
also be portable.
8
4.5 SYSTEM STUDY
• FEASIBILITY STUDY
The feasibility of the project is analyzed in this phase and business proposal is put
forth with a very general plan for the project and some cost estimates. During system analysis
the feasibility study of the proposed system is to be carried out. This is to ensure that the
proposed system is not a burden to the company. For feasibility analysis, some
understanding of the major requirements for the system is essential.
1. Economical Feasibility
2. Technical Feasibility
3. Social Feasibility
• ECONOMICAL FEASIBILITY
This study is carried out to check the economic impact that the system will have on
the organization. The amount of fund that the company can pour into the research and
development of the system is limited. The expenditures must be justified. Thus the developed
system as well within the budget and this was achieved because most of the technologies
used are freely available. Only the customized products had to be purchased.
• TECHNICAL FEASIBILITY
This study is carried out to check the technical feasibility, that is, the technical
requirements of the system. Any system developed must not have a high demand on the
available technical resources. This will lead to high demands on the available technical
resources. This will lead to high demands being placed on the client. The developed system
must have a modest requirement, as only minimal or null changes are required for
implementing this system.
9
• SOCIAL FEASIBILITY
The aspect of study is to check the level of acceptance of the system by the user. This
includes the process of training the user to use the system efficiently. The user must not feel
threatened by the system, instead must accept it as a necessity. The level of acceptance by the
users solely depends on the methods that are employed to educate the user about the system
and to make him familiar with it. His level of confidence must be raised so that he is also able
to make some constructive criticism, which is welcomed, as he is the final user of the system.
10
CHAPTER 5
SYSTEM DESIGN
11
5.2 DATA FLOW DIAGRAM
1. The DFD is also called as bubble chart. It is a simple graphical formalism that can be used
to represent a system in terms of input data to the system, various processing carried out on
this data, and the output data is generated by this system.
2. The data flow diagram (DFD) is one of the most important modeling tools. It is used to
model the system components. These components are the system process, the data used by
the process, an external entity that interacts with the system and the information flows in the
system.
3. DFD shows how the information moves through the system and how it is modified by a
series of transformations. It is a graphical technique that depicts information flow and the
transformations that are applied as data moves from input to output.
4. DFD is also known as bubble chart. A DFD may be used to represent a system at any level
of abstraction. DFD may be partitioned into levels that represent increasing information flow
and functional detail.
12
5.3 UML DIAGRAMS
UML stands for Unified Modeling Language. UML is a standardized general-purpose
modeling language in the field of object-oriented software engineering. The standard is
managed, and was created by, the Object Management Group.
The goal is for UML to become a common language for creating models of object
oriented computer software. In its current form UML is comprised of two major components:
a Meta-model and a notation. In the future, some form of method or process may also be
added to; or associated with, UML.
The Unified Modeling Language is a standard language for specifying, Visualization,
Constructing and documenting the artifacts of software system, as well as for business
modeling and other non-software systems.
The UML represents a collection of best engineering practices that have proven
successful in the modeling of large and complex systems.
The UML is a very important part of developing objects oriented software and the
software development process. The UML uses mostly graphical notations to express the
design of software projects.
GOALS:
The Primary goals in the design of the UML are as follows:
1. Provide users a ready-to-use, expressive visual modeling Language so that they can
develop and exchange meaningful models.
2. Provide extendibility and specialization mechanisms to extend the core concepts.
3. Be independent of particular programming languages and development process.
4. Provide a formal basis for understanding the modeling language.
5. Encourage the growth of OO tools market.
6. Support higher level development concepts such as collaborations, frameworks, patterns
and components.
7. Integrate best practices.
13
5.3.1 Use case diagram
A use case diagram in the Unified Modeling Language (UML) is a type of behavioral
diagram defined by and created from a Use-case analysis. Its purpose is to present a graphical
overview of the functionality provided by a system in terms of actors, their goals (represented
as use cases), and any dependencies between those use cases. The main purpose of a use case
diagram is to show what system functions are performed for which actor. Roles of the actors
in the system can be depicted.
14
5.3.2 State diagram
A state diagram, as the name suggests, represents the different states that objects in
the system undergo during their life cycle. Objects in the system change states in response to
events. In addition to this, a state diagram also captures the transition of the object's state
from an initial state to a final state in response to events affecting the system.
15
5.3.3 Activity diagram
The process flows in the system are captured in the activity diagram. Similar to a state
diagram, an activity diagram also consists of activities, actions, transitions, initial and final
states, and guard conditions.
16
5.3.4 Sequence diagram
A sequence diagram represents the interaction between different objects in the system.
The important aspect of a sequence diagram is that it is time-ordered. This means that the
exact sequence of the interactions between the objects is represented step by step. Different
objects in the sequence diagram interact with each other by passing "messages".
17
5.3.5 Collaboration diagram
18
5.3.6 Component diagram
The component diagram represents the high-level parts that make up the system. This
diagram depicts, at a high level, what components form part of the system and how they are
interrelated. A component diagram depicts the components culled after the system has
undergone the development or construction phase.
The deployment diagram captures the configuration of the runtime elements of the
application. This diagram is by far most useful when a system is built and ready to be
deployed.
19
5.4 MODULES DESCRIPTION
• Data Collection
The dataset used in this paper is from kaggle site . This step was done by the original
owners of the dataset. And the composition of the dataset understand the relationship among
different features. A plot of the core features and the entire dataset. The dataset is further split
into 2/3 for training and 1/3 for testing the algorithms. Furthermore, in order to obtain a
representative sample, each class in the full dataset is represented in about the right
proportion in both the training and testing datasets. The various proportions of the training
and testing datasets used in the paper.
• Data Preprocessing
The data which was collected might contain missing values that may lead to
inconsistency. To gain better results data need to be preprocessed so as to improve the
efficiency of the algorithm. The outliers have to be removed and also variable conversion
need to be done. In order to overcoming these issues we use map function.
• Model Selection
Machine learning is about predicting and recognizing patterns and generate suitable
results after understanding them. ML algorithms study patterns in data and learn from them.
An ML model will learn and improve on each attempt. To gauge the effectiveness of a
model, it’s vital to split the data into training and test sets first. So before training our models,
we split the data into Training set which was 70% of the whole dataset and Test set which
was the remaining 30%. Then it was important to implement a selection of performance
metrics to the predictions made by our model.
20
CHAPTER 6
IMPLEMENTATION
6.1 ALGORITHM
This is an equation that involves calculating the inner products of a new input vector
(x) with all support vectors in training data. The coefficients B0 and ai (for each input) must
be estimated from the training data by the learning algorithm.
21
Fig 6.1.1: Support Vector Machine
Deep Learning deals with training multi-layer artificial neural networks, also called
Deep Neural Networks. After Rosenblatt perceptron was developed in the 1950s, there was a
lack of interest in neural networks until 1986, when Dr.Hinton and his colleagues developed
the backpropagation algorithm to train a multilayer neural network. Today it is a hot topic
with many leading firms like Google, Facebook, and Microsoft which invest heavily in
applications using deep neural networks. A fully connected multi-layer neural network is
called a Multilayer Perceptron (MLP).
22
Fig 6.1.2: Multi-layer ANN
It has 3 layers including one hidden layer. If it has more than 1 hidden layer, it is
called a deep ANN. An MLP is a typical example of a feedforward artificial neural network.
In this figure, the ith activation unit in the lth layer is denoted as ai(l). The number of layers
and the number of neurons are referred to as hyper parameters of a neural network, and these
need tuning. Cross-validation techniques must be used to find ideal values for these. The
weight adjustment training is done via back propagation. Deeper neural networks are better at
processing data. However, deeper layers can lead to vanishing gradient problems. Special
algorithms are required to solve this issue.
23
These rules can be used to perform predictions. When we have our dataset
categorized into 3 category so now Random forest helps to make classes from the dataset.
Random forest is clusters of decision trees all together, if you input a training dataset with
features and labels into a decision tree, it will formulate some set of rules, which will be used
to make the predictions.
24
6.1.5 BOOSTING
Boosting is an ensemble modelling technique that was first presented by Freund and
Schapire in the year 1997, since then, Boosting has been a prevalent technique for tackling
binary classification problems. These algorithms improve the prediction power by converting
a number of weak learners to strong learners.
The principle behind boosting algorithms is first we built a model on the training
dataset, then a second model is built to rectify the errors present in the first model. This
procedure is continued until and unless the errors are minimized, and the dataset is predicted
correctly. Let’s take an example to understand this, suppose you built a decision tree
algorithm on the Titanic dataset and from there you get an accuracy of 80%. After this, you
apply a different algorithm and check the accuracy and it comes out to be 75% for KNN and
70% for Linear Regression. We see the accuracy differs when we built a different model on
the same dataset. But what if we use combinations of all these algorithms for making the
final prediction? We’ll get more accurate results by taking the average of results from these
models. We can increase the prediction power in this way. Boosting algorithms works in a
similar way, it combines multiple models (weak learners) to reach the final output (strong
learners). In this article, we will understand the math behind different types of boosting
algorithms. There are mainly 3 types of boosting algorithms: AdaBoost algorithm Gradient
descent algorithm Xtreme gradient descent algorithm
Gradient boosting classifiers are a group of machine learning algorithms that combine
many weak learning models together to create a strong predictive model. Decision trees are
usually used when doing gradient boosting. Gradient boosting models are becoming popular
because of their effectiveness at classifying complex datasets, and have recently been used to
win many Kaggle data science competitions.
25
6.1.7 Logistic Regression:
The logistic regression is a predictive analysis. Logistic regression is used to describe data
and to explain the relationship between one dependent binary variable and one or more
nominal, ordinal, interval or ratio-level independent variables
6.2.1 PYTHON
Python is an interpreted high-level programming language for general-purpose
programming. Created by Guido van Rossum and first released in 1991, Python has a
design philosophy that emphasizes code readability, notably using significant whitespace.
Python features a dynamic type system and automatic memory management. It supports
multiple programming paradigms, including object-oriented, imperative, functional and
procedural, and has a large and comprehensive standard library.
Before we take a look at the details of various machine learning methods, let's start
by looking at what machine learning is, and what it isn't. Machine learning is often
categorized as a subfield of artificial intelligence, but I find that categorization can often be
misleading at first brush. The study of machine learning certainly arose from research in
this context, but in the data science application of machine learning methods, it's more
helpful to think of machine learning as a means of building models of data.
26
Fundamentally, machine learning involves building mathematical models to help
understand data. "Learning" enters the fray when we give these models tunable
parameters that can be adapted to observed data; in this way the program can be considered
to be "learning" from the data. Once these models have been fit to previously seen data,
they can be used to predict and understand aspects of newly observed data. I'll leave to the
reader the more philosophical digression regarding the extent to which this type of
mathematical, model-based "learning" is similar to the "learning" exhibited by the human
brain. Understanding the problem setting in machine learning is essential to using these
tools effectively, and so we will start with some broad categorizations of the types of
approaches we'll discuss here.
• Emotion analysis
• Object recognition
• Fraud detection
• Fraud prevention
27
6.2.3 Modules in python
• Tensorflow
TensorFlow is a free and open-source software library for dataflow and differentiable
programming across a range of tasks. It is a symbolic math library, and is also used
for machine learning applications such as neural networks. It is used for both research and
production at Google.
TensorFlow was developed by the Google Brain team for internal Google use. It was
released under the Apache 2.0 open-source license on November 9, 2015.
• Numpy
It is the fundamental package for scientific computing with Python. It contains various
features including these important ones:
• Pandas
28
• Matplotlib
• Scikit – learn
29
6.3 SAMPLE CODE
credentials_1 = {
'host':'xxx.yyy.com',
'port':'nnnn',
'username':'user',
'password':'password',
'database':'location',
'schema':'SMHEALTH'
}
def load_data_from_database(table_name):
return (
spark.read.format("jdbc").options(
driver = "com.ibm.db2.jcc.DB2Driver",
url = "jdbc:db2://" + credentials_1["host"] + ":" + credentials_1["port"] + "/" +
credentials_1["database"],
user = credentials_1["username"],
password = credentials_1["password"],
dbtable = credentials_1["schema"] + "." + table_name,
partitionColumn = "patientid",
lowerBound = 1,
upperBound = 5000,
numPartitions = 10
).load()
)
observations_df = load_data_from_database("OBSERVATIONS")
observations_df.show(5)
from pyspark.sql.functions import col
systolic_observations_df = (
observations_df.select("patientid", "dateofobservation", "numericvalue")
.withColumnRenamed("numericvalue", "systolic")
.filter((col("code") == "8480-6"))
)
systolic_observations_df.show(5)
diastolic_observations_df = (
observations_df.select("patientid", "dateofobservation", "numericvalue")
.withColumnRenamed('numericvalue', 'diastolic')
.filter((col("code") == "8462-4"))
)
hdl_observations_df = (
observations_df.select("patientid", "dateofobservation", "numericvalue")
.withColumnRenamed('numericvalue', 'hdl')
30
.filter((col("code") == "2085-9"))
)
ldl_observations_df = (
observations_df.select("patientid", "dateofobservation", "numericvalue")
.withColumnRenamed('numericvalue', 'ldl')
.filter((col("code") == "18262-6"))
)
bmi_observations_df = (
observations_df.select("patientid", "dateofobservation", "numericvalue")
.withColumnRenamed('numericvalue', 'bmi')
.filter((col("code") == "39156-5"))
)
merged_observations_df = (
systolic_observations_df.join(diastolic_observations_df, ["patientid",
"dateofobservation"])
.join(hdl_observations_df, ["patientid", "dateofobservation"])
.join(ldl_observations_df, ["patientid", "dateofobservation"])
.join(bmi_observations_df, ["patientid", "dateofobservation"])
)
merged_observations_df.show(5)
from pyspark.sql.functions import datediff
merged_observations_with_age_df = (
merged_observations_df.join(patients_df, "patientid")
.withColumn("age", datediff(col("dateofobservation"),
col("dateofbirth"))/365)
.drop("dateofbirth")
)
merged_observations_with_age_df.show(5)
31
CHAPTER 7
SYSTEM TESTING
Testing is the process where the test data is prepared and is used for testing the modules
individually and later the validation given for the fields. Then the system testing takes place
which makes sure that all components of the system property functions as a unit. The test
data should be chosen such that it passed through all possible condition. Actually testing is
the state of implementation which aimed at ensuring that the system works accurately and
efficiently before the actual operation commence. The following is the description of the
testing strategies, which were carried out during the testing period.
Testing has become an System integral part of any system or project especially in the field
of information technology. The importance of testing is a method of justifying, if one is
ready to move further, be it to be check if one is capable to with stand the rigors of a
particular situation cannot be underplayed and that is why testing before development is so
critical. When the software is developed before it is given to user to user the software must be
tested whether it is solving the purpose for which it is developed. This testing involves
various types through which one can ensure the software is reliable. The program was tested
logically and pattern of execution of the program for a set of data are repeated. Thus the
code was exhaustively checked for all possible correct data and the outcomes were also
checked.
To locate errors, each module is tested individually. This enables us to detect error and
correct it without affecting any other modules. Whenever the program is not satisfying the
required function, it must be corrected to get the required result. Thus all the modules are
individually tested from bottom up starting with the smallest and lowest modules and
proceeding to the next level. Each module in the system is tested.
32
For example the job classification module is tested separately. This module is tested with
different job and its approximate execution time and the result of the test is compared with
the results that are prepared manually. The comparison shows that the results proposed
system works efficiently than the existing system. Each module in the system is tested
separately. In this system the resource classification and job scheduling modules are tested
separately and their corresponding results are obtained which reduces the process waiting
time.
After the module testing, the integration testing is applied. When linking the modules
there may be chance for errors to occur, these errors are corrected by using this testing. In
this system all modules are connected and tested. The testing results are very correct. Thus
the mapping of jobs with resources is done correctly by the system.
When that user fined no major problems with its accuracy the system passers through a
final acceptance test.This test confirms that the system needs the original goals, objectives
and requirements established during analysis without actual execution which elimination
wastage of time and money acceptance tests on the shoulders of users and management, it is
finally acceptable and ready for the operation.
33
7.2. TEST CASES
Test Case 1 The user gives the input An output is Predict A result is Predict
in the form of Traffic Traffic Result. Traffic Result.
(Unit testing
of Dataset) Dataset.
Test Case 2 The user gives the input An output is Predict A result is Predict
in the form of Traffic Traffic Result. Traffic for test data
(Unit testing using SVM
of Accuracy) Dataset. algorithm got
accuracy up to 98%.
Test Case 3 The user gives the input An output is Predict A result is Predict
in the form of Traffic Traffic Result. Traffic for test data
(Unit testing using SVM
of Machine Dataset. algorithm got
Learning accuracy up to 96%.
Algorithms)
Test Case 4 The user gives the input An output is Predict A result is Predict
in the form of traffic test Traffic Result. Traffic for test data
(Integration using SVM
testing of data . algorithm got
Dataset) accuracy up to 98%.
Test Case 6 The user gives the input An output is Predict A result is predict
in the form of Traffic Traffic Result. Traffic Result.
(Data Flow
Testing) Dataset
34
Test Case 7 The user gives the input An output is Predict A result is predict
in the form of Traffic Traffic Result. Traffic Result.
(User
interface Dataset
Testing)
Test Case 9
The user uploads test An output is user A result is user
(User data in to application. upload test data successfully upload
interface Successfully. dataset in to
Testing-Event application.
based)
35
7.3 RESULTS AND DISCUSSIONS
36
Fig 7.3.3: Sign Up Page
37
Fig 7.3.5: Input page
38
CHAPTER 8
CONCLUSION AND FUTURE ENHANCEMENT
8.1 CONCLUSION
Although Machine Learning algorithms are used in data analysis, the ML community
has addressed them in any detail. The suggested approach improves the complexity issues
throughout the dataset and provides more accuracy than the currently used algorithms.
Also we have planned to integrate the web server and the application. Also the things
algorithms will be further improved to much more higher accuracy.
39
9. REFERENCES
[1] Fei-Yue Wang et al. Parallel control and management for intelligent transportation
systems: Concepts, architectures, and applications. IEEE Transactions on Intelligent
Transportation Systems, 2010.
[2] Yongchang Ma, Mashrur Chowdhury, Mansoureh Jeihani, and Ryan Fries. Accelerated
incident detection across transportation networks using vehicle kinetics and support vector
machine in cooperation with infrastructure agents. IET intelligent transport systems,
4(4):328–337, 2010.
[3] Rutger Claes, Tom Holvoet, and Danny Weyns. A decentralized approach for
anticipatory vehicle routing using delegate multiagent systems. IEEE Transactions on
Intelligent Transportation Systems, 12(2):364–373, 2011.
[4] Mehul Mahrishi and Sudha Morwal. Index point detection and semantic indexing of
videos - a comparative review. Advances in Intelligent Systems and Computing, Springer,
2020.
[6] H Qi, RL Cheu, and DH Lee. Freeway incident detection using kinematic data from probe
vehicles. In 9th World Congress on Intelligent Transport SystemsITS America, ITS Japan,
ERTICO (Intelligent Transport Systems and Services-Europe), 2002.
[7] Z. Zhao, W. Chen, X. Wu, P. C. Y. Chen, and J. Liu. Lstm network: a deep learning
approach for short-term traffic forecast. IET Intelligent Transport Systems, 11(2):68–75,
2017.
[8] C. Zhang, P. Patras, and H. Haddadi. Deep learning in mobile and wireless networking: A
survey. IEEE Communications Surveys Tutorials, 21(3):2224–2287, thirdquarter 2019.
40
[9] Chun-Hsin Wu, Jan-Ming Ho, and D. T. Lee. Travel-time prediction with support vector
regression. IEEE Transactions on Intelligent Transportation Systems, 5(4):276–281, Dec
2004.
[10] Yan-Yan Song and LU Ying. Decision tree methods: applications for classification and
prediction. Shanghai archives of psychiatry, 27(2):130, 2015.
[11] Yiming He, Mashrur Chowdhury, Yongchang Ma, and Pierluigi Pisu. Merging mobility
and energy vision with hybrid electric vehicles and vehicle infrastructure integration. Energy
Policy, 41:599–609, 2012.
[12] Jason Brownlee. Bagging and random forest ensemble algorithms for machine learning.
Machine Learning Algorithms, pages 4–22, 2016.
41