Engineering and Technology Journal: Hadeel N. Abdullah, Nuha H. Abdulghafoor
Engineering and Technology Journal: Hadeel N. Abdullah, Nuha H. Abdulghafoor
Engineering and Technology Journal: Hadeel N. Abdullah, Nuha H. Abdulghafoor
KEYWORDS ABSTRACT
Object Detection, Object Object detection and tracking are key mission in computer visibility
Tracking, Fast Principle applications, including civil or military surveillance systems. However,
Component Purist there are major challenges that have an effective role in the accuracy of
(FPCP), Blob Analysis, detection and tracking such as the ability of the system to track the target
Kalman Filter and the response speed of the system in different environments as well as
the presence of noise in the captured video sequence. In this proposed
work, a new algorithm to detect moving objects from video data is
designed by the Fast Principle Component Purist (FPCP). Then, we used
an ideal filter that performs well to reduce noise through the
morphological filter. The Blob analysis is used to add smoothness to the
spatial identification of objects and their areas, and finally, the detected
object is tracked by Kalman Filter. The applied examples demonstrated
the efficiency and capability of the proposed system for noise removal,
detection accuracy and tracking.
How to cite this article: H. N. Abdullah and N. H. Abdulghafoor “Automatic objects detection and tracking
using FPCP, Blob analysis and kalman filter,” Engineering and Technology Journal, Vol. 38, No. 02, pp.
246-254, 2020.
DOI: https://doi.org/10.30684/etj.v38i2A.314
1. Introduction
The process of tracking is one of the main tasks in computer visibility [1]. It has an important turn in
many areas of research, such as movement guessing, recognition and analysis of human and
nonhuman Vitality, 3D representation, mobility in vehicles, and others. Object tracking is the most
common attribute in automated monitoring applications because the individual human employer
cannot manage the controlled area, especially when the number of cameras rises. In addition, in the
medicinal application, the operator cannot sometimes analyze the video taken by the device;
especially in crucial cases, the detection and tracking system is more efficient than the human. It is
also used in anti-theft systems, traffic management systems, and others. The tracking system can
track single or multiple animation objects in different environments. In general, the object detection
246
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
and tracking system include the different stages like background subtraction, object detection, and
object tracking, as shown in the block diagram in Figure 1.
The first basic step in many fields of image processing and computer vision is the background
subtraction, also known as Foreground Detection, which extracts the foreground of the image for
subsequent processing (such as object selection and identification). These are the most important
areas of the picture, which are called objects such as humans, cars, texts, etc. This stage maybe after
the pre-processing phase of the image, which may include noise reduction of images, and before the
subsequent treatment phase such as morphology, etc. A more common way to detect moving objects
in videos is background subtraction. The basic principle is to detect moving objects from the
difference between the current frame and the reference frame, often called a "background-image" or
"background model". Some common methods in this area include the use of frame differentiation,
optical flow, analysis of principal components, and background mixture models [2].
Object modelling represents the object of interest in a scene. To represent an object, features are
extracted that uniquely defines an object. These features or the descriptors of an object are then used
to track the object. A feature is an image pattern that differentiates an object from its neighbourhood.
The features of an object are converted into descriptors, also referred to as appearance features, using
some operations around the features [1]. The commonly used object representations for tracking are
centroid, multiple points, rectangular patch, and complete object contour, etc. while the descriptors of
an object such as probability densities of object appearance (Histogram), template, blob analysis, etc.
The object tracking is to select and give individual paths to each object in the video sequence.
Objects can be humans on the street, cars on the road, players on the pitch, or from a group of
animals. The object is tracked to extract the object, identify the object and track it, and the decisions
related to their activities. Trace objects can be classified as points tracking, kernel tracking, and trace
shadow images. The general techniques of tracking such as Kalman Filter, Particle Filter, Mean Shift
Method, etc.
The main contribution of this work is as follows:
1- The proposed algorithm uses the FPCP technique to extract the motion areas from different
backgrounds of the captured video frame without the need for further input. As a result, so the
outputs have good speed and accuracy.
247
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
2- Using the method of analyzing the blob spatially simultaneously to select the effective pixels in
the motion zones and to determine the area of the object at the same time. In addition to using
efficient tracking technique, Kalman's Filter is thus an efficient and integrated way to track multiple
objects in the same captured video frame.
This paper is ordered as follows: Section 2 explains the related works, Section 3 explains
methodologies (Mathematical Background), Section 4 explains the proposed algorithm, Section 5
explains the results and discussion. Finally the conclusions in Section 6.
2. Related Works
Since the last few decades, many researchers have proven algorithms for detecting and tracking
objects. In this section, we demonstrated some of these algorithms related to the proposed system.
According to [3], motion brim is extracted in polar-log coordinate; then the gradient operator is
employed to compute the optical flow directly in every motion regions. Finally, the object is tracked.
In the proposed work in [4], the active background is reconstructed and the object size is determined
as a preliminary task, to extract and track the object in the foreground. The method in [5] is the
object detection is done by Gaussian Mixture Model (GMM), and Kalman Filter does the tracking. In
this method, Object detection is determined based on the size of the foreground. Therefore Errors
will occur in determining the object such as the object and its shadow are merged as an object or
representing two adjacent compounds as a single object. The paper [6] developed; the algorithm
includes optical flow and the motion vector estimation for object detection and tracking. The
detection and tracking system in [7] is sophisticated depend on optical flow for detection; the object
tracking is done by blob analysis.
In Prabhakar et al. [8], a moving object tracking system using morphological processing and blob
analysis, which able to distinguish between car and pedestrian in the same video. In the paper [9], the
foreground is extracted from the background using multiple-view architecture. After that, the forward
movement date and editing schemes are used to detect the animated objects. Finally, by detecting the
center of gravity of the moving object, it is used to trace the object based on the Kalman Filter.
In the method [10], animated objects are represented as groups of spatial and temporal points using
the Gabor 3D filter, which works on the spatial and temporal analysis of the sequential video and is
then joint by using the Minimum Spanning Tree. The proposed technique described in [11], split into
three stages; Foreground segmentation stage by using Mixture of Adaptive Gaussian model, tracking
stage by using the blob detection and evaluation stage which includes the classification according to
the feature extraction.
After exploring some of the published research on the detection and tracking of the object, it was
found that the discovery and tracking of the object is a complex task because of many elements of
dynamic tracking such as determining the type of camera moving or static, the random. Change of
the speed of the object, the intensity of light and darkness, etc.
248
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
The normal process to solve eq. (3) by the substitutional minimization as follow:
Lk+1 = arg L min ||L + Sk − D ||F s.t. rank (L) ≈ t (4)
𝑆k+1 = arg S min ||𝐿𝑘+1 + S − D ||F + 𝜆 ||𝑆||1 (5)
The eq. (3) can be solved by taking a partial Singular Value Decomposition SVD of (D - Sk) with
respect to t. while the eq. 4 can be solved by element-wise shrinkage. The background of the videos
is supposed to lie in a low-level sub-space, and the moving objects should be in the foreground as if
they were gradually soft in the spatial and temporal direction. The proposed method integrates the
Frobenius and l1-norm base into a unified framework for simultaneous noise reduction and detection.
The Frobenius base uses the low-level property in the background; the contrast is improved by the l 1
norm standard [13].
The matrix P, which includes moving object information, is obtained through the detection process.
An integral part of the morphological expansion and erosion processes is a structural element of a flat
shape. There is a binary flat structure element with a living value, either 2-D or multi-dimensional, in
which the real pixels are included in the morphological calculation, and false pixels are not. The
middle pixel of the structure element, called the parent, determines the pixel in the image being
processed.
249
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
Kalman Filter has the ability to rating the tracking locations with minimal datum on the location of
the object. Initially, the status St and measurement Xt paradigm are determined to predict the next
site. The paradigm matrixes are defined as [5]:
Prediction
Ẍt = AẌt−1 + Bu (8)
St = ASt-1 AT + Q (9)
Where; A - state transition matrix, B - coverts control input and Q - process noise covariance.
Correction: The measurement update equations are given as:
Kt = St-1HT (HSt-1HT + R)−1 (10)
Ẍt+1 = Ẍt + Kt(Yt − HẌt) (11)
St+1 = (I − KtH)St (12)
Where K- Kalman gain, S- measurement matrix, R- measurement error covariance and H-model
matrices. The prediction of the next state St+1 is done by integrating the actual measurement with the
pre-estimate of the situation St-1.
250
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
11. Update the current status in the present frame to assign the new detection tracks or remove any
invisible or lost tracks.
12. Data association for the same detection object for the present frame.
13. Display the resulting frames.
14. End while.
5. Results and Discussion
The algorithm proposed in MATLAB (2018b) has been applied, and their experiments were
performed on a Computer type MSI GV63 with Intel Core i7 8750H, NVIDIA RTX 2060 6G,
256GB SSD+1TB and 16 GB RAM. It has three stages are foreground detection, filtering and
tracking. The proposed algorithm detects the movable objects accurately and keeps track of their
appearance in the sequence video frames. Video data has been used in any format as an input to the
proposed work, and good results have been obtained in various article conditions on this indoor,
outdoor, light traffic and dense traffic. The efficiency of the proposed algorithm was evaluated; the
experiential outcomes were as follows:
Figure 2(a) shows the original framework. In Figure 2(b) the foreground was extracted by the FPCP
detection and showed holes and noise on the frame. In order to clarify and soften the frame, the
morphology was done.
The foreground frame is shown after the morphological process in Figure 2(c) and the final output
frame as in Figure 2(d), which includes movable and tracked objects.
Several-sampled video is used in the various environment in order to test the performance of the
proposed algorithm. The experiential outcomes are given in Figure 3. The first column (Figure 3(a))
shows the sampled frames of the video then the second column (Figure 3(b)) shows a clean
foreground extracted frames by FPCP detection are given. The third column (Figure 3(c)) includes
the detected and tracked objects by marking it with a circumferential box.
Figure 2: A sampled result (a) original frame (b) Foreground before and (c) after applying the
morphological filter and (d) The final result
251
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
Figure 3: The proposed algorithm results on several-captured video, (a) Input Video, (b) the filtered
foreground and (c) the resultant video
Table 1 explains the mean execution times (sec) of the proposed system when implemented on 365
sampled video frames by Matlab version R2018b. The consuming time of evaluation reduces
drastically. Accuracy is a measure of the performance efficiency of the object tracking system. The
detection and tracking system precision can be calculated using the following formula:
The total number of detected objects by system
Accuracy = The total number of actual objects in video (13)
The proposed algorithm for the different video input has been tested with different methods to
evaluate its accuracy. The accuracy of the proposed tracker in different input scenes was compared
and compared with other tracking systems, as shown in Table 2. It shows that the detection and
tracking accuracy rate using the proposed algorithms is 100%. The results are optically acceptable
except for an algorithm t.
The proposed algorithm for the different video input has been tested with different methods to
evaluate its accuracy. The accuracy of the proposed tracker in different input scenes was compared
and compared with other tracking systems, as shown in Table 2. It shows that the detection and
tracking accuracy rate using the proposed algorithms is 100%. The results are optically
acceptable except for an algorithm that proves that this multi-object tracking method is
validated. It was concluded that the proposed algorithm was still competitive, although some
results were closest to another, with little degradation in some because of the cost of some
complicated calculations.
The detection precision of the suggested algorithm is compared with other known and present
methods. The comparison shows the efficient performance of the proposed method on some of the
selected frames shown in Figure 4. We compared the proposed algorithm with the most
representative algorithms and for different frame sizes and settings for the tested videos; we used
grey or chromatic video sequences. The results were comparable to the proposed algorithm with
other algorithms.
To evaluate the visual performance of the proposed algorithm, we compared the proposed algorithm
to 3 algorithms. The videos examined contain different background scenes, and multiple moving
objects both outdoors and indoors (pedestrians, vehicles, etc.). We have chosen the following most
methods to compare with our proposed method: (1) GMM method [5], (2) optical flow [7], (3)
MODT [9].
Visual results are shown on the videos tested in Figure 4. Individual and group infantry, small
dynamic background, and multiple traffic surveys as shown in Figure 4, the proposed algorithm is
closest to Ground Truth (GT). Some of the results of the tested algorithms consider the foreground
object as the background. The main reason is that the parts of the object remain static in the video
252
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
and that the proposed algorithm has overcome this effect, and obviously the detection effect is better
than other algorithms.
Figure 4: The comparison results for several experiment captured video: (a) Input Frame, (b) The
Ground Truth, (c) GMM Method [5], (d) Optical Flow [7], (e) MODT [9], and (f) The Proposed
Algorithm
6. Conclusions
Object detection and tracking are the main and affront mission in many computer visibility
implementations, such as monitoring, car salt works, routing, and automation. The proposed
algorithm consists of three stages; the first stage foreground detection and filtering of various types
of noise from images using FPCP technique, the second stage of the identification of animation
objects and their region by blob analysis method and finally, the Kalman Filter is used for tracking
the objects. This algorithm presents several benefits, such as multiple object detection and tracking in
different environments. The disadvantages of this technique using one method will not produce
perfect results because its accuracy is influenced by different operators such as the low resolution of
captured video, change in weather, etc. In the future, we hope to expand our scope of detection and
tracking of objects in overcrowded scenery or the appearance of severe contrast in lighting and real-
time scenes.
253
Engineering and Technology Journal Vol. 38, Part A, (2020), No. 02, Pages 246-254
References
[1] J. Cheng, J. Yang, Y. Zhou and Y. Cui, “Flexible background mixture models for foreground
segmentation,” Image and Vision Computing, Vol. 24, No. 5, pp.473-482, 2006.
[2] Y. Wu, J. Lim and M.H. Yang, “Online object tracking: A benchmark,” Proceedings of the IEEE
conference on computer vision and pattern recognition, pp. 2411-2418, 2013.
[3] H.Y. Zhang, “Multiple moving objects detection and tracking based on optical flow in a polar-log image,”
International Conference on Machine Learning and Cybernetics, IEEE, Vol. 3, pp. 1577-1582, 2010.
[4] N.A. Mandellos, I. Keramitsoglou, and C.T. Kiranoudis, “A background subtraction algorithm for detecting
and tracking vehicles,” Expert Systems with Applications, Vol. 38, No. 3, pp.1619-1631, 2011.
[5] R.Y. Bakti, I.S. Areni and A.A. Prayogi, “Vehicle detection and tracking using gaussian mixture model and
kalman filter,” International Conference on Computational Intelligence and Cybernetics, IEEE, pp. 115-119,
2016.
[6] K., Kale, S. Pawar and P. Dhulekar, “Moving object tracking using optical flow and motion vector
estimation.” 4th International Conference on Reliability, Infocom Technologies and Optimization (ICRITO)
(Trends and Future Directions), IEEE, pp. 1-6, 2015.
[7] S. Aslani and H. Mahdavi-Nasab, “Optical flow-based moving object detection and tracking for traffic
surveillance,” International Journal of Electrical, Electronics, Communication, Energy Science and
Engineering, Vol. 7, No. 9, pp.789-793, 2013.
[8] P. Telagarapu, M.N. Rao and G. Suresh, “A novel traffic-tracking system using morphological and Blob
analysis”. International Conference on Computing, Communication and Applications, IEEE, pp. 1-4, 2012.
[9] W.C. Hu, C.H. Chen, T.Y. Chen, D.Y. Huang and Z.C. Wu, “Moving object detection and tracking from
video captured by moving camera,” Journal of Visual Communication and Image Representation, 30, pp.164-
180, 2015.
[10] K.S. Ray and S. Chakraborty, “Object detection by spatiotemporal analysis and tracking of the detected
objects in a video with variable background,” Journal of Visual Communication and Image Representation, 58,
pp.662-674, 2019.
[11] T. Mahalingam and M. Subramoniam, “A robust single and multiple moving object detection, tracking and
classification”. Applied Computing and Informatics, 2018.
[12] T. Bouwmans and E.H. Zahzah, “Robust PCA via principal component pursuit: A review for comparative
evaluation in video surveillance,” Computer Vision and Image Understanding, 122, pp.22-34, 2014
[13] P. Rodriguez and B. Wohlberg, “Fast principal component pursuit via alternating minimization,” IEEE
International Conference on Image Processing, pp. 69-73, 2013.
[14] Q. Li, R. Li, K. Ji and W. Dai, “Kalman filter and its application,” 2015 8th International Conference on
Intelligent Networks and Intelligent Systems (ICINIS), IEEE, pp. 74-77, 2015.
254