urfd
urfd
urfd
Abstract
Since falls are a major public health problem in an ageing society, there is considerable demand for low-cost fall
detection systems. One of the main reasons for non-acceptance of the currently available solutions by seniors is
that the fall detectors using only inertial sensors generate too much false alarms. This means that some daily
activities are erroneously signaled as fall, which in turn leads to frustration of the users. In this paper we
present how to design and implement a low-cost system for reliable fall detection with very low false alarm
ratio. The detection of the fall is done on the basis of accelerometric data and depth maps. A tri-axial
accelerometer is used to indicate the potential fall as well as to indicate whether the person is in motion. If the
measured acceleration is higher than an assumed threshold value, the algorithm extracts the person, calculates
the features and then executes the SVM-based classifier to authenticate the fall alarm. It is a 365/7/24
embedded system permitting unobtrusive fall detection as well as preserving privacy of the user.
Keywords: Fall detection; Depth image analysis; Assistive technology; Sensor technology for smart homes
However, most of them discriminate poorly between data collection is no longer constrained to laboratory
activities of daily living (ADLs) and falls, and none environments. In fact, it is the only technology that
of which is universally accepted by elderly. One of the was successfully used in large scale collection of people
main reasons for non-acceptance of the currently avail- motion data.
able solutions by seniors is that the fall detectors using
only accelerometers generate too much false alarms. 1.2 Camera based approaches to fall detection
This means that some daily activities are erroneously Video-cameras have largely been used for detecting
signaled as fall, which in turn leads to frustration of falls on the basis of single CCD camera [15, 16], multi-
the users. ple cameras [17], specialized omni-directional ones [18]
The main reason of high false ratio of accelerometer- and stereo-pair cameras [19]. Video based solutions of-
based systems is the lack of adaptability together with fer several advantages over others including the ca-
insufficient capabilities of context understanding. In pability of detection of various activities. The further
order to reduce the number of false alarms, many at- benefit is low intrusiveness and the possibility of re-
tempts were undertaken to combine both accelerome- mote verification of fall events. However, the currently
ter and gyroscope [6, 12]. However, several ADLs like available solutions require time for installation, camera
quick sitting have similar kinematic motion patterns calibration and they are not cheap. As a rule, CCD-
with real falls and in consequence such methods might camera based systems require a PC computer or a
trigger many false alarms. As a result, it is not easy to notebook for image processing. While these techniques
distinguish real falls from fall-like activities using only might work well in controlled environments, in order
to be practically applied they must be adapted to non-
accelerometers and gyroscopes. Another drawback of
controlled environments in which neither the lighting
the approaches based on wearable sensors, from the
nor the subject tracking is fully controlled. Typically,
user’s perspective, is the need to wear and carry vari-
the existing video-based devices for fall detection can-
ous uncomfortable devices during normal daily life ac-
not work in nightlight or low light conditions. Addi-
tivities. In particular, the elderly may forget to wear
tionally, the lack of depth information can lead to lots
such devices. Moreover, in [13] it is pointed out that
of false alarms. What is more, their poor adherence to
the common fall detectors, which are usually attached
real-life applications is particularly related to privacy
to a belt around the hip, are inadequate to be worn
preserving. Nevertheless, these solutions are becom-
during the sleep and this results in the lack of ability of
ing more accessible, thanks to the emergence of low-
such detectors to monitor the critical phase of getting cost cameras, the wireless transmission devices, and
up from the bed. the possibility of embedding the algorithms. The ma-
In general, the solutions mentioned above are some- jor problem is acceptance of this technology by the se-
how intrusive for people as they require wearing con- niors as it requires the placement of video cameras in
tinuously at least one device or smart sensor. On the private living quarters, and especially in the bedroom
other hand, these systems, comprising various kinds and the bathroom.
of small sensors, transmission modules and processing The existing video-based devices for fall detecting
capabilities, promise to change the personal care, by cannot work in nightlight or low light conditions. In ad-
supplying low-cost wearable unobtrusive solutions for dition, in most of such solutions the privacy is not pre-
continuous all-day and any-place health and activity served adequately. On the other hand, video cameras
status monitoring. An example of such solutions with offer several advantages in fall detection over wear-
a great potential are smart watches and smartphone- able devices-based technology, among others the abil-
based technologies. For instance, in iFall application ity to detect and recognize various daily activities. Ad-
[14], data from the accelerometer is evaluated using ditional advantage is low intrusiveness and the possi-
several threshold-based algorithms and position data bility of remote verification of fall events. However,
to determine the person’s fall. If a fall is inferred, a no- the lack of depth information may lead to many false
tification is raised requiring the user’s response. If the alarms. The existing technology permits reaching quite
user does not respond, the system sends alerts message high performance of fall detection. However, as men-
via SMS. tioned above it does not meet the requirements of the
Despite several shortcomings of the currently avail- users with special needs.
able wearable devices, the discussed technology has a Recently, Kinect sensor has been proposed to achieve
great potential, particularly, in the context of grow- fall detection [20, 21, 22]. The Kinect is a revolutionary
ing capabilities of signal processors and embedded sys- motion-sensing technology that allows tracking a per-
tems. Moreover, owing to progress in this technology, son in real-time without having to carry sensors. It is
Kwolek and Kepski Page 3 of 14
the world’s first low-cost device that combines an RGB to short overview of the algorithm. A threshold-based
camera and a depth sensor. Thus, if only depth images detection of the person fall is described in Section 4.
are used it preserves the person’s privacy. Unlike 2D In Section 5 we give details about extraction of the
cameras, it allows tracking the body movements in 3D. features representing the person in depth images. The
Since the depth inference is done using an active light classifier responsible for detecting human falls is pre-
source, the depth maps are independent of external sented in Section 6. The experimental results are dis-
light conditions. Owing to using the infrared light, the cussed in Section 7. Section 8 provides some concluding
Kinect sensor is capable of extracting the depth maps remarks.
in dark rooms. In the context of reliable fall detection
systems, which should work 24 hours a day and 7 days 2 The embedded system for human fall
a week it is very important capability, as we already detection
demonstrated in [21].
This Section is devoted to presentation of the main
1.3 Overview of the method ingredients of the embedded system for human fall
In order to achieve reliable and unobtrusive fall de- detection. At the beginning, the architecture of the
tection, our system employs both the Kinect sensor embedded system for fall detection is outlined. Next,
and a wearable motion-sensing device. When both de- the PandaBoard is drafted briefly. Following that, the
vices are used our system can reliably distinguish be- wearable device is presented in detail. Then, the Kinect
tween falls and activities of daily living. In such a con- sensor and its usefulness for fall detection are discussed
figuration of the system the number of false alarms shortly. Finally, data processing, feature extraction
is diminished. The smaller number of false alarms is along with classification modules are discussed briefly
achieved owing to visual validation of the fall alert in the context of the limited computational resources
generated on the basis of motion data only. The au- of the utilized embedded platform.
thentication of the alert is done on the basis of depth
data and analysis of the features extracted on depth 2.1 Main ingredients of the embedded system
maps. Owing to the determined in advance parame- Our fall detection system uses both data from Kinect
ters describing the floor the system analyses not only and motion data from a wearable smart device contain-
the shape of the extracted person but also the distance ing accelerometer and gyroscope sensors. On the basis
between the person’s center of gravity and the floor. of data from the inertial sensor the algorithm extracts
In situations in which the use of the wearable sensor motion features, which are then used to decide if a fall
might not be comfortable, for instance during changing took place. In the case of the fall the features repre-
clothes, bathing, washing oneself, etc., the system can senting the person in the depth images are dispatched
detect falls using depth data only. In the areas of the to a classifier, see Fig. 1.
room being outside of the Kinect field of view the sys-
tem can operate using data from motion-sensing device 2.2 Embedded platform
consisting of an accelerometer and a gyroscope only. The computer used to execute depth image analysis
Thanks to automatic extraction of the floor no calibra- and signal processing is the PandaBoard ES, which
tion of the system is needed and Kinect can be placed is a mobile development platform, enabling software
according to the user preferences at the height of about developers access to an open OMAP4460 processor-
0.8 − 1.2 m. Owing to using of depth maps only our based development platform. It features a dual-core
system preserves privacy of people undergoing moni- 1 GHz ARM Cortex-A9 MPcore processor with Sym-
toring as well as it can work at nighttime. The price of metric Multiprocessing (SMP), a 304 MHz PowerVR
the system along with working costs are low thanks to SGX540 integrated 3D graphics accelerator, a pro-
the use of low-cost Kinect sensor and low-cost Pand- grammable C64x DSP, and 1 GB of DDR2 SDRAM.
aBoard ES, which is a low-power, single-board com- The board contains wired 10/100 Ethernet along with
puter development platform. The algorithms were de- wireless Ethernet and Bluetooth connectivity. The
veloped with respect to both computational demands PandaBoard ES can support various Linux-based op-
as well as real-time processing requirements. erating systems such as Android, Chrome and Linux
The rest of the paper is organized as follows. Sec- Ubuntu. The booting of the operating system is from
tion 2 gives an overview of the main ingredients of SD memory card. Linux is well suited operating sys-
the system, together with the main motivations for tem for real-time embedded platforms since it provides
choosing the embedded platform. Section 3 is devoted various flexible inter-process communication methods,
Kwolek and Kepski Page 4 of 14
among others message queues. Another advantage of also allow on-board storage of data for later analy-
using Linux in an embedded device is rich availability sis. The x-IMU consists of triple axis 16-bit gyroscope
of tools and therefore it has been chosen for managing and triple axis 12-bit accelerometer. The first sensor
the hardware and software of the selected embedded measures acceleration, the rate of change in velocity
platform. across time, whereas the gyroscope delivers us the rate
The data acquired by x-IMU inertial device with of change of the angular position over time (angular
256 Hz are transmitted wirelessly via Bluetooth to the velocity) with a unit of [deg/s]. The acceleration is
processing device, whereas the Kinect sensor was con- measured in units of [g].
nected to the device via USB, see Fig. 1. The fall detec- The measured acceleration components were median
tion system runs under Linux operating system. The filtered with a window length of three samples to sup-
application consists of five main concurrent processes press the sensor noise. The accelerometric data were
that communicate via message queues, see Fig. 1. Mes- utilized to calculate the acceleration’s vector length.
sage queues are appropriate choice for well structured Figure 2 shows a sample plot of acceleration vector
data and therefore they were selected as a communi- length vs. time for a person walking up and down the
cation mechanism between the concurrent processes. stairs, and after that sitting down. The plot depicts
They provide asynchronous communication that is motion data of a person older than 65 years of age.
managed by Linux kernel. The first process is respon- As illustrated on the discussed plot, for typical daily
sible for acquiring data from the wearable device, the activities of an elderly the acceleration assumes quite
second one acquires depth data from the Kinect, third considerable values. As we can observe, during a rapid
process continuously updates the depth reference im- sitting down the acceleration value equal to 3.5 has
age, fourth one is responsible for data processing and been exceeded. Such a value is assumed very often as
feature extraction, whereas the fifth process is ac- a decision threshold in simple threshold-based algo-
countable for data classification and triggering the fall rithms for fall detection [10, 8]. Therefore, in order
alarm. The extraction of the person on the basis of to reduce the number of false alarms, in addition to
the depth reference maps has been chosen since the the measurements from the inertial sensor we employ
segmentation can be done with relatively low compu- the Kinect sensor whenever it is only possible. The
tational costs. The dual-core processor allows parallel depicted plots were obtained for the IMU device that
execution of processes responsible for the data acqui- was worn near the pelvis region. It is worth noting that
sition and processing. the attachment of the wearable sensor near the pelvis
region or lower back is recommended in the literature
2.3 The inertial device [11] because such a body part represents the major
The person movement is sensed by an x-IMU [23], component of body mass and undergoes movement in
which is a versatile motion sensing platform. Its host most activities.
of on-board sensors, algorithms, configurable auxil-
iary port and real-time communication via USB, Blue- 2.4 Kinect sensor
tooth or UART make it powerful smart motion sens- The Kinect sensor simultaneously captures depth and
ing sensor. The on-board SD card, USB-based battery color images at a frame rate of about 30 fps. The
charger, real-time clock and motion trigger wake up device consists of an infrared laser-based IR emitter,
Kwolek and Kepski Page 5 of 14
Figure 5 Flow chart of the algorithm for fall detection. where Ax (t), Ay (t), Az (t) is the acceleration in the x−,
y−, and z−axes at time t, respectively. The SVT otal
Kwolek and Kepski Page 7 of 14
Figure 6 Acceleration (top row) and angular velocity (bottom row) over time for walking downstairs and upstairs, picking up an
object, sitting down - standing up and falling.
contains both the dynamic and static acceleration 5.1 Extraction of the object of interest in depth maps
components, and thus it is equal to 1 g for standing, In order to make the system applicable in a wide
see plots of acceleration change curves in upper row range of scenarios we elaborated a fast method for up-
of Fig. 6. As we can observe on the discussed plots, dating the depth reference image. The person was de-
during the process of falling the acceleration attained tected on the basis of a scene reference image, which
the value of 6 g, whereas during walking downstairs was extracted in advance and then updated on-line.
and upstairs it attained the value of 2.7 g. It is worth In the depth reference image each pixel assumes the
noting that the data were acquired by x-IMU, which median value of several pixels values from the past im-
was worn by a middle aged person (60+). The plots ages, see Fig. 7. In the set-up stage we collect a number
shown in the bottom row illustrate the corresponding of the depth images, and for each pixel we assemble a
change of angular velocities. As we can see, the change list of the pixel values from the former images, which is
of the angular velocities during the process of falling then sorted in order to extract the median. Given the
is the most significant in comparison to non-fall activ- sorted lists of pixels the depth reference image can be
ities. However, in practice, it is not easy to construct a updated quickly by removing the oldest pixels and up-
reliable fall detector with almost null false alarms ratio dating the sorted lists with the pixels from the current
using the inertial data only. Thus, our system employs depth image and then extracting the median value. We
a simple threshold-based detection of falls, which are found that for typical human motions, satisfactory re-
then verified on the basis of analysis of the depth im- sults can be obtained using 13 depth images. For the
ages. If the value of SVT otal is greater than 3 g then Kinect acquiring the images at 30 Hz we take every
the system starts the extraction of the person and then fifteenth image.
executes the classifier responsible for the final decision
The images shown in the 3rd row of Figure 8 are the
about the fall, see also Fig. 5.
binary images with the foreground objects, which were
obtained using the discussed technique. In the middle
5 Extraction of the features representing row there are the raw depth images, whereas in the
person in depth images upper one there are the corresponding RGB images.
In this Section we demonstrate how the features rep- The RGB images are not processed by our system and
resenting the person undergoing monitoring are ex- they are only depicted for illustrative purposes. In the
tracted. At the beginning we discuss the algorithm for image #410 the person closed the door, which then
person delineation in the depth images. Then, we ex- appears on the binary image being a difference map
plain how to automatically estimate the parameters of between the current depth image and the depth ref-
the equation describing the floor. Finally, we discuss erence image. As we can see, in frame 610, owing to
the features representing the lying person, given the adaptation of the depth reference image, the door dis-
extracted equation of the floor. appears on the binary image and the person under-
Kwolek and Kepski Page 8 of 14
going monitoring is properly separated from the back- In the detection mode the foreground objects are ex-
ground. Having on regard that the images are acquired tracted through differencing the current image from
with 25 frames per second as well as the number of such a reference depth map. Subsequently, the fore-
frames that were needed to update of the depth refer- ground object is determined through extracting the
ence image, the time required for removing the moved largest connected component in the thresholded dif-
or moving objects in the scene is about six seconds. In ference map.
the binary image corresponding to the frame 810 we The images shown in the middle row of Fig. 8 are the
can see a chair, which has been previously moved, and raw depth images. As we already mentioned, the near-
which disappears in the binary image corresponding est neighbor-based interpolation is executed on the
to frame 1010. Once again, the update of the depth depth maps in order to fill the holes in the maps and
reference image has been achieved in about six sec- to get the maps with meaningful values for all pixels.
onds. As we can observe, the updated depth reference Thanks to such an interpolation the delineated persons
image allows us to extract the person’s silhouette in contain a smaller amount of artefacts.
the depth images. In order to eliminate small objects
the depth connected components were extracted. Af- 5.2 V-disparity based ground plane extraction
terwards, small artifacts were removed. Otherwise, the In [24] a method based on v-disparity maps between
depth images can be cleaned using morphological ero- two stereo images has been proposed to achieve reliable
sion. obstacle detection. Given a depth map provided by the
Figure 8 Delineation of person using depth reference image. RGB images (upper row), depth (middle row) and binary images
depicting the delineated person (bottom row).
Kwolek and Kepski Page 9 of 14
Kinect sensor, the disparity d can be determined in the noting that ordinary HT operating on thresholded v-
following manner: disparity images often gives incorrect results. For visu-
alization purposes the accumulator values were divided
by 1000. As we can see on Fig. 10, the highest peak
b·f of the accumulator is for a line with Θ approximately
d= (2)
z equal to zero degrees. This means that it corresponds
to a vertical line, i.e. line corresponding to the room
where z is the depth (in meters), b stands for the
horizontal baseline between the cameras (in meters), walls, see Fig. 9c. In order to simplify the extraction of
whereas f stands for the (common) focal length of the the peak corresponding to the floor, only the bottom
cameras (in pixels). The IR camera and the IR projec- half of the v-disparity maps is subjected to processing
tor form a stereo pair with a baseline of approximately by HT, see also Fig. 9c. Thanks to such an approach
b = 7.5 cm, whereas the focal length f is equal to 580 as well as executing the HT on a predefined range of Θ
pixels. and ρ, the line corresponding to floor can be estimated
Let H be a function of the disparities d such that reliably.
H(d) = Id . The Id is the v-disparity image and H
accumulates the pixels with the same disparity from
a given line of the disparity image. Thus, in the v-
disparity image each point in the line i represents the
number of points with the same disparity occurring in
the i-th line of the disparity image. Figure 9c illustrates
the v-disparity image that corresponds to the depth
image acquired by the Kinect sensor and depicted on
Fig. 9b.
a) b) c)
Figure 10 Accumulator of the Hough transform operating on
v-disparity values from the image shown on Fig. 9c.
termined on the basis of the following equation: Given the delineated person in the depth image along
with the automatically extracted parameters of the
|aXc + bYc + cZc + d| equation describing the floor, the aforementioned fea-
D= √ (3)
a2 + b2 + c2 tures are easy to calculate.
where Xc , Yc , Zc stand for the coordinates of the per- 6 The classifier for fall detection
son’s centroid. The parameters should be re-estimated
subsequent to each change of the Kinect location or At the beginning of this Section we discuss the dataset
orientation. A relevant method for estimating 3D cam- that was recorded in order to extract the features for
era extrinsic parameters has been proposed in [26]. It training as well as evaluating of the classifier. After
operates on three sets of points, which are known to that, we overview the SVM-based classifier.
be orthogonal. These sets can either be identified us-
ing a user interface or by a semi-automatic plane fitting 6.1 The training dataset
method. A dataset consisting of images with normal activi-
ties like walking, sitting down, crouching down and
5.3 Depth features for person detection lying has been composed in order to train the classi-
The following features were extracted in a collection fier responsible for examination whether a person is
of the depth images in order to acknowledge the fall lying on the floor and to evaluate its detection per-
hypothesis, which is signaled by the threshold-based formance. In total 612 images were selected from UR
procedure: Fall Detection Dataset (URFD)[1] and another image
• h/w - a ratio of width to height of the person’s sequences, which were recorded in typical rooms, like
bounding box office, classroom, etc. The selected image set consists
• h/hmax - a ratio expressing the height of the per- of 402 images with typical ADLs, whereas 210 images
son’s surrounding box in the current frame to the depict a person lying on the floor. The aforementioned
physical height of the person depth images were utilized to extract the features dis-
• D - the distance of the person’s centroid to the cussed in Subsection 5.3. The whole UR Fall Detec-
tion dataset consists of 30 image sequences with 30
floor
falls. Two types of falls were performed by five per-
• max(σx , σz ) - standard deviation from the cen-
sons, namely from standing position and from sitting
troid for the abscissa and the applicate, respec-
tively. [1]
http://fenix.univ.rzeszow.pl/~mkepski/ds/uf.html
Figure 11 Multivariate classification scatter plot for features utilized for training of the fall classifier.
Kwolek and Kepski Page 11 of 14
on the chair. All RGB and depth images are synchro- a fall occurred, and thus it is the classifier’s ability to
nized with motion data, which were acquired by the identify a condition correctly.
x-IMU inertial device. The specificity is the number of true negative (TN)
Figure 11 depicts the scatter plot, in which a collec- decisions divided by the number of actual negative
tion of scatter plots is organized in a two-dimensional cases (number of true negatives plus number of false
matrix simultaneously to provide correlation informa- positives). It is the probability of non-fall, given that a
tion among the attributes. As we can observe, the non-fall ADL took place, and thus it shows how good
overlaps in the attribute space are not too significant. a classifier is at avoiding false alarms. The accuracy
Thus, a linear SVM was utilized for classifying lying is the number of correct decisions divided by the total
poses and typical ADLs. Although the non-linear SVM number of cases, i.e. the sum of true positives plus sum
has usually better effectiveness in classification of non- of true negatives divided by total instances in popu-
linear data than its linear counterpart, it has much lation. That is, the accuracy is the proportion of true
higher computational demands for prediction. results (both true positives and true negatives) in the
population. The precision or positive predictive value
6.2 Support Vector Machines (SVM) (PPV) is equal to true positives divided by sum of
The basic idea of the SVM classification is to find a true positives and false positives. Thus, it shows how
separating hyperplane that corresponds to the largest many of the positively classified falls were relevant. In
possible margin between the points of different classes Tab. 1 are shown results that were obtained in 10-fold
[27]. The optimal hyperplane for an SVM means cross-validation by the classifier responsible for the ly-
the one with the largest margin between the two ing pose detection and the aforementioned dataset. As
classes, so that the distance to the nearest data we can see, both specificity and precision are equal
point of both classes is maximized. Such a largest to 100%, i.e. the ability of the classifier to avoid false
margin means the maximal width of the tile paral- alarms and its exactness assume perfect values.
lel to the hyperplane that contains no interior data Table 2 shows results of experimental evaluation of
points and thus incorporating robustness into deci- the system for fall detection, which were obtained on
sion making process. Given a set of datapoints D: depth image sequences from URFD dataset. They were
n
D = {(xi , yi )|xi ∈ Rp , yi ∈ {−1, 1}}i=1 where each ex- obtained on thirty image/acceleration sequences with
ample xi is a point in p-dimensional space and yi is the falls and thirty image/acceleration sequences with typ-
corresponding class label, we search for vector ω ∈ Rp ical ADLs like sitting down, crouching down, picking-
and bias b ∈ R, forming the hyperplane H: ω T x+b = 0 up an object from the floor and lying on the sofa. The
that seperates both classes so that: yi (ω T xi + b) ≥ 1. number of images in the sequences with falls is equal
The optimization problem that needs to be solved is: to 3000, whereas the number of images with sequences
minω,b 21 ω T ω subject to: yi (ω T xi +b) ≥ 1. The problem with ADLs is equal to 9000. All images have corre-
consists in optimizing a quadratic function subject to
sponding motion data. In the case of incorrect response
linear constraints, and can be solved with an off-the-
of the system the remaining part of the sequence has
shelf Quadratic programming (QP) solver. The linear
been omitted. This means that the detection scores
SVM can perform prediction with p summations and
were determined on the basis of the number of the cor-
multiplications, and the classification time is indepen-
rectly/incorrectly classified sequences. As we can ob-
dent of the number of support vectors. We executed
serve, the Threshold UFT method [10] achieves good
LIBSVM software [28] on a PC computer to train the
results. The results obtained by SVM-classifier oper-
fall detector.
ating on only depth features are slightly worse than
results of Threshold UFT method. The reason is that
7 Experimental results the update of the depth reference image was realized
We evaluated the SVM-based classifier and compared without the support of the motion information. This
it with a k-NN classifier (5 neighbors). The classifiers means that a simplified system has been built using
were evaluated in 10-fold cross-validation. To exam- only blocks, which are indicated in Fig. 5 as numer-
ine the classification performances we calculated the als in circles. In particular, in such a configuration of
sensitivity, specificity, precision and classification ac- the system all images are processed in order to ex-
curacy. The sensitivity is the number of true positive tract the depth reference image. The algorithm using
(TP) responses divided by the number of actual pos- both motion data from accelerometer and depth maps
itive cases (number of true positives plus number of for verification of IMU-based alarms achieves the best
false negatives). It is the probability of fall, given that performance. Moreover, owing to the use of the IMU
Kwolek and Kepski Page 12 of 14
Estimated
Precision=100%
Sens.=99.05% Spec.=100%
Fall 208 0
k-NN Accuracy=99.67%
No fall 2 402 Precision=100%
Sens.=99.05% Spec.=100%
is higher than an assumed threshold value, the algo- 10. Bourke, A.K., O’Brien, J.V., Lyons, G.M.: Evaluation of a
rithm extracts the person, calculates the features and threshold-based tri-axial accelerometer fall detection algorithm. Gait &
Posture 26(2), 194–199 (2007)
then executes the SVM-based classifier to authenticate 11. Kangas, M., Konttila, A., Lindgren, P., Winblad, I., Jamsa, T.:
the fall alarm. We demonstrate that person surround- Comparison of low-complexity fall detection algorithms for body
ing features together with the distance between the attached accelerometers. Gait & Posture 28(2), 285–291 (2008)
12. Bourke, A.K., Lyons, G.M.: A threshold-based fall-detection algorithm
person center of gravity and floor lead to reliable fall using a bi-axial gyroscope sensor. Medical Engineering & Physics
detection. The parameters of the floor equation are 30(1), 84–90 (2008)
determined automatically. The extraction of the per- 13. Degen, T., Jaeckel, H., Rufer, M., Wyss, S.: SPEEDY: A fall detector
in a wrist watch. In: Proc. of the 7th IEEE Int. Symp. on Wearable
son is only executed if the accelerometer indicates that Comp., p. 184. IEEE Computer Society, Washington, DC, USA (2003)
he/she is in motion. The person is extracted through 14. Sposaro, F., Tyson, G.: ifall: An Android application for fall monitoring
the differencing the current depth map from the on-line and response. In: IEEE Int. Conf. on Engineering in Medicine and
updated depth reference map. The system permits un- Biology Society, pp. 6119–6122 (2009)
15. Anderson, D., Keller, J.M., Skubic, M., Chen, X., He, Z.: Recognizing
obtrusive fall detection as well as preserves privacy of falls from silhouettes. In: Annual Int. Conf. of the Engineering in
the user. However, a limitation of the Kinect sensor Medicine and Biology Society, pp. 6388–6391 (2006)
is that sunlight interferes with the pattern-projecting 16. Rougier, C., Meunier, J., St-Arnaud, A., Rousseau, J.: Monocular 3D
head tracking to detect falls of elderly people. In: Annual Int. Conf. of
laser, so the proposed fall detection system is most the IEEE Eng. in Medicine and Biology Society, pp. 6384–6387 (2006)
suitable for indoor use. 17. Cucchiara, R., Prati, A., Vezzani, R.: A multi-camera vision system for
fall detection and alarm generation. Expert Syst. 24(5), 334–345
(2007)
Acknowledgements 18. Miaou, S.-G., Sung, P.-H., Huang, C.-Y.: A customized human fall
This work has been supported by the National Science Centre (NCN) detection system using omni-camera images and personal information.
within the project N N516 483240. Distributed Diagnosis and Home Healthcare, 39–42 (2006)
Author details 19. Jansen, B., Deklerck, R.: Context aware inactivity recognition for
1
AGH University of Science and Technology, 30 Mickiewicza Av., 30-059 visual fall detection. In: Proc. IEEE Pervasive Health Conference and
Kraków, Poland. 2 University of Rzeszow, 16c Rejtana Av., 35-959 Workshops, pp. 1–4 (2006)
Rzeszów, Poland. 20. Kepski, M., Kwolek, B., Austvoll, I.: Fuzzy inference-based reliable fall
detection using Kinect and accelerometer. In: The 11th Int. Conf. on
References Artificial Intelligence and Soft Computing. LNCS, vol. 7267, Springer,
1. Chan, M., Esteve, D., Escriba, C., Campo, E.: A review of smart pp. 266–273 (2012)
homes - present state and future challenges. Computer Methods and 21. Kepski, M., Kwolek, B.: Fall detection on embedded platform using
Programs in Biomedicine 91(1), 55–81 (2008) Kinect and wireless accelerometer. In: 13th Int. Conf. on Computers
2. Cook, A.M.: The future of assistive technologies: A time of promise Helping People with Special Needs. LNCS, vol. 7383, Springer, pp.
and apprehension. In: Proc. of the 12th Int. ACM SIGACCESS Conf. 407–414 (2012)
on Comp. and Accessibility, pp. 1–2. ACM, New York, USA (2010) 22. Mastorakis, G., Makris, D.: Fall detection system using Kinect’s
3. Hoflinger, F., Muller, J., Zhang, R., Reindl, L.M., Burgard, W.: A infrared sensor. J. of Real-Time Image Processing, 1–12 (2012)
wireless micro inertial measurement unit (imu). Instrumentation and 23. 3D Orientation Sensor IMU. http://www.test.org/doe/
Measurement, IEEE Transactions on 62(9), 2583–2595 (2013) 24. Labayrade, R., Aubert, D., Tarel, J.-P.: Real time obstacle detection in
4. Buesching, F., Kulau, U., Gietzelt, M., Wolf, L.: Comparison and stereovision on non flat road geometry through ”v-disparity”
validation of capacitive accelerometers for health care applications. representation. In: Intelligent Vehicle Symposium, 2002. IEEE, vol. 2,
Comp. Methods and Programs in Biomedicine 106(2), 79–88 (2012) pp. 646–6512 (2002)
5. Heinrich, S., Rapp, K., Rissmann, U., Becker, C., Kőnig, H.-H.: Cost 25. Fischler, M.A., Bolles, R.C.: Random sample consensus: A paradigm
of falls in old age: a systematic review. Osteoporosis Int. 21, 891–902 for model fitting with applications to image analysis and automated
(2010) cartography. Commun. ACM 24(6), 381–395 (1981)
6. Noury, N., Fleury, A., Rumeau, P., Bourke, A.K., Laighin, G.O., Rialle, 26. Deklerck, R., Jansen, B., Yao, X.L., Cornelis, J.: Automated estimation
V., Lundy, J.E.: Fall detection - principles and methods. In: IEEE Int. of 3d camera extrinsic parameters for the monitoring of physical
Conf. on Eng. in Medicine and Biology Society, pp. 1663–1666 (2007) activity of elderly patients. In: XII Mediterranean Conference on
7. Yu, X.: Approaches and principles of fall detection for elderly and Medical and Biological Engineering and Computing. IFMBE
patient. In: 10th Int. Conf. on E-health Networking, Applications and Proceedings, vol. 29, pp. 699–702 (2010)
Services, pp. 42–47 (2008) 27. Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3),
8. Igual, R., Medrano, C., Plaza, I.: Challenges, issues and trends in fall 273–297 (1995)
detection systems. BioMedical Engineering OnLine 12(1), 1–24 (2013) 28. Chang, C.-C., Lin, C.-J.: Libsvm: A library for Support Vector
9. Mubashir, M., Shao, L., Seed, L.: A survey on fall detection: Principles Machines. ACM Trans. Intell. Syst. Technol. 2(3), 1–27 (2011)
and approaches. Neurocomputing 100, 144–152 (2013)
Kwolek and Kepski Page 14 of 14
29. Stone, E.E., Skubic, M.: Unobtrusive, continuous, in-home gait Engineering 60(10), 2925–2932 (2013)
measurement using the microsoft kinect. IEEE Trans. on Biomedical