International Core Journal of Engineering 2020-26 | Page 101
IV. E XPERIMENTS
Experiments are conducted on a computer with Intel Core
i7 CPU(3.60GHz). The code is implemented by MATLAB
2018. The threshold parameter in the de-noising process is
set to 16 in the experiments.
3) In the de-noising processing, a threshold parameter is
used to process the Difference_Gabor, to get the
Difference_Gabor_Threshold. The Difference_Gabor_T-
hreshold is converted into a binary graph, by setting non-
zero pixel values to 1, and others to 0. Then the method
described in the postprocessing part is used to process the
binary graph to obtain, Bitmap_Gabor.
Fig. 5. The Gabor function act on the highway dataset. (a):Total difference
image; (b): Total difference after binary processing; (c): Total difference
after de-noising processing; (d): detection result: Bitmap_Gabor. A. The Cdnet dataset
The Change Detection.net dataset ( Cdnet ) is the first
large-scale benchmark dataset established by the IEEE
Change Detection Workshop organizing committee in 2012.
It is a video dataset devoted to the evaluation of change and
motion detection approaches. In 2014, the IEEE Change
Detection Workshop organizing committee expanded the
dataset by adding 22 additional videos in 5 new categories.
The Cdnet dataset covers most application scenarios of
moving target detection and has become a main platform for
fair comparison of algorithms in the research of moving
target detection.
C. Postprocessing
The binary image is processed by the morphological
closing [19], filled and smoothed by median filter [20] to
obtain the detection result. The functions in MATLAB 2018
are used in the experiments. The proposed method is evaluated on the Cdnet 2014
dataset [21] which totally includes 53 scenarios, categorized
into 11 classes. The baseline category is the best test data to
evaluate the generality of exercise objective testing
algorithms. It contains four catalogs that are highway, office,
pedestrian, and PETS2006.
An example of detection using the texture features is
shown in Fig. 5.
(a)
(b)
(c)
(d)
First, the original image is processed by the
morphological closing which performs closing with the
structuring element. To complete this step, the
IMCLOSE(IM,SE) function is used . IM can be any numeric
or logical class and any dimension, and must be non-sparse.
SE is a single structuring element object. In our experiment,
IM is the input binary image, SE is the shape of a rectangle
of 10 by10 for doing closing operations.
B. Results
The results of the proposed algorithm on the baseline
dataset are shown in Fig. 7, 8, 9, and 10. In all these figures,
(a) is the original video frame of the baseline dataset, (b) is
the final result of the moving target extracted with the
proposed method, (c) is the binary graph, Bitmap_Gabor,
obtained only by using the texture based inter-frame
difference detection, (d) is a binary graph, Bitmap_Lab,
obtained only based on color based inter-frame difference
detection, and (e) is the binary graph, Binary_image, based
on both texture features and color features.
Then the region and holes in the image are filled. The
IMFILL(BW1, LOCATIONS) function is used, which is
based on the principle of a flood-fill operation on
background pixels of the input binary image BW1. It starts
from the points specified in LOCATIONS, and the parameter
of LOCATIONS is chosen to be holes.
Fig. 7 shows the experimental results conducted on the
Highway dataset. It can be seen that both texture features
(Bitmap_Gabor) and color features (Bitmap_Lab) can detect
the moving cars well, while texture features work better on
the little cars and color features work better on the closest car.
And it can be seen that combination of the two features
(Binary_image) produces the best result for the car detection.
Finally, the binary image is operated by the median
filtering operation to smoothing the noise in the image. The
MEDFILT2 (A, [M N]) function performs median filtering
of the matrix A in two dimensions. Each output pixel
contains the median value in the M-by-N neighborhood
around the corresponding pixel in the input image.
Fig. 8 shows the experiment results conducted on the
Office dataset. Fig. 9 shows the experiment results conducted
on the Pedestrian dataset. It can be seen that for the two
datasets, texture features (Bitmap_Gabor) can detect the full
body of the person successfully, as shown in Fig. 8(c) and
Fig. 9(c), while the color features (Bitmap_Lab) only detects
few parts of the body, as shown in Fig. 8(d) and Fig. 9(d).
The combination results shown in Fig. 8(e) and Fig. 9(e) are
mainly the same as the results from the texture features.
D. Combination
1) Binary_image is obtained by combining the
difference binary image representing color features,
Bitmap_Lab, and the different binary image representing
texture features, Bitmap_Gabor. The combination is done
by the logical AND operation.
2) Apply the Binary_image as a mask on the original
frame F2 to get the moving target.
Fig. 10 shows the experiment results conducted on the
PETS2006 dataset. It can be seen that texture features
(Bitmap_Gabor) perform much better than color features
(Bitmap_Lab), while there still exits some holes in the result
of the texture features and the combination result. The person
closest to the center of the image is detected successfully,
while the two people on the bottom right corner are only
partially detected. This may be due to that the two people on
the bottom right corner have just entered into the view
shortly.
An example for the combination is shown in Fig. 6.
Fig. 6. The combination result
79