automatic vehicle detection using various object …1.4 image processing with mathcad and matlab....

13
International Journal on Future Revolution in Computer Science & Communication Engineering Volume: 1 Issue: 1 16 28 _____________________________________________________________________________________ 16 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________ Automatic Vehicle Detection Using Various Object Detecting Algorithm and Thresholding Methods Aarti Sharma, M. Tech Scholar (CSE), ECB Rajasthan (India) Abstract:-The digital image processing deals with developing a digital system to performs experiments and operations on a digital image with the use of computer algorithms. An image is nothing more than a 2D mathematical function f(x,y) where x and y are two horizontally and vertically co-ordinates. Object recognition is one of the most important applications of image processing. Vehicle detection from a satellite image or aerial image is one of the most interesting and challenging research topics from past few years. Vehicle detection from satellite image is one of the applications of object detection. The traffic and crowd is increasing everyday in all over the world. Satellites images are normally used for weather forecasting and geographical applications. So, Satellites images may be also good for the detecting traffic using Image processing. This thesis used simple morphological recognition method for vehicle detection using image processing technique in Matlab which is best method for detection of cars, trucks and buses. We can easily compute the total numbers of vehicles in the desired area in the satellite image and vehicles are shown under the bounding box as a tiny spots. Here we compare two algorithms like pixel thresholding and Otsu thresholding method. According to our result Pixel level thresholding is better than Otsu method. Keywords:- Vehicles at surface, object detection algorithm, thresholding values, airel images. __________________________________________________*****_________________________________________________ 1. Introduction 1.1 Introduction of Image processing In world of computer science Image processing is a quickly growing area. Its growth has filled by technological advances in, computer processors, digital imaging and mass storage devices. Analog imaging is over taking by digital systems, for their affordability & flexibility. Most seen examples in daily routine, are medicines, video production remote sensing, photography, film and security monitoring. Huge volumes of digital images data every day produced by these and other sources, it’s too huge that, we cannot check it manually [1]. 1.2 Digital images An image may be defined as a two-dimensional function, f(x, y), where x and y are spatial (plane) coordinates, and the amplitude of at any pair of coordinates (x, y) is called the intensity or gray level of the image at that point. When x, y, and the amplitude values of f are all finite, discrete quantities, we call the image a digital image. The field of digital image processing refers to processing digital images by means of a digital computer. Note that a digital image is composed of a finite number of elements, each of which has a particular location and value. These elements are referred to as picture elements, image elements, pels, and pixels. Pixel is the term most widely used to denote the elements of a digital image [2]. Three types of passive image sensing technologies are given below, which have been made for target detection. (1)Image Processing: Video Camera Nowadays using video cameras to collect data for recognition studies can be achieved at a lower cost compared to other image sensing techniques [Nooralahiyan et al., 1997]. On the other hand, the use of video camera images is not suitable under certain monitoring conditions, for example during the night or foggy conditions [Koch et al., 2006; Cheng et al., 2005]; which may imply the unsuitability of them for the current research. (2)Image Processing: Satellite Satellite image detection is suitable when surveillance of a wide area is required although the installation & maintenance costs may be high [Roper, 2005]. The other disadvantages are the inability to monitor & the low ground resolution on a cloudy day. 1.2.1Advantages with digital imaging Benefits of digital image processing: low cost processing, consistently high quality of the image, & the ability to manipulate all aspects of the process. As long as computer processing speed was increasing, cost of storage memory started dropping, this field started growing [6]. There are various benefits with digital cameras compared to other means of acquiring data. Main purposes were to make large spatial resolution of CCD cameras. A resolution of 3120 X 2068 pixels allows for a rather large span in the size of the features detected by the cameras. This is very crucial for reliable fractal analysis of scaling relations. The pixels size 9X 9m is of the same order as the grain size in photographic films. The main advantages of the CCD cameras are making it superior quality to video cam-recorders or ordinary cameras. 1.3 Image Processing Techniques This portion handles acquisition, processing & formation of images. It is shown the best way in diagram where the processes involved & evolution of the considered information (images).

Upload: others

Post on 28-Feb-2020

11 views

Category:

Documents


0 download

TRANSCRIPT

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

16 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Automatic Vehicle Detection Using Various Object Detecting Algorithm and

Thresholding Methods

Aarti Sharma,

M. Tech Scholar (CSE), ECB

Rajasthan (India)

Abstract:-The digital image processing deals with developing a digital system to performs experiments and operations on a digital image with

the use of computer algorithms. An image is nothing more than a 2D mathematical function f(x,y) where x and y are two horizontally and

vertically co-ordinates. Object recognition is one of the most important applications of image processing.

Vehicle detection from a satellite image or aerial image is one of the most interesting and challenging research topics from past few years.

Vehicle detection from satellite image is one of the applications of object detection. The traffic and crowd is increasing everyday in all over the

world. Satellites images are normally used for weather forecasting and geographical applications. So, Satellites images may be also good for the

detecting traffic using Image processing. This thesis used simple morphological recognition method for vehicle detection using image processing

technique in Matlab which is best method for detection of cars, trucks and buses. We can easily compute the total numbers of vehicles in the

desired area in the satellite image and vehicles are shown under the bounding box as a tiny spots. Here we compare two algorithms like pixel

thresholding and Otsu thresholding method. According to our result Pixel level thresholding is better than Otsu method.

Keywords:- Vehicles at surface, object detection algorithm, thresholding values, airel images.

__________________________________________________*****_________________________________________________

1. Introduction

1.1 Introduction of Image processing

In world of computer science Image processing is a quickly

growing area. Its growth has filled by technological advances in,

computer processors, digital imaging and mass storage devices.

Analog imaging is over taking by digital systems, for their

affordability & flexibility. Most seen examples in daily routine, are

medicines, video production remote sensing, photography, film and

security monitoring. Huge volumes of digital images data every

day produced by these and other sources, it’s too huge that, we

cannot check it manually [1].

1.2 Digital images

An image may be defined as a two-dimensional function, f(x, y),

where x and y are spatial (plane) coordinates, and the amplitude of

at any pair of coordinates (x, y) is called the intensity or gray level

of the image at that point. When x, y, and the amplitude values of f

are all finite, discrete quantities, we call the image a digital image.

The field of digital image processing refers to processing digital

images by means of a digital computer. Note that a digital image is

composed of a finite number of elements, each of which has a

particular location and value. These elements are referred to as

picture elements, image elements, pels, and pixels. Pixel is the term

most widely used to denote the elements of a digital image [2].

Three types of passive image sensing technologies are given

below, which have been made for target detection.

(1)Image Processing: Video Camera

Nowadays using video cameras to collect data for recognition

studies can be achieved at a lower cost compared to other image

sensing techniques [Nooralahiyan et al., 1997]. On the other hand,

the use of video camera images is not suitable under certain

monitoring conditions, for example during the night or foggy

conditions [Koch et al., 2006; Cheng et al., 2005]; which may

imply the unsuitability of them for the current research.

(2)Image Processing: Satellite

Satellite image detection is suitable when surveillance of a wide

area is required although the installation & maintenance costs may

be high [Roper, 2005]. The other disadvantages are the inability to

monitor & the low ground resolution on a cloudy day.

1.2.1Advantages with digital imaging

Benefits of digital image processing: low cost processing,

consistently high quality of the image, & the ability to manipulate

all aspects of the process. As long as computer processing speed

was increasing, cost of storage memory started dropping, this field

started growing [6].

There are various benefits with digital cameras compared to other

means of acquiring data. Main purposes were to make large spatial

resolution of CCD cameras. A resolution of 3120 X 2068 pixels

allows for a rather large span in the size of the features detected by

the cameras. This is very crucial for reliable fractal analysis of

scaling relations. The pixels size 9X 9m is of the same order as the

grain size in photographic films. The main advantages of the CCD

cameras are making it superior quality to video cam-recorders or

ordinary cameras.

1.3 Image Processing Techniques

This portion handles acquisition, processing & formation of

images. It is shown the best way in diagram where the processes

involved & evolution of the considered information (images).

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

17 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Figure 1: Image processing technique [7]

1.4 Image processing with MathCAD and MatLab.

The MathCAD & MatLab environments are perfectly suited to

image processing. Particularly, MatLab's matrix-oriented language

is perfectly suited for manipulating images, which are not more

than visual renderings of matrices. The result is a very easy &

economical way of expressing image processing operations. In

addition both programs have Image Processing Toolboxes which

provide a flexible & powerful environment for image analysis &

processing. Both programs were used to perform different

calculations on images [8].

There are various benefits of using MatLab & MathCAD for image

analysis. One is the ability to have directed access to any portion of

available information what in general is not possible with many

commercial image analysis systems. We can stop any calculation

anytime with the help of this program, change a portion of the

calculation procedure and then restart the calculations from the

point which was affected by the changes without recompiling the

code as it usually happens with programming in C, or even

restarting the calculations from the starting. New techniques are

developed and also these types of abilities are very helpful in

researches. Main disadvantage of this programs is the relatively

slow computational speed compared to compiled C code. It is

because of the need of the code to be translated first into a machine

code and only then to be executed. Therefore, complex image

processing applications can be better implemented by the use of

high level programming languages as C or C++, rather than using

the softwares like MathCAD or MatLab.

2. Literature survey

This survey included various types of images, image segmentation,

threshold techniques and previous work for detecting the vehicles.

A literature survey has been conducted in order to understand the

current and past research trends in the area of automated moving

vehicle recognition systems. And this chapter discuss about the

previous work for detecting the vehicles using pixel level and Otsu

thresholding techniques.

2.1 Previous work

Semantic analysis of changes in satellite imagery requires the

detection of changes as a first step. Change detection is a well

studied problem in computer vision and many modern change

detection algorithms exist. The most notable of these rely on a

technique called background modeling. In this approach, a number

of images of the scene are used to learn what the normal

background appearance of the scene should look like so that given

a new image, the pixels with abnormal appearance can be detected

as changes [22].

Most of the literature addresses 2-d background modeling problem

in which the images of the scene can be registered to a common 2-

d frame. This is trivially the case with stationary cameras or is

possible through various registration techniques. Early attempts at

2-d background modeling [25, 26, and 27] use a single Gaussian

density at each pixel to model background intensity plus

acquisition noise. These densities are updated with simple adaptive

filters as new observations of the pixel are made. In this approach,

the single Gaussian cannot model the pixel variation present in

many real world scenes such as frequently occluded surfaces

(roads), slow moving objects, and vegetation blowing in the wind.

Stauffer and Grimson [28] proposed the first fully general

algorithm to model a more realistic dynamic background. They

used a mixture of Gaussians [29] density at each pixel which is a

widely used and well studied distribution for modeling multi-

modal data. Because of its fast and robust performance on all types

of realistic scenes, the Stauffer-Grimson algorithm has become the

standard technique for change detection.

In the case of non-stationary cameras, the registration step required

for 2-d background modeling works reasonably well only if the

camera motion is mostly limited to pan/tilt or the scene is mostly

planar. This is because the image registration techniques rely on

global image transformations such as homographies, or polynomial

warping for satellite imagery. The latter allows for more distortion

of the image but is still a global transformation and cannot do local

alignments of objects with 3-d relief between images, which is the

case for satellite imagery. These limitations call for a 3-d solution

to the problem.

3-d change detection is a relatively less researched field in

computer vision. Earlier approaches [30,31] used manually

constructed 3-d site models to make correspondences between

images so that change detection algorithm can be applied. The

overhead of constructing 3-d geometry is infeasible for this kind of

approach to be used in modern applications. Heller et al. [32] use

stereo pairs of satellite images to reconstruct 3-d geometry of the

scene and then compare reconstructed geometry from different

pairs of images to detect 3-d changes to the scene. This algorithm

is more applicable but it depends on having stereo pairs and it

cannot detect appearance changes on the surfaces of the scene such

as moving vehicles and shadows. Recently, a new approach that

combines the power of Stauffer-Grimson style appearance

modeling with automated 3-d geometry discovery has been

proposed [32]. This volumetric appearance modeling (VAM)

approach is suitable for change detection from satellite imagery for

several reasons. First, it is not assumed that there is a fixed set of

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

18 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

images available at the beginning and the model of the world is

built up incrementally as images are observed in an unbounded

sequence. This type of processing is called an on-line algorithm in

computer science terminology. This makes it possible to discard

each image after processing and reduces storage requirements

significantly making the method applicable to satellite domain

where image sizes are on the order of GBytes. Second, VAM

makes no assumptions about frame-to-frame continuity. Third,

even though a camera model, namely a cubic rational polynomial

camera [33] for the case of satellite imagery, is supplied with each

image to be processed, in general it is necessary to calibrate and

register such projection models due to errors in platform motion

and internal parameters. These adjustments are outside the scope of

this paper and in this work it is assumed that the cameras are

correct.

However, the VAM framework is capable of supporting automated

camera calibration and registration [34]. Finally, it has been shown

[35] that the framework tolerates the amount of variability in

viewpoint, lighting, and atmospheric conditions such as haze that is

typically observed in satellite imagery.

2.2 Related Work for Object Recognition

Semantic analysis of changes in satellite imagery requires the

classification of change regions as the second step. The objects of

interest in this analysis are vehicles. But the classification

technique used is general and applies to several object categories

such as buildings, roads, bridges, etc. which could also be objects

of interest in satellite imagery [24]. From a probabilistic point of

view, being on a changing surface in the scene, and being on an

object surface are independent events for a pixel in an image.

Hence the output map of a change classification algorithm can

simply be computed as the product of the change map as given by

the Voxel World and the classification map as given by a

conventional classification algorithm. However the classification

module can proceed in a completely independent manner as well

and the algorithm used is in the realm of the classical classification

algorithms.

Object recognition is the problem of declaring existence of an

instance from an object class in a given image patch. The input

image may be larger and contain many instances of the class as in

the case of satellite imagery. However the algorithms usually

convolve the input image with the classification module which

boils down to the classification of a patch in the image that is of the

extent of the object instance. Thus, the classification module

explicitly achieves position invariance during recognition. By

carrying out processing at multiple scales, it is also possible to

achieve scale invariance. In the following discussion, the object

image will refer to the image patch that contains the whole extent

of the object as seen in the image and the image domain will refer

to this patch and not the whole input image.

2.3 Related work on Road Vehicle Recognition

This section describes research on road vehicles rather than

military ones. In general, research on road vehicle recognition is

performed for traffic management or automated driver assistance

system.

(a) [Sampan, 1997] A PhD thesis was published for traffic

monitoring by using a circular array, consisting of 152

microphones, but 143 of them were of interest. After the data were

pre-processed to only maintain the components between 2700Hz

and 5400Hz, the 30-dimension feature vectors were extracted from

the energy over each 0.2 seconds in the time domain. PCA was

processed to reduce the dimension to 24 before performing

classification with either kNN, Multilayer Perceptron (MLP) or

Adaptive Fuzzy Logic System (AFLS). Although the exact number

of vehicles in each class was not clear, it was stated that they

varied between each class; and there were data of 1327 vehicles in

total. Problems caused by this inequality of the sizes of training

sets was addressed in the thesis by considering the effect on the

training as well as introducing the learning factor, a method

intended to take care of the issue. Moreover, it attempted to deal

with the inequality derived from the fact some classes are easier to

be learned than others by allocating misclassification costs for each

pattern although there was no logical explanation regarding how

they were determined. The obtained classification accuracies were

97.95%, 92.24% and 78.67% for 2-class, 4- class, and 5-class

experiments respectively. Althugh the use of such a large number

of microphones may not be appropriate for developing a cost

effective and compact recognition system, this study played a good

role in initiating research on acoustic road vehicle recognition with

fairly accurate results.

(b) [Nooralahiyan et al., 1997, 1998] Researchers at the

University of Leeds published two similar papers on their acoustic

road vehicle classification studies for traffic monitoring in the late

1990s. The focus was on classification only thus detection

algorithms were not included. The study was motivated by the

progress in automated speech recognition research that apparently

let them believe feasibility of the short term spectrum, particularly

at low frequencies, for the task. Therefore, the resultant algorithm

choice was made heuristically but with influence of speech

recognition studies. In the first paper [Nooralahiyan et al., 1997],

they first conducted a feasibility study using acoustic signals of

four vehicles; such as a small saloon, a medium saloon, a

motorcycles, and a light goods vehicle. All of the recordings were

collected under relatively controlled conditions, particularly in

terms of the level of background noise and vehicle speed. At this

stage, the feature extraction methods used were: FFT and

autocorrelation method Linear Predictive Coding (LPC) both in

MATLAB, and also software for computational modelling of

hearing, which was developed by Patterson et al. [Patterson et al.,

1995]. For the latter, Equivalent Rectangular Bandwidth (ERB)

based gammatone filterbanks, covering between 100Hz and

12kHz, were used to simulate the movement of the basilar

membrane in the cochlea. Moreover, the same software platform

was also used to simulate auditory nerve activity patterns of the

cochlea’s inner hair cells. Within these methods, FFT was omitted

after the initial stage because its classification outcome was not

satisfactory. An unsupervised Kohonen Self-Organising Map

(SOM) was used for this first phase of the study.

They then carried out another set of experiments using signals

collected at various urban road sites where they had little control

compared with the first stage [Nooralahiyan et al., 1998]. In total,

over 200 recordings of various vehicles travelling at different

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

19 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

speed of no faster than 40 miles per hour (i.e. approximately 64 km

per hour), collected at more than 20 sites were utilised. The

collected feature vectors were classified by the supervised Time

Delay Neural Network (TDNN) with two hidden layers and no

feedback, between the following four classes; buses or lorries,

saloon cars, motorcycles, and light goods vehicle or vans. The

reported classification accuracies achieved, particularly with

adaptively changed threshold, were above 84%. These results were

seen as good, therefore, it might be a good reference for the the

current research.

(c) [Wu et al., 1998, 1999] Another example was reported in two

similar papers published firstly in a conference proceedings and

then in a journal. The adapted method was called “eigenfaces

methods”, which had been used in human face image recognition

beforehand. It is also known as Karhunen-Loeve expansion or

PCA. The feature vectors of acoustic signals were collected with

FFT power spectrum, which was then normalised per frame to unit

power. Some adjustment was also applied to reduce the impact of

insignificant parts before PCA was performed. The class of

unknown samples was determined according to the distance of the

new feature vector from the reference, which was prepared in

advance during training by using only selected data, recorded at the

same location under very similar conditions. Recordings of saloon

type cars were used as the training samples during the experiment.

The distributions of training and test sample sets were exhibited in

graphs, but no numerical values for classification accuracy nor

sample size was included. It was pointed out that the algorithm was

susceptible to the change in recording conditions; furthermore,

more data would be required to study sample distributions that

usefully display the signals’ distinguishable attributes. (d) [Jacyna

et al., 2005; Necioglu et al., 2005] A group of researchers at the

MITRE Corporation in the USA have published work on

developing acoustic road vehicle classification algorithms for

networked sensor systems [Jacyna et al., 2005]. Their approach

was based on a hierarchical network topology, and the

classification was performed between two road vehicle classes: the

“light” and the “heavy”. Firstly they looked into algorithm

development for a simple classification that can be performed at

each sensor node. The feature vectors were extracted by a

combination of FFT and 8-band filterbank, and then classified by a

so-called “linear-weighted classifier”. Although there was not

much information about the actual experimental results nor the

total number of different vehicles involved in the experiment, other

than mentioning the number of 125ms long audio file segments for

car and truck signals; it was reported that they managed to achieve

the minimum error rate of approximately 13%.

Secondly, according to another paper [Necioglu et al., 2005], a

more sophisticated classification algorithm that can be processed

on the features collected by the each node described above, but at

the next stage of the hierarchical network topology, was examined.

Although the feature extraction algorithm was similar to that of the

first one, this time they disclosed more information regarding the

experiment and the data set, in which the effects of small parameter

variation were also examined. The selected classification algorithm

for this stage was ML estimation with GMM. It was reported that

when the feature vectors, which were the filterbank output of

spectral energy, were scaled logarithmically; the minimum error

rate was found to be nearly 7%.

(e) [Sobreira-Seoane et al., 2008] An algorithm that has some

degree of similarity to the one studied for the current research in a

sense as they both consist of two stages, such as vehicle presence

detection and classification, was reported at a conference, which

some of the findings of this PhD research was also presented at. It

was realised that the obvious difference between the two

algorithms would be that Sobreira-Seoane purely relied on acoustic

signals. For vehicle detection, Short Time Energy (STE) and

envelope peaks were used. The chosen feature extraction

algorithms for classification stage were; Zero-Crossing Rate

(ZCR), “Spectral Centroid (centre gravity of spectral power

distribution)”, “Spectral Rolloff Point”, Subband Energy Ratio

(SBER), and MFCC. kNN and LDA algorithms were utilised as the

classifiers. The primary results showed that the combination of

kNN and three feature extraction algorithms; Spectral Rolloff

Point, SBER and MFCC, performed comparatively better, and

achieved nearly 90% accuracy.

(f) [Lu et al., 2008a,b] A vehicle detection algorithm, based on

studies in mammalian perception and neurology, was suggested in

articles in IEEE conference proceedings. This was another research

initiated from an augment that “there exists similarity between

vehicle and speech recognition”[Lu et al., 2008a, p.1336].

However, once again, they included no evidence nor discussion to

support such a view point. Their recommendation was to use

gammatone filterbank on STFT for feature extraction, dimension

reduction and the two-step decision making processes to first

detect the presence of vehicles before classifying them into four

vehicle groups. They conducted experiments to compare MFCC

and gammatone auditory filterbank on STFT. They showed that the

gammatone filterbanks outperformed MFCC under noisy

environment whereas MFCC was better in less noisy conditions. In

addition, a technique called Spectro-Temporal Representation

(STR) was described. This enhanced the correlation between

neighbouring signal frames, and the use of Nonlinear Hebbian

Learning (NHL) for reducing the dimension improved the

detection accuracy. In the best case, under less noisy environment

where the SNR was 10dB, it achieved about 95% accuracy, on

average under two kinds of noise although the performance

declined significantly as SNR decreased. They argued that the

algorithm could be expanded to classification tasks without

difficulty

(g) [Anami and Pagi, 2009] Recognition algorithms for two-wheel

vehicles were investigated. Vectors extracted by three time domain

methods, such as ZCR, STE and Root Mean Square(RMS) plus

two frequency domain feature extraction methods, “Mean and

Standard Deviation of Spectrum Centroid (CMEAN and CSD)”

were applied to the NN for classifying between two classes:

“bikes” and “scooters”. The recorded acoustic signals of 118

vehicles were utilised, of which 58 were for training and the rest

was for testing. It achieved 73.33% accuracy between the two

classes. The observed effects of vehcie ages and how well they

were maintained within the same class were also mentioned. Such

information can be useful for certain applications that have only

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

20 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

known and limited targets, however, it does not apply to the

current project.

(f) [Amit Satpathy 2014] proposes two sets of novel edge-texture

features, Discriminative Robust Local Binary Pattern (DRLBP)

and Ternary Pattern (DRLTP), for object recognition. By

investigating the limitations of Local Binary Pattern (LBP), Local

Ternary Pattern (LTP) and Robust LBP (RLBP), DRLBP and

DRLTP are proposed as new features. They solve the problem of

discrimination between a bright object against a dark background

and vice-versa inherent in LBP and LTP. DRLBP also resolves the

problem of RLBP whereby LBP codes and their complements in

the same block are mapped to the same code. Furthermore, the

proposed features retain contrast information necessary for proper

representation of object contours that LBP, LTP, and RLBP

discard. Our proposed features are tested on seven challenging data

sets: INRIA Human, Caltech Pedestrian, UIUC Car, Caltech 101,

Caltech 256, Brodatz, and KTH-TIPS2- a. Results demonstrate that

the proposed features outperform the compared approaches on

most data sets.

(g) [Xueyun Chen, Shiming Xiang, Cheng-Lin Liu, and Chun-

Hong Pan 2014] Detecting small objects such as vehicles in

satellite images is a difficult problem. Many features (such as

histogram of oriented gradient, local binary pattern, scale-invariant

feature transform, etc.) have been used to improve the performance

of object detection, but mostly in simple environments such as

those on roads. Kembhavi et al. proposed that no satisfactory

accuracy has been achieved in complex environments such as the

City of San Francisco. Deep convolutional neural networks

(DNNs) can learn rich features from the training data automatically

and has achieved state-of-the-art performance in many image

classification databases. Though the DNN has shown robustness to

distortion, it only extracts features of the same scale, and hence is

insufficient to tolerate large-scale variance of object. In this letter,

we present a hybrid DNN (HDNN), by dividing the maps of the

last convolutional layer and the maxpooling layer of DNN into

multiple blocks of variable receptive field sizes or max-pooling

field sizes, to enable the HDNN to extract variable-scale features.

Comparative experimental results indicate that our proposed

HDNN significantly outperforms the traditional DNN on vehicle

detection.

3. Experimental Result and Analysis

Satellite resolution and object-oriented detection method are being

better day by day in satellite images. Comparing to traditional data

acquired methods, in large area satellite images the heavy traffic

data can be faster and widely acquired. In this work first of all the

image quality is improved, and then the multiscale segmentation

technology is used and the vehicles are detected from satellite

images from supervised classification method [44]. The simple

morphological recognition method is used for vehicle detection

using image processing technique in Matlab which is one of the

best methods for detection of trucks, cars, bikes and buses etc.

4.1 Data Information

This thesis work is used five images named New 5, D10, New

3, New 4 and New 7 as a reference. And step by step process

is described on the image New 5. Images are shown in 4.1,

4.2, 4.3, 4.4, 4.5 figures.

Figure 4.1: New 5 image (reference image no. 1)

Figure 4.2: D10 image (reference image no. 2)

Figure 4.3: New 3 image (reference image no. 3)

Figure 4.4: New 4 image (reference image no. 4)

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

21 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Figure 4.5: New 7 image (reference image no. 5)

4.2 Methodology

Morphological recognition algorithms are used to develop an

automated system in MATLAB R2013a. In which satellite images

are taken as input and converted into gray scale image for pre-

processing. After conversion these images are converted into

binary images after image complement. After conversion canny

edge detection method has done and passed this detection to the

dilation process. The area is selected after filtration and dilation

where number of vehicles is maximum and vehicles are recognized

from the image in the form of bounding box. The number of

vehicles is counted by blob analysis. Here we are using reference

image New 5.

Figure 4.6: Block Diagram of vehicle detection from satellite image using canny edge

The steps in block diagram are elaborated below:

1) Satellite Image Acquisition:

a) Read Satellite image

b) Resize Image

2) Necessary Operations

a. RGB to Gray Scale conversion

3) Image segmentation process

a. Gray image to Binary conversion

b. Edge Detection

4) Image Enhancement

a. Filling Holes on images

b. Creating Holes Edge Detected Images

c. Filteration of image using Bewareopen

command using High pass filter

4.3 Process to detect vehicles from satellite images

Process to detect vehicle include many steps. Descriptions of these

steps are as following:

Figure 4.7: Acquired Image

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

22 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Figure 4.8: RGB to Gray Scale Conversion

Figure 4.8: Binary Converted Image

Figure 4.9: Canny Edge Detection

Figure 4.10: Filling contour holes for vehicle detection.

Figure 4.11: Filtration using high pass filter

Figure 4.12: cropped area

4.3.9 Detected Vehicles using Blob Analysis

Blob Analysis is a fundamental technique of machine

vision based on analysis of consistent image regions. As

such it is a tool of choice for applications in which the

objects being inspected are clearly discernible from the

background. Diverse set of Blob Analysis methods allows

creating tailored solutions for a wide range of visual

inspection problems.

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

23 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Figure 4.13: Detected Vehicles shown by green bounding box

Figure 4.14: pop up window after detecting the vehicles

4.4 Result

Same method is applied on the reference images new image

4, new image 3, new image 7, and D10. Figure 4.15, 4.17,

4.19, 4.21 and 4.23 shows the gray level graphs of reference

new image 5, new image 7, D10, New image 4 and new

image 3 respectively. In this approach we applied both pixel

level and Otsu thresholding method for all of the images. The

overall result will be shown as the comparison table and the

result shows that pixel level thresholding method is faster

than Otsu method.

Figure 4.15: Graph of New Image 5

Figure 4.16: Detected vehicle of new image 5

Figure 4.17: Graph of New image 7

Figure 4.18: Detected vehicle of new image 7

0

500

1000

1500

2000

2500

Displays number of gray value

0 50 100 150 200 250

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

24 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Figure 4.19: Graph of image D10

Figure 4.20: Detected vehicle of Image D10

Figure 4.21: Graph of New image 4

Figure 4.22: Detected vehicle of new image 4

Figure 4.23: Graph of New image 3

Figure 4.24: Detected vehicle of new image 3

Table 1 show the comparison between pixel level and Otsu

method which we used in this proposed work.

Table 2 shows that in the new 3 image when we calculates

the vehicles from pixel level thresholding method it

calculated 18 vehicles and when we calculate from Otsu

method then it calculate 11 vehicles. So it works 63.65%

faster.

In the image new 4 image when we calculates the vehicles

from pixel level thresholding method it calculated 15 vehicles

and when we calculate from Otsu method then it calculate

15 vehicles. So it works same.

.

In the image new 5 when we calculates the vehicles from

pixel level thresholding method it calculated 23 vehicles and

when we calculate from Otsu method then it calculate 22

vehicles. So it works 4.54% faster.

In the image D10 when we calculates the vehicles from pixel

level thresholding method it calculated 7 vehicles and when

we calculate from Otsu method then it calculate 8 vehicles.

So it works 12.5 % slower.

In the image new 7 when we calculates the vehicles from

pixel level thresholding method it calculated 11 vehicles and

when we calculate from Otsu method then it calculate 11

vehicles. So it works same.

The overall result shows that pixel level thresholding method

is faster than Otsu thresholding method.

Sr. No Pixel level

(This Method)

Otsu Method

Previous vehicle

0

1000

2000

3000

4000

5000

6000

7000

8000

Displays number of gray value

0 50 100 150 200 250

0

500

1000

1500

2000

2500

3000

3500

4000

4500

Displays number of gray value

0 50 100 150 200 250

0

500

1000

1500

2000

2500

3000

3500

Displays number of gray value

0 50 100 150 200 250

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

25 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

detection technique

1 No image

Complemented used

Image Complemented

Required

2 Fast Slow compare to Pixel

3 Not mostly used in

GUI

Maximum used in GUI

approach

4. Low Resolution

Camera is also used

Only High Resolution

Images are used here

Table-1 Comparison of Pixel Level and Otsu Threshold Method

for Vehicle Detection

4.4.1 Comparison table with reference images

Table 4.1: Comparison of pixel and Otsu method in reference

images

Overall Result of Vehicle Detection from Satellite Image

Reference

Images

Vehicle Detection

Thresholding Techniques

Result in

%

Pixel Level Otsu

Method

New 3 image 18 Vehicles 11 Vehicles 63.65%

fast

New4 Image 15 Vehicles 15 Vehicles 0% Same

New5 image 23 Vehicles 22 Vehicles 4.54% fast

D10 image 7 Vehicles 8 Vehicles -12.5% slow

New7 Image 11 vehicles 11 Vehicles 0 % Same

Reference Images Vehicle Detection Thresholding Techniques

Pixel Level Otsu Method

New 3 image

18 Vehicles

11 Vehicles

New4 Image

15 Vehicles

15 Vehicles

New5 image

23 Vehicles

22 vehicles

D10 image

7 Vehicles

8 Vehicles

New7 Image

11 vehicles

11 Vehicles

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

26 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

Table 2: comparison of pixel level and Otsu method in

percentage

% faster/slowe calculation = (pixel level thresholding-Otsu

Method)/Otsu Method

Overall Average among 5 Images would be 11.138%

Overall System Average = New3+New4+New5+D10+New7/5

Overall System Average = 63.65+0+4.54+ (- 12.5)+0/5

Pixel level Thresholding Overall System Average = 11.138%

Our Research says that pixel level threholding method is faster

than the Otsu method.

4.4.2 Profiler Time of Otsu Level Threholding

Figure 4.25: Profile time of Otsu level Thresholding

4.4.3 Profiler Time of Pixel Level Threholding:

Figure 4.26: Profile time of Pixel level Thresholding

5.1 Conclusion

By the time the numbers of vehicles and traffic have been

increasing day by day. With this increase, it is becoming harder to

track each vehicle for the purpose of traffic management and law

enforcement. Thus, a type of system is required, which is capable

of give appropriate solutions to the traffic issues and hence this

Vehicle detection from satellite image is developed using

morphological recognition algorithm in MATLAB and this thesis

compare two thresholding techniques to get the best result.

Morphological recognition algorithms are used to develop an

automated system in MATLAB R2013a. In which satellite images

are taken as input and converted into gray scale image for pre-

processing. After conversion these images are converted into

binary images after image complement. After conversion canny

edge detection method has done and passed this detection to the

dilation process. The area is selected after filtration and dilation

where number of vehicles is maximum and vehicles are recognized

from the image in the form of bounding box. The number of

vehicles is counted by blob analysis. Here we compare two

algorithms like pixel thresholding and Otsu thresholding method.

According to our result Pixel level thresholding is better than Otsu

method. Here image complementation are not used which make

system more powerful and make system highly applicable. The

result concluded that pixel level thresholding is better than Otsu

method for detecting the vehicles from satellite imaginary.

5.2 Future Scope

This approach makes easier to detect vehicles from

stationary images, but for moving cars or vehicles this

method will not be applicable, therefore this problem

can also be overcome in near future in order to get

more appropriate consequences.

The effect of light on vehicles reduces accuracy when

this approach is applied, so there is some scope to

improvement here.

REFERENCE

[1] S.Jayalakshmi, M.Sundaresan, "A Study of Iris

Segmentation Methods using Fuzzy CMeans and K-

Means Clustering Algorithm", International Journal of

Computer Applications (0975 – 8887) Volume 85 – No

11, January 2014.

[2] http://elearning.vtu.ac.in/17/e-Notes/DIP/Unit1-SH.pdf.

[3] Megha Soni,Anand Khare, Asst. Prof. Saurabh Jain, "A

SURVAY OF DIGITAL IMAGE PROCESSING AND ITS

PROBLEM", International Journal of Scientific and

Research Publications, Volume 4, Issue 2, February 2014

1 ISSN 2250-3153..

[4] http://folk.uio.no/walmann/Publications/PhD/node36.h

tml.

[5] Yugha R1, Dr S Uma, Swarnalatha S, Poovizhi.M,

"Multilevel Authentication System For Providing

Security", IPASJ International Journal of Computer

Science (IIJCS), Volume 3, Issue 3, March 2015.

[6] http://www.wisegeek.org/what-is-image-

processing.htm.

[7] http://www.ncsa.illinois.edu/People/kindr/phd/PART1.P

DF.

[8] K. Sumithra, S. Buvana, R. Somasundaram, "A Survey on

Various Types of Image Processing Technique",

International Journal of Engineering Research &

Technology (IJERT) Vol. 4 Issue 03, March-2015.

[9] Ian Darwin, Valerie Quercia and Tim O’Reilly, "X Window

System User’s Guide", March, 1995.

[10] "Digital Image Processing" book, By Rafael C Gonzalez,

page no. 29.

[11] http://shodhganga.inflibnet.ac.in/bitstream/10603/826

9/9/09_chapter%201.pdf. Page no. 1-4.

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

27 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________

[12] By Carolyn Asbury, "Brain Imaging Technologies and

Their Applications in Neuroscience".

[13] http://fas.org/irp/imint/docs/rst/Intro/Part2_1.html.

[14] http://bookboon.com/en/digital-image-processing-part-

one-ebook.

[15] Rafael C. Gonzalez, Richard E. Woods, "Digital Image

Processing Third Edition". 2008 by Pearson Education,

Inc.

[16] Ramya R, Anand Kumar S, Krinish N K, Suraj V, "Simulink

Component Recognition Using Image Processing",

International Journal of Scientific & Technology

Research Volume 4, Issue 02, February 2015.

[17] AMALOPRAVAM.G,HARISH NAIK T,JYOTI KUMARI,

"Transformation of Digital Images using Morphlogical

Operations", ISSN: 2278-9669, January 2013

(http://ijcsit.org)..

[18] e-Study Guide for: Artificial Intelligence by Stuart

Russell, ISBN 9780136042594,By Cram101 Textbook

Reviews.

[19] Trapti Sahu, Shabahat Hasan, "A Neural Network Based

Image Abstraction Technique", www.ijset.com, Volume

No.1, Issue No.2 pg:143-147, 01 April 2012..

[20] http://www.uptu.ac.in/pdf/sub_ecs_702_30sep14.pdf

(Notes Subject: Digital Image Processing Subject Code:

Ecs-70) Asst. Prof. Neeti Chadha CSE Department,

AKGEC Gzb page no. 5, 6.

[21] Naoko Evans, "Automated Vehicle Detection and

Classification using Acoustic and Seismic Signals",

September, 2010, page no.40-47.

[22] Ozcanli Ozbay, Ozge Can, "Recognition of Vehicles as

Changes in Satellite Imagery", 2010.

[23] "Remote sensing imagery in vegetation mapping: a

review" Plant Ecol (2008) 1 (1): 9-23. doi:

10.1093/jpe/rtm005..

[24] T. Blaschke, "Object based image analysis for remote

sensing", ISPRS Journal of Photogrammetry and Remote

Sensing, Volume 65, Issue 1, January 2010, Pages 2–16.

[25] Christof Ridder, Olaf Munkelt, and Harald Kirchner,

"Adaptive background estimation and foreground

detection using Kalman-filtering," in Proceedings of

International Conference on Recent Advances in

Mechatronics, 1995.

[26] S Huwer and H Niemann, "Adaptive change detection

for real-time surveillanace applications," Visual

Surveillance, 2000.

[27] T Kanade, R Collins, A Lipton, P Burt, and L Wixson,

"Advances in cooperative multi-sensor video

surveillance," in Proceedings of DARPA Image

Understanding Workshop, vol. 1, 1998.

[28] Chris Stauffer and W E Grimson, "Adaptive background

mixture models for real time tracking," in Proceedings of

IEEE Conf. on Computer Vision and Pattern Recognition

(CVPR)., vol. 2, 1999, pp. 246-252.

[29] Richard O Duda, Peter E Hart, and David G Stork, Pattern

Classification.: Wiley-Interscience, 2001.

[30] Oscar Firschein and Thomas M Strat, RADIUS: Image

Understanding for Imagery Intelligence.: Morgan

Kaufmann, 1997.

[31] Andres Huertas and Ramakant Nevatia, "Detecting

changes in aerial views of man-made structures," in

Proc. of International Conference on Computer Vision,

1998.

[32] Aaron J Heller, Yvan G Leclerc, and Quang-Tuan Luong,

"A framework for robust 3-d change detection," in

Proceedings for International Symposium on Remote

Sensing, SPIE., 2001.

[33] Richard I Hartley and Tushar Saxena, "The cubic rational

polynomial camera model," in DARPA Image

Understanding Workshop, 1997.

[34] Thomas Pollard, Ibrahim Eden, Joseph L Mundy, and

David Cooper, "A Volumetric Approach to Change

Detection in Satellite Images," Photogrammetric

Engineering and Remote Sensing, vol. 75, no. 12, p. to

appear, December 2009.

[35] Thomas Pollard, Comprehensive 3-d change detection

using volumetric appearance modeling, 2008.

[36] Gonzalez, R., and Woods, R., Digital Image

Processing, second edition, Prentice Hall, 2002.

[37] Lizhu Xie, Liying Wei, “Research on Vehicle Detection in

High Resolution Satellite Images”, 2013 Fourth Global

Congress on Intelligent Systems.

[38] X. Yang, G. A. Tang, F. D. Deng, EDARS Experiment

Tutorial of Remote Sensing Image Processing. Beijing,

CHN: Science press, 2009.

[39] X. L. Tian, “An Algorithm Based on the Satellite for Image

Enhancement,” COMPUTER KNOWLEDGE AND

TECHNOLOGY, vol. 11, pp. 315–317, Feb. 2008.

[40] Bhabatosh Chanda and Dwijest Dutta Majumder, 2002,

Digital Image Processing and Analysis.

[41] http://elearning.vtu.ac.in/17/e-

Notes/DIP/Unit%205%20&%206%20-%20SH.pdf.

[42] Otsu, Nobuyuki, “A threshold selection method from

gray-level histograms”, IEEE transactions on systems,

man, and cybernetics, vol. SMC9, no.1, January 1979,

pp.62-66

[43] The MathWorks, Inc., Image Processing Toolbox, URL:

http://www.mathworks.com/access/helpdesk/help/tool

box/images/

[44] Lizhu Xie, Liying Wei. “Research on Vehicle Detection in

High Resolution Satellite Images”, 2013 Fourth Global

Congress on Intelligent Systems.

[45] Sumalatha Kuthadi, “DETECTION OF OBJECTS FROM

HIGH-RESOLUTION SATELLITE IMAGES", May 2005, page

no. 5-10.

[46] Rocio Alba-Flores, "Evaluation of the Use of High-

Resolution Satellite Imagery in Transportation

Applications", August 2005, page no. 18-22.

[47] http://www.eecs.berkeley.edu/~fateman/kathey/node3

.html.

[48] http://en.wikipedia.org/wiki/K-means_clustering.

International Journal on Future Revolution in Computer Science & Communication Engineering

Volume: 1 Issue: 1 16 – 28

_____________________________________________________________________________________

28 IJFRCSCE | June 2015, Available @ http://www.ijfrcsce.org _______________________________________________________________________________________