Hybrid Features based on Zernike Moments and Single Value Decomposition (SVD) for Vehicle Classification System

DOI : 10.17577/IJERTV5IS030403

Download Full-Text PDF Cite this Publication

Text Only Version

Hybrid Features based on Zernike Moments and Single Value Decomposition (SVD) for Vehicle Classification System

Jayadurga R Dr. Gunasundari R

Research Scholar, Associate Professor & Head (i/c), Department of Computer Science, Department of Information Technology,

Karpagam Un,iversity Eachanari Post, Karpagam University, Eachanari Post,

Coimbatore- 641 021,Tamilnadu, India Coimbatore- 641 021,Tamilnadu, India

Abstract – Vehicle identification and classification in a traffic environment containing mixed background is an important application of machine vision. The goal of this paper is to build a classifier for vehicle identification using hybrid features. The proposed work is to identify a car from a non-car amidst the mixed background taken from University of Illinois at Urbana-Champaign (UIUC) standard database. UIUC image database contains mixer of car and non-car image database. Every image is divided into equal sized small sub-block images. The Zernike features and Singular Value Decomposition (SVD) features are extracted from each sub-block of the image. The features of the vehicle objects are fed to the artificial neural classifier after normalization. The performance of the classifier is compared with various literature methods of similar work. Quantitative evaluation shows improved results of 95.6% compared with the literature papers. A critical evaluation of this approach under the proposed standards is presented.

Keywords: Neural Classifier, Zernike Moments, Singular Value Decomposition, Hybrid Feature, Vehicle Classification, Normalization.

  1. INTRODUCTION

    Vehicle identification and classification are necessary components in an artificially intelligent traffic monitoring system. Vehicle identification plays a major role in applications such as vehicle security system, traffic monitoring system, etc [1-3]. It is expected that these artificially intelligent traffic monitoring system venture onto the street of the world, thus requiring identification and classification of car objects commonly found on the road side. In reality, these vehicle classification systems face two types of problem. (i) Vehicles of same category with large variation in appearance. (ii) Vehicles with different viewing conditions like hidden cars, mixed background containing buildings, people, etc. This article makes an attempt on Hybrid feature based on Zernike and Singular Value Decomposition (SVD) for vehicle

    Vehicle classification is a major area where researchers design computational systems that can identify and classify vehicles automatically. Vehicle classification has been a

    focus of investigation over last decades [8-10]. Agarwal et

    al. [7] proposed a new approach to vehicle classification that makes use of a sparse, part-based representation model. This proposed work gives a promising result in the identification of vehicles from a group of non-vehicle category. Nagarajan and Balasubramanie [8] have proposed their work based on wavelet features towards object identification and classification with mixed background. Nagarajan and Balasubramanie [9]-[11] have presented their work based on moment invariant, statistical and spectral features to identify the vehicles with mixed background respectively. Devendran et. al. [12] proposes an SVD based features for classifying the natural scenes in real world environment. Selvan and Ramakrishnan [13] introduced a new work for image texture classification based on wavelet and singular value decomposition models. Roman W. Swiniarski and Larry Hargis [14] describes an application of rough sets model which includes SVD features used for artificial neural-network- based texture images recognition.

  2. ZERNIKE MOMENT FEATURES Zernike moments are a set of complex polynomials

    Vnm (x, y) which form a complete orthogonal set over

    the unit disk of x2 y2 1, in polar coordinates. These polynomials are of the form given in Equation (1).

    Vnm (x, y) Vnm (r, ) Rnm (r) exp( jm ) (1)

    Where n is positive integer or zero and m is an integer subject to constraints n m is even and m n .

    classification. The derived Zernike and Singular Value

    r x2 y 2 is the length of the vector from the origin

    Decomposition (SVD) features from various images are

    normalized and fed to the artificial neural classifier. The vehicle of interest being a car and non-car images are

    to the pixel (x, y) .

    arctan( y )

    x

    is the angle

    identified and classified.

    between the vector r and x axis in counter clockwise

    direction. Rnm (r) is a radial polynomial defined in

    Equation (2).

    R (r)

    (n m ) / 2

    (1)s (n s)!

    r n2s

    (2)

    Thus, the magnitudes of the moments of

    nm

    s0

    n m s!

    n m

    • s

    • s

    • s

    • s

    ! !

    zernike can be used as RST (Rotation, Scaling &

    2 2

    Transformation) invariant image features. 36 zernike

    The two-dimensional zernike moment of order n with

    repetition m for function f (x, y) is defined in Equation (3).

    features are extracted for every block of an image.

  3. SINGULAR VALUE DECOMPOSITION (SVD)

Znm

n 1

f (x, y) V * (x, y)dxdy

(3)

SVD and PCA are the two of the major tools for data

analysis, reduction in dimensionality, data processing and

unitdisk nm

unitdisk nm

nm

nm

Where V *

(x, y) Vnm

(x, y)

compression. SVD works with the principle of data matrix. The m x n data matrix X is decomposed as given in

To compute the zernike moment of a digital image, It is required to change the integrals with summations as given

Equation (6).

X=USVT (6)

in the Equation (4).

Where U is m x n, V is n x n and S is m x n. The

A n 1 f (x, y)V * (x, y),

(4)

diagonal entries of S are known as singular values. While

nm nm

x y

Where x2 y2 1.

The defined features of zernike moments are only invariant to rotation (R). To achieve scale(S) and translation(T) invariance, the image needs to be normalized first by using the regular zernike moments.

The translation(T) invariance is achieved by translating

Eigen decomposition is defined only for square matrices. SVD is defined for rectangular matrices too. SVD is a generalization of Eigen decomposition. The column of U corresponding to the singular values form a basis for the column space of X and those of V form the basis for the row space of X. If the input images form the rows of X, then the columns of V form a basis of that space. These basis vectors can also be ordered according to their importance as given by the singular values. If the images are of size (20×20), the feature set is of size (1×20). Hence,

the original image

f (x, y) to

20 features are extracted from every samples using SVD.

f (x x, y y),where x m10

m00

and

y m01 .

m00

IV VEHICLE CLASSIFIER USING NEURAL NETWORKS

In other words, the original images center is moved to

m

m

the centroid before the zernike moments calculation. Scale invariant is achieved by enlarging or reducing each shape

Artificial Neural Network (ANN) classifier is built with back-propagation algorithm [10]-[11] that learns to classify a vehicle image as a car or non-car image. The number

p>so that the 0th regular moment

' equals to the total

of input nodes to the ANN is equal to the dimension of the feature space obtained from the hybrid features. The

00

00

number of shape pixels in the image, for a scaled image

f ( x, y), its regular moments

number of output nodes is usually determined by the application [15] which is 1 (either Yes/No) where, a

threshold value nearer to 1 represents Car Image and a

m

m

pq

pq

pq

pq

' pq2 m ,

mpq is the moments of

value nearer to 0 represents Non Car Image. The neural

f (x, y).

Being objective is to make

m

m

m

m

00

00

00

00

' ,

m

m

,

,

let

.

00

00

m00

classifier is trained with different choices for the number of hidden layer. The final architecture is chosen with single hidden layer shown in Figure 1 that results with better performance.

Substituting

m00

into ' ' 2

The connections carry the outputs of a layer to the input

, m00

is obtained.

of the next layer have a weight associated with them. The node outputs are multiplied by these weights before

The features of the zernike moments is rotational invariance. If f (x, y) is rotated by an angle , then the

reaching the inputs of the next layer. The output neuron (7) will be representing the existence of a particular class of object.

zernike moment Znm of the rotated image is obtained by

Equation (5).

Z

Z

Z

Z

'

'

nm nme jm

(5)

Figure 1. The Three Layered Neural Architecture

UIUC Car database with Image Size 100 x 40

UIUC Car database with Image Size 100 x 40

Image Block Size

METHODE- I

10 Blocks of

N l 1

l l

l 1

(7)

Size 20 X 20 Each

O j k f w jmO

m 0 m

V. PROPOSED WORK

This paper addresses the issues to classify vehicles containing side views of cars amidst mixed background. The vehicles of interest to be classified are car images and non-car images taken from University of Illinois at Urbana-Champaign (UIUC) database. The image data set consists of 1000 real images for training and testing having 500 in each class. The sizes of the images are uniform with the dimension 100×40 pixels. The proposed framework consists of 10 Squared Blocks of size 20×20 each.

Thirty six zernike features are extracted from each block as mentioned in the previous section using Equation

(1) to Equation (5). The zernike features are calculated from each squared single block of the sub-image. A total of 360 (36 features x 10 blocks) features are extracted from a single image.

Twenty SVD features are calculated from each single squared block of the sub-image. A total of 200 (20 features x 10 blocks) SVD features are extracted from a single image Equation (6).

Zernike and SVD features put together 560 (360 Zernike + 200 SVD) data are normalization using equation (8). Data normalization returns the deviation of each column of D from its mean normalized by its standard deviation. This is known as the Zscore of D. For a column vector V, Z score is calculated from equation (8). This process improves the performance of the neural classifier. The overall flow of the framework is shown in Figure 2.

Z=(Vmean(V))/std(V) (8)

VI .IMPLEMENTATION

The proposed method was trained with different kinds of cars vehicle against a variety of mixed background of positive class. The negative images with natural scenes, buildings, and road views are also used for training. The training is done with 400 images (200 positive and 200 negative) against all the three methods. The testing of images are done with 1000 images (500 positive and 500 negative) taken from the UIUC image database [17].

Hybrid of Zernike and SVD features

Hybrid of Zernike and SVD features

Data Normalization

Data Normalization

Building Neural classifier ANN

Building Neural classifier ANN

Result and analysis

Result and analysis

Figure 2. The proposed hybrid (Zernike + SVD) framework for vehicle classification

Figure 3. The performance of neural network training for 10 Blocks of size 20×20 each.

The feed-forward network for learning is done for 10 blocks of size 20×20. The input nodes is a hybrid based on Zernike and SVD features of size 560. Optimal structure validation is done with several repeated experiments and the structure given below leads to better results. Thus the optimal structure (Figure 1) of the neural classifier is 560-

2-1. The Performance graph of the neural classifier is shown in Figure 3.

VII DISCUSSION

In object classification problem, the four quantities of results category are given below.

  1. True Positive (TP): Classify a car image into class of cars.

  2. True Negative (TN): Misclassify a car image into class of Non-cars.

  3. False Positive (FP): Classify a non-car image into class of non-cars.

  4. False Negative (FN): Misclassify a non-car image into class of cars.

The objective of any classification is to maximize the number of correct classification denoted by True Positive Rate (TPR) and False Positive Rate (FPR) where by minimizing the wrong classification denoted by True Negative Rate (TNR) and False Negative Rate (FNR).

Figure 4. Sample results of the vehicle classifier of the category car images with mixed background.

It is evident from Table 1 that the proposed method has the highest overall classification accuracy of 95.6% compared to the literature methods [9-11]. The proposed work is compared with the work in the literature shown in Figure 6. The proposed work gives a significant improvement in classification accuracy. The novelty of the proposed work is that the input images are not pre- processed. The mixed backgrounds are not removed using background removal method as found in the literature [10-

12].

TPR Number of true positive TP

Total no.of positiveindata set nP

TNR Number of truenegative TN

Total no.of negative in data set nN

FPR Number of false positive FP

Total no.of positive in data set nP

FNR

FNR

Number of falsenegative FN

Total no.of negative in data set nF

(9)

(10)

(11)

(12)

The testing samples are 500 for positive (nP) and 500 for negative (nN) respectively. Most classification algorithm includes a threshold parameter for classification accuracy which can be varied to lie at different trade-off points between correct and false classification. The comparison of experimental methods for the proposed methods is shown in Table I which is obtained with an activation threshold value of 0.7. Classified images of category car and non-car objects as resultant sample images are shown below in the Figure 4 and Figure 5 respectively.

Figure 5. Sample results of the vehicle classifier of the category non-car images containing trees, road view, bike, wall, buildings

and persons.

Threshold for classification : 0.7

Classifying Positive Images (Car Images)

Classifying Negative Images (Non-Car Images)

TPR

TNR

FPR

FNR

10 Blocks of size 20×20 each

93.6%

6.4%

97.6%

2.4%

Overall Classification Accuray (TPR+FPR)/2 is 95.6%

Table 1. Comparison of Experimental Methods

Figure 6. Comparison of proposed work with the previous Literature.

VIII CONCLUSION

Thus an attempt is made to build a system that classifies the vehicle amidst mixed background is achieved to a certain extent. The novelty of this paper is that the input features are the hybrid of two different literatures namely Zernike feature and SVD features. Thus the goal is to classify vehicle objects of real-world images containing side views of cars images with mixed background with that of non-car images with natural scenes is presented. Further work extension can be made to improve the performance of the classifier system with various hybrid feature extraction methods.

ACKNOWLEDGEMENTS

The authors would like to thank the Research & Development department of Karpagam University. They would also like to thank their management for the constant support towards Research and Development activities.

REFERENCES

  1. B. Nagarajan and V. Devendran, Vehicle Classification under Cluttered Background and Mild Occlusion Using Zernike Features, Elsevier Procedia Engineering, International Conference on Communication Technology and System Design, 2012. 30, 201-209.

  2. B. Nagarajan and V. Devendran, Singular Value Decomposition based for Vehicle Classification under Cluttered Background and Mild Occlusion, Elsevier Procedia Engineering, International Conference on Communication Technology and System Design, 2012. 30, 529-534.

  3. B. Nagarajan and P. Balasubramanie, Hybrid Feature based Object Classification with Cluttered Background Combining Statistical and Central Moment Textures, ICGST- Graphic Vision and Image Processing (GVIP) Journal, June 2009, Vol. 9, Issue 3, 1-7.

  4. Hsieh J.W., Yu S. H., Chen Y. S. and Hu W. F. Automatic Traffic Surveillance System for Vehicle Tracking and Classification, IEEE Transansaction on Intelligent Transport System, 2006. 7(2), 175-187.

  5. Shan Y., Sawhney H. S. and Kumar R. Vehicle Identification between Non-Overlapping Cameras without Direct Feature Matching, Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV05), 2005. Digital Object Identifier: 10.1109/ICCV.2005.247

  6. Sun Z., Bebis G. and Miller R, Monocular Precrash Vehicle Detection : Features and Classifiers, IEEE Transaction on Image Processing, 2006, 15, 2019-2034.

  7. Agarwal S., A. Awan and D. Roth. Learning to Detect Objects in Images via a Sparse, Part-Based Representation, IEEE Transaction on Pattern Analysis and Machine Intelligent, 2004, 26(11), 1475- 1490.

  8. Nagarajan B. and P. Balasubramanie, Wavelet feature based Neural Classifier system for Object Classification with Complex Background, Proceeding International Conference on Computational Intelligence and Multimedia Applications (ICCIMA07), IEEE CS Press, 2007, 1, 302-307.

  9. Nagarajan B. and P. Balasubramanie, Object Classification in Static Images with Cluttered Background Using Statistical Feature Based Neural Classifier, Asian Journal of Information Technology, 2008, 7(4), 162-167.

  10. Nagarajan B. and P. Balasubramanie, Neural Classifier System for Object Classification with Cluttered Background Using Invariant Moment Features, International Journal of Soft Computing, 2008, 3(4), 302-307.

  11. Nagarajan B. and P. Balasubramanie, Neural Classifier for Object Classification with Cluttered Background Using Spectral Texture Based Features, Journal of Artificial Intelligence, 2008, 1(2), 61-69.

  12. Devendran V, A. Chitra, Hemalatha Thiagarajan and Amitabh Wahi, Singular Value Decomposition based Features for Classification of Scenes in Real World Environments, National Journal of Technology, Mar 2009, Vol. 5, No. 1, 38-46.

  13. Selvan S., and S. Ramakrishnan, SVD-Based Modeling for Image Texture Classification Using Wavelet Transformation, IEEE Transaction on Image Processing, Nov. 2007, Vol. 16, No. 11, 2688- 2696.

  14. Roman W. Swiniarski and Larry Hargis, Rough sets as a front end of neural-networks texture classifiers, Journal of Neurocomputing, Feb 2001, Vol. 36, No. 1-4, 85-102.

  15. Khotanzand A. and C. Chung, Application of MultiLayer Perceptron Neural Networks to Vision Problem,

    Neural Computing & Applications, Springer-Verlag London Limited, 1998, 249-259.

  16. B.Yegnanarayana, Artificial Neural Networks, 1999, Prentice-Hall of India, New Delhi.

  17. Agarwal and Roth, UIUC car dataset, 2002, Available : http://L2r.cs.uiuc.edu/`cogcomp/Data/Car.

Leave a Reply