Continuous rotation invariant features for gradient-based texture classification

Continuous rotation invariant features for gradient-based texture classification

Computer Vision and Image Understanding xxx (2014) xxx–xxx Contents lists available at ScienceDirect Computer Vision and Image Understanding journal...

4MB Sizes 4 Downloads 177 Views

Computer Vision and Image Understanding xxx (2014) xxx–xxx

Contents lists available at ScienceDirect

Computer Vision and Image Understanding journal homepage: www.elsevier.com/locate/cviu

Continuous rotation invariant features for gradient-based texture classification q Kazim Hanbay a,⇑, Nuh Alpaslan b, Muhammed Fatih Talu b, Davut Hanbay b, Ali Karci b, Adnan Fatih Kocamaz b a b

Bingol University, Department of Informatics, Turkey Inonu University, Department of Computer Engineering, Turkey

a r t i c l e

i n f o

Article history: Received 1 March 2014 Accepted 14 October 2014 Available online xxxx Keywords: HOG CoHOG Hessian matrix Eigen analysis Rotation invariance Texture classification

a b s t r a c t Extracting rotation invariant features is a valuable technique for the effective classification of rotation invariant texture. The Histograms of Oriented Gradients (HOG) algorithm has been proved to be theoretically simple, and has been applied in many areas. Also, the co-occurrence HOG (CoHOG) algorithm provides a unified description including both statistical and differential properties of a texture patch. However, HOG and CoHOG have some shortcomings: they discard some important texture information and are not invariant to rotation. In this paper, based on the original HOG and CoHOG algorithms, four novel feature extraction methods are proposed. The first method uses Gaussian derivative filters named GDF-HOG. The second and the third methods use eigenvalues of the Hessian matrix named Eig(Hess)HOG and Eig(Hess)-CoHOG, respectively. The fourth method exploits the Gaussian and means curvatures to calculate curvatures of the image surface named GM-CoHOG. We have empirically shown that the proposed novel extended HOG and CoHOG methods provide useful information for rotation invariance. The classification results are compared with original HOG and CoHOG algorithms methods on the CUReT, KTH-TIPS, KTH-TIPS2-a and UIUC datasets show that proposed four methods achieve best classification result on all datasets. In addition, we make a comparison with several well-known descriptors. The experiments of rotation invariant analysis are carried out on the Brodatz dataset, and promising results are obtained from those experiments. Ó 2014 Elsevier Inc. All rights reserved.

1. Introduction Texture is an important characteristic of the appearance of objects and is a powerful visual cue, used in describing and recognizing object surfaces [1]. Texture analysis plays an important role in image processing, pattern recognition, and computer vision [2–8]. Texture classification methods usually consist of two steps of feature extraction and classification. Feature extraction involves simplifying the amount of resources required to describe a large set of data accurately. To enhance the overall quality of texture classification, both the quality of the texture features and the quality of the classification algorithm must be improved [9–14]. There has been intensive research in developing robust features for texture classification with strong invariance to rotation, scale, translation, illumination changes [15–23]. Rotation invariant feature extraction is a difficult problem, thus many algorithms were proposed to achieve the rotation invariance [24,25]. q

This paper has been recommended for acceptance by Yasutaka Furukawa.

⇑ Corresponding author.

E-mail address: [email protected] (K. Hanbay).

The pioneer works to achieve rotation-invariant texture classification include generalized co-occurrence matrices (GCM) [26], polarograms [27], texture anisotropy [28], the methods based on Markov random field (MRF) [29] and autoregressive model. The wavelet based algorithms achieved effective classification performance [30–37]. Recently, the statistical based approaches have attracted considerable attention [38–40]. However, many of these approaches achieve the rotation invariance by shifting the discrete orientations. For example, the method of local binary pattern (LBP) [18] is proposed to achieve rotation invariance [41]. The gradient based features such as edges or orientation angles are widely used as feature descriptors in image processing. In order to identify objects in images effectively, gradient based edge features have been developed, which are edge orientation histogram [42], Histograms of Oriented Gradients (HOG) [43,44], co-occurrence HOG (CoHOG) [45], multilevel edge energy features [46], shapelets [47], and edge density [48]. The HOG method distributes the gradients into several orientation bins. HOG encapsulates changes in the magnitude and orientation of contrast over a grid of small image patches. HOG features have shown satisfactory performance in their ability to recognize a range of different object

http://dx.doi.org/10.1016/j.cviu.2014.10.004 1077-3142/Ó 2014 Elsevier Inc. All rights reserved.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

2

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

types including natural objects as well as artificial objects. CoHOG (Co-Occurrence Histograms of Oriented Gradients), an extension of HOG to represent the spatial relationship between gradient orientations, has been proposed and its effectiveness for pedestrian detection, human detection and medical image analysis has been demonstrated in [49–51]. According to a review [52], the extraction of basic features in images is based on two mathematical concepts: differential geometry and scale-space theory. The differential geometry approach uses the assumption that features can be obtained from the image based on local anisotropic variations of pixel intensities. This concept is strong and effective. Recently, the differential feature extraction approaches have attracted more attentions [53,54]. Among the different feature extraction methods, gradient-based methods were widely used in the past. These methods are effective in defining and describing significant image features. Hessian matrix information is a robust differential method and it has been widely used in many publications to extract the image features [55,56]. Compared with the conventional gradient, the Hessian matrix and its Eigen analysis are more reliable and robust in revealing the fundamental directions in data. In this study, we propose four novel methods to improve the classification performance of the HOG and CoHOG algorithms. The proposed four methods are based on Gaussian derivatives filters and Hessian matrix. Instead of using the conventional gradient operator in HOG and CoHOG algorithms, the second-order partial derivatives in Gaussian derivatives filters and Hessian matrix are more proper and stable to calculate the intensity and texture variations of image surface. The rest of this paper is organized as follows. In Section 2, we will give a review of the original HOG and CoHOG algorithms. Section 3 presents the proposed new HOG and CoHOG algorithms based on Gaussian derivatives filters, Hessian matrix and Gaussian–mean curvatures. In Section 4, we test the performance of novel feature extraction algorithms on four standard texture datasets and discuss the effect of the normalization step on the classification performance. In Section 5, a series of rotation analysis experiments are performed. Furthermore, the characteristics of proposed descriptors are discussed in detail. The comparison results with the state-of-the-art the texture classification methods performed on Brodatz and UIUC datasets are shown in Section 6. The conclusions are given in Section 7. 2. Related works

f x ðx; yÞ ¼ Iðx þ 1; yÞ  Iðx  1; yÞ f x ðx; yÞ ¼ Iðx; y þ 1Þ  Iðx; y  1Þ

where fx and fy denotes x and y components of image gradient respectively. I(x, y) denotes the pixel intensity at position (x, y). The magnitude and orientation is computed in Eqs. (2) and (3):

qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi f x ðx; yÞ2 þ f y ðx; yÞ2   f y ðx; yÞ hðx; yÞ ¼ tan1 f x ðx; yÞ

mðx; yÞ ¼

This section gives an overview of the HOG feature extraction process. The basic concepts of the HOG are the local object appearance and shape, which can be characterized by the distribution of the local intensity gradients or edge directions [57,58]. The gradients orientations are strong against lighting changes since the forming histogram provides rotational invariance. For each key point, a local HOG descriptor from a block is computed. The block size is not restricted to construct an extensive set of texture features, which allow extracting high-discriminated features in order to improve classification accuracy and reduce computational time of classification algorithms [59,60]. HOG is a window based algorithm computed local to a detected interest point. The window is centered upon the point of interest and divided into a regular square grid (n  n) [43,44]. This method consists of several steps. First, the grayscale image was filtered to obtain x and y derivatives of pixels. The filter kernels were used to compute discrete derivative in the x and y direction. The gradient values at every image pixel were computed as follows:

ð2Þ ð3Þ

Second, the image intensity gradients are divided into layers based on their orientation. The original HOG descriptor uses unsigned gradients in conjunction with 9 bins (a bin corresponds to 20°) to construct the histograms of oriented gradients. Therefore, there are 9 layers of orientated gradient. Finally, orientation histogram of every cell and larger spatial blocks n  m are normalized. To normalize the cells’ orientation histograms, they should be grouped into blocks. Since a cell has k orientations, the feature dimension of each block is n  m  k for each block. v denotes feature vector in a block h(i,j) denotes unnormalized histogram of the cell in the position (i, j) in a block. Although there are three different methods for block normalization, L1-Norm normalization is implemented as: 0

hði;jÞ 0 hði;jÞ ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðe ¼ 1Þ kv k1 þ e

ð4Þ

where e is the small constant [43]. Here, e is set to 1 empirically. 2.2. Co-Occurrence Histograms of Oriented Gradients (CoHOG) The CoHOG feature descriptor is based on a co-occurrence matrix which is obtained from a 2D histogram of pairs of gradient orientations [45]. It performs on grayscale images. The co-occurence matrix expresses the distribution of gradient orientations at given offset over an image as shown in Fig. 1. The combinations of neighbor gradient orientations provide reliable features of objects in images and this is very advantageous for object classification problems. The co-occurrence matrix C is obtained from n  m image of gradient orientations, and formulated in Eq. (5);

C i;j ¼

n1 X m1 X 1 if Iðp; qÞ ¼ i and Iðp þ x; q þ yÞ ¼ j p¼0 q¼0

2.1. Histograms of Oriented Gradients (HOG)

ð1Þ

0

otherwise

ð5Þ

where I indicates a gradient orientation image, i and j indicates gradient orientations and x, y denotes vertical and horizontal offsets. The gradient orientations from I are calculated in Eq. (6);

h ¼ tan1

v  h

ð6Þ

where v and h are the vertical and the horizontal components of gradient calculated by appropriate filters. Then, the orientations in the range (0, 2p) are quantized into eight labels. Each label is used for representing an orientation. Thus, the size of the co-occurrence matrix C becomes 8  8. Six offsets are used in experiments. The cooccurrence matrix contains information on the local textures by using short-range offsets and the global textures by using longrange offsets [45,61]. The co-occurrence matrices are computed for each tiled regions with all offsets. Hence, the number of CoHOG descriptor features is m  n  d2 where d is the number of gradient orientation bins, m is the number of tiled regions and n is the number of offsets. Finally, the CoHOG descriptor is determined as a vector by concatenating the components of all the co-occurrence matrices. The size of the original CoHOG descriptor is 2  2  6  82 = 1536.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

Tiled Regions

Offsets

3

C

Fig. 1. Process diagram of CoHOG algorithm.

3. The proposed novel HOG and CoHOG methods In this section, we introduce four novel derivative methods to improve the HOG and CoHOG feature extraction algorithms. In original HOG algorithm, the gradient orientations of the pixels are calculated by using gradient information of the input image. In original CoHOG algorithm, in the same way, the labeling process of the pixels is performed in the specified offset by using pixel orientations obtained from gradient information. This labeling process directly affects the quality and distinctiveness of the information contained in the co-occurrence matrix created in the next step. Therefore, the calculation of gradient information is the most basic and important step in both algorithms. In this study, we propose an efficient global derivative scheme that uses Gaussian derivative filter and Hessian matrix for feature extraction. Thus, by obtaining distinctive and robust characteristics of the images, significant improvement has been provided in the classification performances of algorithms. In this paper, the first gradient calculation method proposed to improve HOG algorithm uses x–y separable Gaussian derivative filters. In addition, the other proposed method uses the eigenvalues of the Hessian matrix. The proposed first descriptor to strengthen the gradient step of CoHOG algorithm uses eigenvalues of the Hessian matrix. Instead of directly using the responses of gradient, our second descriptor uses the Gaussian curvature and mean curvature informations. Furthermore, for the developed new CoHOG algorithm, a new formulation is used in labeling process. For brevity, novel HOG algorithm using proposed separable Gaussian derivative filters is named as GDF-HOG. Also, novel HOG algorithm using the eigenvalues of the Hessian matrix is named as the Eig(Hess)-HOG. The proposed novel Co-HOG algorithm using the eigenvalues of the Hessian matrix is named as the Eig(Hess)-CoHOG. Finally, the novel Co-HOG algorithm using Gaussian and mean curvature information is named as GM-CoHOG.

3.1. The proposed GDF-HOG algorithm

change in terms of the definition of uniform histogram. To overcome the most prominent shortcoming of the definition of ‘‘uniform’’ histogram in HOG, we propose a novel extension in which the local texture patterns are subjected to further treatment and then computed in Gaussian derivative filters way. We can find the continuous rotation invariant features via Gaussian function. Therefore, we employ the Gaussian derivative filters approach to represent and classify texture images. Since the rotation, scaling and translation operations are each a linear transformation, they retain the shape of the image. So, even though the coordinates of the image change, the exact state of shape does not change. Gaussian function is a continuous and linear-dependent function. Thus this function can be used in the calculation of first and secondorder rotation invariant derivative. The Gaussian first and second derivative filters could be rotated at any angle by linear combination of two basis filters [62]. The gradient computation which is calculated via the Gaussian function and two-dimensional convolution provides more prominent texture and intensity information than conventional gradient. Thus the Gaussian derivative filters are usually an appropriate model to extract the basic features of the texture patterns. The Gaussian function can be formulated as:

Gðx; yÞ ¼ e

ðx2 þy2 Þ r2

ð7Þ

The primary first-order differential quantity for an image is the gradient. Gradient is a 2-D vector quantity. It has both direction and magnitude informations which vary at each point [63]. Image derivatives can be calculated as follows:

rI ¼



I  Gx



I  Gy

ð8Þ

where ⁄ denotes the convolution, Gx and Gy are the first partial derivatives of G with regard to x and y respectively. The responses of the oriented Gaussian first derivative filter Gh in Eq. (9):

Ih ¼ I  Gh ¼ I  ðcosðhÞGx þ sinðhÞGy Þ HOG is a well-known feature descriptor which computes the features through a gradient orientation histogram within the local region. The original HOG algorithm uses a gradient computation method which is based on calculation difference of neighboring pixels. The primary weak point of original HOG is that it cannot describe the characteristics of textures efficiently and distinctively. Another weak point of HOG is that it is mathematically weak and sensitive to noise since the label of a local histogram is easy to

¼ cosðhÞI  Gx þ sinðhÞI  Gy

ð9Þ

Since the two basic filters are x–y separable, the response of each filter could be calculated by twice one-dimensional convolutions [41]. Also, the computation cost of two-dimensional convolution is more than twice one-dimensional convolutions. The four basic one-dimensional Gaussian derivative filters are given by

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

4

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

f1 ¼

2t

r

2

t2

e r2 ;

t2

f 2 ¼ e r2 ; f3 ¼ f4 ¼

 2t 2t2

r2 2t

r2

r2

 2 t  1 e r2 ;

ð10Þ

t2

e r2



where t controls the length of the derivative filter. The generation step of five basic Gaussian filter responses is shown in Table 1. In this section, we focus on novel gradient computation method for the HOG algorithm. So, we only used Gx and Gy filters for novel gradient computation. The other filters are used for the second-order derivative calculation. The standard deviation (r) of the Gaussian derivatives filters is selected as 1. For the calculation of the gradient in the proposed GDF-HOG algorithm, we used Gx and Gy filters based on the convolution with the derivatives of the Gaussian according to the Eq. (10). In GDFHOG algorithm, gradient information is calculated as follows:

Ix ¼ I  Gx Iy ¼ I  Gy

ð11Þ

From Eq. (11), we can obtain the gradient magnitude of image I with regard to h in Eq. (12): Ih ¼ cosðhÞI  Gx þ sinðhÞI  Gy qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ¼ ðI  Gx Þ2 þ ðI  Gy Þ2 cosh

I  Gx

ðI  Gx Þ2 þ ðI  Gy Þ2 qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ¼ ðI  Gx Þ2 þ ðI  Gy Þ2 sinðh þ uÞ

þ sinh

!

I  Gy ðI  Gx Þ2 þ ðI  Gy Þ2

ð12Þ

where u ¼

the information about maximum, minimum and saddle points of the function have been obtained by looking at the minor determinants of Hessian matrix. The Hessian of an image is defined as second-order partial derivative matrix of gray level image. The Hessian matrix H, as a real-valued matrix, has real-valued eigenvalues. Hessian matrix of one point in a gray image I for a scale r is computed as

IG arctan IGyx .

Thus, when h ¼

p  u; 2

gradient magnitude of

input image Ih obtains

Imaggradient ¼

qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðI  Gx Þ2 þ ðI  Gy Þ2

ð13Þ

To calculate gradient magnitude in proposed GDF-HOG algorithm, Eq. (13) is used. It should be noted that there are major mathematical differences between classical gradient computation and the proposed gradient computation. The proposed gradient computation is invariant to image rotation at any angle. In addition, the proposed gradient method is mathematically strong and robust to noise. Classification results of the GDF-HOG obtained on various datasets are higher than the original HOG method. 3.2. The proposed Eig(Hess)-HOG and Eig(Hess)-CoHOG algorithms Instead of using the Gaussian derivative filters in original HOG and CoHOG algorithms, our second method uses the Hessian matrix to calculate the eigenvalues of image surface. In order to analyze the local behavior of an image Ix, we apply local analysis of an image by using Hessian matrix (second fundamental form represented in Appendix A). The Hessian matrix is a square and symmetric matrix, and consists of second-order partial derivatives of the function [64]. According to differential geometry concepts, Table 1 Generation step of separable two-dimensional Gaussian filters using the Gaussian one-dimensional filters. Basic filters

Filter in x

Filter in y

Gx Gy Gxx Gxy Gyy

f1 f2 f3 f4 f2

f2 f1 f2 f4 f3

Hr ðx; yÞ ¼

Dxx Dyx

Dxy Dyy



 ¼

I  Gxx I  Gxy

I  Gxy I  Gyy

ð14Þ

where ⁄ denotes the convolution, Dxx, Dyy, Dxy are the second-order derivative of the image along direction of x, y, xy respectively. Gxx, Gyy, Gxy are the second-order derivative filter of the image along direction of x, y, xy respectively. The generation of these three filters is detailed in the former section. In Eq. (14), r is implicitly included in the calculation of second-order derivatives. The Hessian matrix contains more differential information than the gradient computation. Especially, first order operations (i.e. gradient) are insufficient to describe the behavior of nonlinear functions. But the second order differential operations (i.e. Hessian) enable a more accurate analysis in detail about function curves [64]. From this viewpoint, the state-of-the-art geometrybased operators try to estimate the shape of the underlying image region by estimating second-order partial derivatives such as Laplacian and Hessian approaches [41,65]. In many state-of-theart studies, the object edges, corners, and shape information are obtained by the Hessian matrix [65,66]. The basic idea behind eigenvalue and eigenvector analysis of the Hessian matrix is to extract the principal directions and principal curvatures in image surface. Thus, the local second-order structure of the image can be examined. Since this directly gives the direction of the smallest curvature in image. The eigenvalues of the Hessian are called principal curvatures and invariant under rotation. The eigenvalues of the Hessian matrix can be defined as follows [41]:

sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðI  Gxx  I  Gyy Þ2 I  Gxx  I  Gyy k¼ þ ðI  Gxy Þ2 þ 4 2

ð15Þ

where k are the eigenvalues of the Hessian matrix. Fig. 2 presents the procedure for computing eigenvalues of the Hessian matrix at a specific Gaussian standard deviation. The Eq. (15) allows the extraction of the principal directions in image. So, the local second-order structure of the image is examined. In this paper, we focus on the novel derivative method for the HOG and CoHOG algorithms. Thus, k1 and k2 can be used instead of gradient calculation step of the original algorithms. In original HOG algorithm, gradient orientation and magnitude of the pixels are calculated by using conventional gradient. Instead of using the conventional gradient, gradient magnitude and orientations can be calculated by using the eigenvalues k1 and k2 . In proposed Eig(Hess)-HOG algorithm, the gradient magnitude is calculated as:

qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ðk1 Þ2 þ ðk1 Þ2   k2 h ¼ arctan k1

Ih

gradient

¼

ð16Þ ð17Þ

Then we calculated the gradient orientations of proposed algorithm as in Eq. (17). The obtained gradient orientations are labeled in the range (0–180°) with eight different labels as mentioned in Section 2. In the experimental analysis, the novel Eig(Hess)-CoHOG algorithm performs this labeling process for seven groups, while the original method pixels using the gradient often label eight different groups. Discarding high order derivatives can simplify

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

5

Fig. 2. Procedure for computing eigenvalues of the Hessian matrix. The original texture image is taken from the KTH-TIPS2-a dataset. The Gaussian standard deviation r = 1.

computation, but loss of high order information leads to reduction of classification accuracy. Since the first order derivatives cannot comprehensively characterize a complex texture image, more order derivatives should be included. Thus, the co-occurrence matrix in novel Eig(Hess)-CoHOG has more distinctive texture information. Owing to discriminative capability of co-occurrence matrix, the novel CoHOG algorithm reduces the size of feature vector from 1536 to 1176. The extraction of a high-dimensional feature descriptor, such as the original CoHOG and LBP, is time consuming, especially when the operation is applied to a whole image with an exhaustive block-based search [18,41]. The novel CoHOG approach helps reduce the computational load especially when it is of interest to carry out real-time operations. Despite the lower feature vector dimensionality, the classification performance of the algorithm is substantially increased.

3.3. The proposed GM-CoHOG algorithm In Section 3.2, to improve gradient calculation of original HOG and CoHOG algorithms, the Hessian matrix is used instead of gradient computation. These second-order differential operations enable us to obtain distinctive information of pixel and texture orientation about the image surfaces. In addition, Hessian matrix and its eigenvalues give us edge informations in the image as well as corner point details of objects in the image. Edge and corner points

are important to obtain important image features accurately. Therefore, very high classification ratio is obtained in the developed Eig(Hess)-HOG and Eig(Hess)-CoHOG algorithms. In rotation analysis on Brodatz dataset, obtained positive results by the use of Hessian matrix and eigenvalues revealed the strength of secondorder differential analysis techniques. The success of second-order differential analysis has urged us to use different second-order differential analyses. Starting from this point, the Gaussian curvature and mean curvature data used in the calculation of the curvature of the functions in differential geometry have been integrated into the gradient computation step of the original CoHOG algorithm. In the original CoHOG algorithm, the gradient orientations are calculated based on the conventional gradient information. More robust gradient orientations are obtained with the use of Gaussian and mean curvature of image. In Section 3.1, first and second-order partial derivatives of the image are calculated. Since the Gaussian derivatives filters could be treated as special difference operators, the first and second partial derivatives of the image I are calculated as:

Ix ¼ I  Gx Iy ¼ I  Gy Ixx ¼ I  Gxx

ð18Þ

Ixy ¼ I  Gxy Iyy ¼ I  Gyy

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

6

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

where Ix and Iy are the first partial derivatives of I with regard to x and y respectively, and Ixx, Iyy and Ixy are the second partial derivatives of I with regard to x and y. The Gaussian curvature K and the mean curvature H of the image surface I can be calculated in Eqs. (19) and (20) [64]:





Ixx ð1 þ I2y Þ þ Iyy ð1 þ I2x Þ þ 2Ix Iy Ixy  32 1 þ I2x þ I2y Ixx Iyy  I2xy 2

ð1 þ I2x þ I2y Þ

ð19Þ

ð20Þ

Labeling process of the pixels is performed by using the Gaussian and mean curvature data. In the labeling process, Eq. (21) is used.

Ol ¼ 1 þ 3ð1 þ signðHÞ þ 1 þ signðKÞÞ

ð21Þ

Through these detailed and advanced geometric derivative calculations, the most detailed pixel analysis is made by calculating the Gaussian and mean curvature data of each image. Thanks to the two curvature information of each pixel, even the most distinguishing characteristics of the two neighboring pixels could be detected. The proposed method has the best discriminating ability for texture images. The classification results of GM-CoHOG algorithm reveal the power of the proposed method. Fig. 3 presents the procedure for computing Gaussian and mean curvatures of an image at a specific Gaussian standard deviation. 4. Experiments on the standard material datasets In this section, the novel HOG and CoHOG methods using our proposed four novel derivative computation model are tested and compared with the original HOG and CoHOG methods in terms of classification rate and feature vector size. Our algorithms are coded by Matlab. All the experiments in this section are run in a release version of our code on an Intel(R) Core(TM) 2 Duo CPU 2.93 GHz, 8 GB RAM personal computer. 4.1. Dissimilarity metric The dissimilarity of training and test histogram is a test of goodness-of-fit, which can be measured with a nonparametric statistic test. There are many metrics for measuring the fit between two image histograms, such as Euclidean distance, log-likelihood ratio, and chi-square statistic. In this paper, a test sample S is assigned to the class of model M that minimizes the chi-square distance:

DðS; MÞ ¼

L X ðSk  M k Þ2 k¼1

Sk þ M k

ð21Þ

where L is the number of bins. Sk and Mk are the values of the sample and model images at the k th bin, respectively. To evaluate the effectiveness of the proposed new HOG and CoHOG methods, we use the nearest neighborhood (NN) classifier with chi-square distance kernel rather than other classifiers such as artificial neural network and support vector machines which have been obtained to produce perfect results [41,67,68]. 4.2. Datasets We tested the new HOG and CoHOG methods by classifying images from CUReT dataset [69], KTH-TIPS dataset [17,70], KTHTIPS2-a [70,71], UIUC [15] and Brodatz datasets [72]. Brodatz dataset is only used to show the rotation invariance properties of the proposed methods.

The CUReT dataset contains 61 different texture classes, and each texture class includes 205 images of a physical texture sample photographed under a range of viewing and illumination angles. For fair comparison with other HOG and CoHOG studies using CUReT dataset, the 92 images per class are chosen with the size of 200  200. Some examples of CUReT dataset are given in Fig. 4a. The KTH-TIPS dataset includes 10 different texture classes of sandpaper, crumpled aluminum foil, linen, sponge, corduroy, styrofoam, cracker, brown bread, orange peel and cotton. This dataset consists of texture images that variation in scale as well as pose and illumination. Images were captured at nine different scales spanning two octaves, viewed under three different illumination directions and different poses, thus giving a total of nine images per scale. Sample images from the KTH-TIPS dataset are given in Fig. 4b. The KTH-TIPS2-a ensures a considerable extension to previous KTH-TIPS dataset. The KTH-TIPS2 includes four physical samples of 11 different materials. The images of each sample texture have various poses, illumination conditions and scales. The images which are not 200  200 pixels in size are deleted, thus there are 4395 images in all. Sample images from the KTH-TIPS2-a dataset are given in Fig. 4c. The UIUC dataset includes 25 classes, as shown in Fig. 4d and 40 images in each texture class. The resolution of each image is 640  480. The dataset includes materials imaged under significant view-point variations. The Brodatz dataset includes 112 different textures images. It is composed of 112 grayscale images representing a large variety of natural grayscale textures. This dataset has been widely used with different levels of complexity in texture classification and texture segmentation. In this paper, a rotation invariant version of the Brodatz dataset is used for texture classification. Some examples of Brodatz datasets are given in Fig. 4e. 4.3. The effect of normalization process The normalization of texture images is a considerable pre-processing step for texture classification methods. The texture images are generally normalized to have zero means and unit standard deviations. This normalization process provides invariance to global affine transformations in the illumination intensity [19]. For all methods with the normalization step, higher classification accuracies are obtained on the CUReT, KTH-TIPS, KTHTIPS2-a, UIUC datasets. Table 2 shows the classification results for the proposed GDF-HOG, Eig(Hess)-HOG, Eig(Hess)-CoHOG and GM-CoHOG algorithms with normalization and without normalization. For all algorithms with the normalization step, higher classification accuracies are obtained on the on the CUReT, KTH-TIPS, KTHTIPS2-a and UIUC datasets. Here, following question may arise: Why does image normalization affect the classification performance positively on the four texture datasets? The question might be answered in the following explanations [41]:  The images are normalized to have unit standard deviation which adjusts the global contrast of the images to a standard level. The normalization process is only invariant to global transformation of the illumination intensity. Therefore, in the dataset in which the intensity of illumination changed locally, the impact of the normalization process on the classification accuracy may not be at the expected level. As, in KTH-TIPS2-a dataset, the illumination intensity often changes locally, positive effect of the normalization process is limited compared to other datasets.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

7

Fig. 3. Procedure for computing Gaussian and mean curvatures of a texture image. The original texture image is taken from the KTH-TIPS2-a dataset. The Gaussian standard deviation r = 1.

 The use of Gaussian local derivative filters to obtain the first and second derivative information already has the strength of local illumination invariance. For all the proposed methods, feature vectors are normalized to have an average intensity of 0 and a standard deviation of 1. This process provides local brightness and contrast invariance. Global normalization process is thus unnecessary. Different features of the dataset show that a single normalization process cannot be applied to them all. In this study, although a positive influence is obtained for the entire datasets with the normalization operation on classification accuracy, suitable normalization process should be applied for different dataset characteristics. KTH-TIPS, KTH-TIPS2-a and UIUC datasets contain materials imaged under significant pose variations, various poses, illumination conditions and scales. Therefore, the differences between the above-mentioned histogram of the images in the three datasets and histogram of the images in CUReT dataset are noticed. Fig. 5 demonstrates the texture images and their histogram distributions for the CUReT and KTH-TIPS2-a datasets. All of the texture images are from the same class in the CUReT and KTH-TIPS2-a datasets. The histograms denote that the images in the CUReT dataset usually have different mean values and standard deviations. However, images histograms from the KTHTIPS2-a dataset do not have such features. Thus, the normalization step eliminates such effects, and it provides higher classification performance on various datasets.

4.4. Comparison of classification performance In this paper, to evaluate the effects of the proposed first and second-order derivatives on the original HOG and CoHOG algorithms, a series of experiments are conducted on large and comprehensive texture datasets. The first experiment is conducted on the proposed GDF-HOG, Eig(Hess)-HOG algorithms and the original HOG algorithm. The experiments are conducted on CUReT, KTHTIPS, KTH-TIPS2-a and UIUC datasets. In order to show the best classification results, the image normalization step is adopted for the all datasets. Table 3 shows that the proposed GDF-HOG and Eig(Hess)-HOG algorithms achieve best classification results on CUReT, KTH-TIPS, KTH-TIPS2-a and UIUC datasets. The GDF-HOG algorithm uses the magnitudes of Gaussian first derivatives but achieving continuous rotation invariance. On the CUReT dataset, the classification accuracy of original HOG algorithm is 89.39%, and GDF-HOG achieves 94.52%. The development is considerable. On the KTH-TIPS dataset, the classification accuracy of original HOG algorithm is 82.75%, and the proposed Eig(Hess)-HOG achieves 95.58%. The improvement in the classification accuracy is utterly incredible. The similar results are true for KTH-TIPS2-a and UIUC datasets as well. Since the UIUC dataset contains big affine and scale variation, GDF-HOG and Eig(Hess)-HOG algorithms cannot get very high classification results. Classification results show that the pair of eigenvalues of image surface is a significant property which describes local structures distinctively and it is more robust on all datasets. Furthermore the major advantage of

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

8

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

Fig. 4. Image examples from standard datasets. (a) CUReT dataset. (b) KTH-TIPS dataset. (c) KTH-TIPS2-a dataset. (d) UIUC dataset. (e) Brodatz album.

our proposed Eig(Hess)-HOG algorithm is its continuous rotation invariance. Generally, the rotation invariant feature extraction may not achieve superior performance than the original HOG without rotation invariance. The Hessian matrix-based feature extraction achieves best classification accuracy on all datasets. According to experimental results, it is also seen that the proposed Gaussian derivative filters and eigenvalues of Hessian matrix

achieve excellent results compared with the original HOG algorithm. The second experiment is done between the novel CoHOG algorithms and the original CoHOG algorithm. Pixel labeling process of the proposed Eig(Hess)-CoHOG algorithms and the original CoHOG algorithm is made with the same formula. So, though feature vector is expected to be the same size in both algorithms, the lower

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

9

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

Table 2 Effect of image normalization on the CUReT, KTH-TIPS, KTH-TIPS2-a and UIUC datasets. I means the images are normalized and II means the normalization is omitted. The bolded values represent the best classification results. Methods

CUReT

Original HOG GD-HOG Eig(Hess)-HOG Original CoHOG Eig(Hess)-CoHOG GM-HOG

KTH-TIPS

KTH-TIPS2-a

UIUC

I (%)

II (%)

I (%)

II (%)

I (%)

II (%)

I (%)

II (%)

89.39 94.52 94.08 87.63 99.67 99.38

75.00 85.18 93.33 80.00 90.00 92.30

82.75 94.02 95.58 97.93 99.00 99.02

77.77 91.80 91.80 82.35 98.18 96.62

84.98 91.08 92.40 97.74 99.28 99.09

83.72 89.42 84.84 97.73 98.34 98.20

66.66 88.63 85.71 77.41 96.82 98.41

60.00 85.18 70.00 76.47 91.66 93.54

CUReT dataset images and their histograms

1200

1500 1200 1000 1000

400

800

800 600

500

400 200

0

500

1000

0

50

100

150

200

250

0

0

50

100

150

200

250

600

300

400

200

200

100

0

0

50

100

150

200

250

0

0

50

100

150

200

250

0

50

100

150

200

250

KTH-TIPS2-a dataset images and their histograms

600

600

500 600

500

500

400 400

400

400

300

300

300

200

200

200

200

100

100

0

0

0

50

100

150

200

250

100 0

50

100

150

200

250

0

0

50

100

150

200

250

0

Fig. 5. Examples of texture images and their histograms for the CUReT and KTH-TIPS2-a datasets.

dimensional feature vector is obtained in the proposed Eig(Hess)-CoHOG algorithm. Because images are characterized with the second-order differential analysis more detailed, co-occurrence matrix is composed of distinctive and sensitive derivative information. Therefore, the distinctive information contained in the co-occurrence matrix led to reducing in the size

of the feature vector. Another important aspect to be emphasized here is that: since Eig(Hess)-CoHOG algorithm has lowdimensional feature vector and a very high classification rate, it is suitable for real-time applications. Pixel labeling process in the proposed GM-CoHOG algorithm is performed with Gaussian and mean curvature information and a new formula that enabled

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

10

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

Table 3 Classification accuracy on the CUReT, KTH-TIPS, KTH-TIPS2-a and UIUC datasets for the proposed novel HOG algorithms. The bolded values represent the best classification results. Method

Feature size

CUReT (%)

KTH-TIPS (%)

KTH-TIPS2-a (%)

UIUC (%)

Original HOG GDF-HOG Eig(Hess)-HOG

128 128 128

89.39 94.52 94.08

82.75 94.02 95.58

84.98 91.08 92.40

66.66 88.63 85.71

Table 4 Classification accuracy on the CUReT, KTH-TIPS, KTH-TIPS2-a and UIUC datasets for the proposed novel CoHOG algorithms. The bolded values represent the best classification results. Method

Feature size

CUReT (%)

KTH-TIPS (%)

KTH-TIPS2-a (%)

UIUC (%)

Original CoHOG Eig(Hess)-CoHOG GM-CoHOG

1536 1176 1536

87.63 99.67 99.38

97.93 99.00 99.02

97.74 99.28 99.09

77.41 96.82 98.41

labeling. Feature vector size of this algorithm is in the same size with the original CoHOG algorithm and more larger than Eig(Hess)-CoHOG we proposed. However, it has the highest classification success on UIUC and KTH-TIPS datasets. The classification results are listed in Table 4. Table 4 shows that the proposed Eig(Hess)CoHOG and GM-CoHOG algorithms achieves best classification accuracies on all datasets. On the CUReT dataset, the classification accuracy of original CoHOG algorithm is 87.93%, the proposed Eig(Hess)-CoHOG achieves 99.67%, and the proposed GM-CoHOG achieves 99.38%. The highest classification rate is obtained for the proposed CoHOG algorithms. Eig(Hess)-CoHOG algorithm achieves 99.34% classification success, while the GM-CoHOG algorithm achieves 99.09% on the KTH-TIPS2-a dataset on which a lot of classification methods achieved an accuracy rate of only 70%. The fundamental challenge of the KTH-TIPS and KTH-TIPS2-a texture datasets lies on its large intra-class differences, so Eig(Hess)-CoHOG and GM-CoHOG algorithms which have powerful intra-class congregate capability could achieve best classification performance. However, the CUReT dataset has opposite properties. The texture images from the same texture class are very alike, while the texture images from different class are also similar. That means the CUReT dataset has both small inter-class and intra-class differences. Eig(Hess)-CoHOG and GM-CoHOG have presented their superiority on CUReT dataset due to their good inter-class distinguish capability. In the experiments carried out on UIUC and KTH-TIPS datasets, very high classification results have been obtained compared to the original CoHOG algorithm. In particular, it must be emphasized that; the KTH-TIPS2-a dataset is obtained under different lighting, poses and scaling. The classification success of the Eig(Hess)-CoHOG algorithm is an indication of how powerful eigenvalues of Hessian matrix are in feature extraction process. On the other hand, the excellent classification rate achieved by GM-CoHOG algorithm shows the value added to the original CoHOG algorithm by the Gaussian and mean curvature calculations. Moreover the important advantage of our proposed Eig(Hess)CoHOG and GM-CoHOG algorithms are their continuous rotation invariance. This property of both of the developed algorithms is analyzed in detail in the following sections. 5. Rotation analysis In order to show the rotation invariant performance of the proposed algorithms clearly, we carry out various implementations on three datasets based on Brodatz dataset [72]. To make an objective comparison with the other methods, the production of testing and training images are performed as described in the literature [35]. Each texture image with the size of 512  512 in Brodatz album is divided into four 256  256 non-overlapping image regions. The

center 128  128 subimage from each region is used for training process. For the generate test set, each 256  256 image region is rotated at angles of 10–160° with 10° increments and, from each rotated image, a 128  128 subimage is selected. In this way, the training set is 4  25 subimages and the test set is 4  16  25 subimages. The dataset 2 includes 60 texture of from Brodatz album also used in [14]. The 60 texture images are D01, D04, D05, D06, D08, D09, D10, D11, D15, D16, D17, D18, D19, D20, D21, D22, D23, D24, D25, D26, D27, D28, D34, D37, D46, D47, D48, D49, D50, D51, D52, D53, D55, D56, D57, D64, D65, D66, D68, D74, D75, D76, D77, D78, D81, D82, D83, D84, D85, D86, D87, D92, D93, D94, D98, D101, D103, D105, D110 and D111. The training and test sets are generated in the same way as dataset 1. Therefore, the training set is 4  60 subimages and the testing set is 4  16  60 subimages. In this way, an objective comparison is performed with other methods in the literature. The dataset 3 contains 12 Brodatz texture classes, and each texture class included seven rotation images of size 512  512 rotated by different angles (0°, 30°, 60°, 90°, 120°, 150°, and 200°). Table 5 shows the classification results for the original HOG and Eig(Hess)-HOG algorithms. The proposed GDF-HOG algorithm obtained poor performance, thus its classification results are not presented. For dataset 1, 2 and 3, the Eig(Hess)-HOG algorithm achieves remarkable classification results. Because the original HOG algorithm does not have the property of rotation invariance, it has a very low classification rate on three of the datasets. The very high classification rate of Eig(Hess)-HOG algorithm that we proposed can be explained as follows:  Eig(Hess)-HOG algorithm is a method developed by modifying the gradient calculation step of the original HOG algorithm. In this method, firstly the Hessian matrix of the image is calculated. Secondly, the k1 and k2 eigenvalues of the Hessian matrix are calculated. The main characteristic of eigenvalues of Hessian is invariant under rotation. Therefore, the Eig(Hess)-HOG algorithm developed by using the magnitude of the eigenvalues has the property of rotation invariance. Therefore, the Eig(Hess)-HOG algorithm achieved high classification accuracy in all datasets. Each image from the Brodatz dataset is rotated Table 5 Classification accuracy on the rotated datasets of Brodatz album for the original HOG and Eig(Hess)-HOG algorithms. The bolded values represent the best classification results. Method

Feature size

Dataset1 (%)

Dataset 2 (%)

Dataset 3 (%)

Original HOG Eig(Hess)-HOG

128 128

5.05 95.81

4.10 94.31

1.30 99.37

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

11

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

at angles of 0–1600 with 100 increments. The proposed method achieves a remarkable classification rate for all angle values. Figs. 6 and 7 show the classification accuracy of subimages with different rotated angles in dataset 1, 2 and 3. The result of the original HOG algorithm is too low along with the rotation angles, while the Eig(Hess)-HOG algorithm with continuous rotation invariance achieves more stable and accurate classification results.

Table 6 Classification accuracy on the rotated datasets of Brodatz album for the original CoHOG, Eig(Hess)-CoHOG and GM-CoHOG algorithms. The bolded values represent the best classification results.

The second experiment about rotation analysis is done among original CoHOG, Eig(Hess)-CoHOG and GM-CoHOG algorithms. Table 6 shows the classification results for the original CoHOG, Eig(Hess)-CoHOG and GM-CoHOG algorithms. For the datasets 1, 2, and 3, the classification results of the original CoHOG algorithm vary from 9% to 68%. These results are very low. The proposed Eig(Hess)-CoHOG and GM-CoHOG algorithms achieve incredible success. In both methods, a high and stable classification results are obtained on the all the datasets. Fig. 7 shows the classification accuracy of subimages with different rotated angles in dataset 1, 2 and 3 for all CoHOG algorithms. The result of the original CoHOG algorithm is unsuccessful along with the rotation angles, while the Eig(Hess)-HOG and GM-CoHOG algorithms with continuous rotation invariance obtain more robust and high classification results. Suggested reasons underlying the success of both methods can be expressed substantially as follows:

Classification accuracy (%)

100 80 60

Original HOG Eig(Hess)-HOG

40 20 0 20

40

60

80

100

140

120

160

Feature size

Dataset 1 (%)

Dataset 2 (%)

Dataset 3 (%)

Original CoHOG Eig(Hess)CoHOG GM-CoHOG

1536 1176

31.51 83.29

15.99 79.75

9.69 94.36

1536

78.28

78.48

98.63

containing higher order differential analysis, reaches high classification rate by pixel labeling process independent of the rotation angle. In addition to all these, Eig(Hess)-CoHOG algorithm that we proposed has the highest classification rate in spite of having smaller feature vector size than two the other methods. (2) There are two fundamental novelties in GM-CoHOG algorithm that we proposed. Firstly, in the pixel labeling process, the second-order horizontal, vertical and diagonal derivatives of image are used instead of the first-order derivative. Instead of using first-order derivative, Gaussian and mean curvatures are used. In the literature, there are almost no studies depending on curvature information in gradient orientations and texture classification. Finally, novel labeling formulation is used as the best strategy to label texture pixels. Another important novelty developed in this algorithm is the labeling formula. This formulation used in proposed algorithms is also used in the analysis of function curves in differential geometry. Therefore, precise measurements and calculations are performed on images.

6. More experiments and discussions 6.1. Comparison with texton dictionary-based descriptors In this section, the novel developed HOG and CoHOG algorithms are compared with the texton dictionary-based methods in terms Classification accuracy (%)

Classification accuracy (%)

(1) The gradient calculation of the original CoHOG algorithm heavily relies on simple pixel difference. However, the eigenvalues of the Hessian matrix are used in the Eig(Hess)-CoHOG algorithm that we proposed. The eigenvalues of an image surface represent the principal curvature details in the image and are rotation invariant. These eigenvalues are also used in the analysis of the non-linear function curvatures in differential geometry. Orientation information of pixels calculated by k1 and k2 eigenvalues of the matrix obtained in the Eig(Hess)-CoHOG algorithm directly affects accuracy and quality of the pixel labeling process in the next step of the algorithm. Eig(Hess)-CoHOG algorithm,

Method

100 80 60

Original HOG Eig(Hess)-HOG

40 20 0 20

40

60

80

100

120

Rotated angle (degree)

Rotated angle (degree)

(a)

(b)

140

160

100 80 60

Original HOG Eig(Hess)-HOG

40 20 0

0

30

60

90

120

150

180 200

Rotated angle (degree)

(c) Fig. 6. Classification accuracy of different rotated angles. (a) Is for dataset 1, the variances of the classification accuracies are VARHOG = 0.0116, VAREig(Hess)-HOG = 0.0011. (b) Is for dataset 2, the variances of the classification accuracies are VARHOG = 0.0061, VAREig(Hess)-HOG = 0.0002. (c) Is for dataset 3, the variances of the classification accuracies are VARHOG = 0.0050, VAREig(Hess)-HOG = 0.00007.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

Classification accuracy (%)

100 80 60 40 CoHOG

20

Eig(Hess)-CoHOG GM-CoHOG

0 20

40

60

100

80

120

140

160

Classification accuracy (%)

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

Classification accuracy (%)

12

100 80 60

CoHOG

40

Eig(Hess)-CoHOG GM-CoHOG

20 0 20

40

60

80

100

120

Rotated angle (degree)

Rotated angle (degree)

(a)

(b)

140

160

100 80 CoHOG

60

Eig(Hess)-CoHOG GM-CoHOG

40 20 0

0

30

60

90

120

150

180 200

Rotated angle (degree)

(c) Fig. 7. Classification accuracy of different rotated angles. (a) Is for dataset 1, the variances of the classification accuracies are VARCoHOG = 0.0265, VAREig(Hess)-CoHOG = 0.0028, and VARGM-CoHOG = 0.0019. (b) Is for dataset 2, the variances of the classification accuracies are VARCoHOG = 0.0399, VAREig(Hess)-CoHOG = 0.0036, and VARGM-CoHOG = 0.0041. (c) Is for dataset 3, the variances of the classification accuracies are VARCoHOG = 0.0926, VAREig(Hess)-CoHOG = 0.0070, and VARGM-CoHOG = 0.0002.

of their classification accuracy and feature vector size. The texton dictionary methods are learned from the training images by clustering the local descriptors, and the representation of each image is the frequency histogram of the textons. Guo et al. [13] propose two types of local descriptors based on Gaussian derivatives filters, both of them have the property of continuous rotation invariance. The first local descriptor directly uses the maximum of the filter responses named continuous maximum responses (CMR). The second local descriptor rectifies the Gaussian filter responses to calculate principal curvatures (PC) of the image surface. The other study to which we compare the algorithms we developed is Joint_Sort texton dictionary-based techniques developed by Guo et al. [13]. The idea of the Joint_Sort method could be extended to other local patch based methods. In order to compare them with the proposed four new HOG and CoHOG algorithms, we conducted the same texture classification experiments of paper [41]. Classification results of CMR, PC and Joint_Sort methods were taken from their own paper [13,41]. The classification results for the texton number of CMR and PC methods are considered as 40. In addition, the texton number of Joint_Sort method is considered as 100, because the highest classification results in the related articles are achieved with these numbers of textons. The feature vector sizes used by the texton dictionary-based methods are calculated with the text on number  class number formulation used in articles. Thus, during the test process of PC and CMR methods, feature vector sizes are found to be 1000 for UIUC dataset, and 4480 for Brodatz album. In the same way, as in Joint_Sort method texton number is 100, they are calculated as 2500 for the UIUC dataset, and 11,200 for Brodatz album. Tables 7 and 8 show the classification results for UIUC and Brodatz datasets. The first experiment is carried out on the Brodatz album of 112 textures, each of which is divided into nine equally sized regions, giving 999 texture samples. The proposed Eig(Hess)-CoHOG algorithm obtained the highest classification rate with a ratio of 98.21% on Brodatz dataset. GM-CoHOG algorithm has a success rate of 97.25%, leaving behind CMR, PC and Joint_Sort techniques. The second experiment is performed on the UIUC dataset. Our proposed Eig(Hess)-CoHOG and GM-CoHOG algorithms

achieve better classification results on the UIUC dataset, and CMR, PC and Joint_Sort methods achieve better performance on the UIUC dataset. Here, there are two important differences between the methods we proposed and other texton dictionary-based methods. First, in texton dictionary-based methods, the number of textons used during the training process significantly affects classification performance. For example, in experiments on the UIUC dataset, CMR method has a success rate of 91.51% for 10 texton. Also, it has a classification success rate of 93.03% for 40 texton. This situation is also true for the PC and Joint_Sort methods. The second important difference is K-means algorithm used in the training stage of the texton dictionary-based methods. During the training stage of the texton dictionary methods, feature vectors obtained in original image size are divided into subsets equal to the number of textons determined using K-means algorithm. This process increases the computational time of the algorithm in a very significant rate. To analyze this situation in a more concrete form. Let us briefly examine only the training phase of PC method regarding texton number to be 40. This method performs the first and second derivative calculations for each of r = 1, 2, 4, 8 values with the purpose of extraction of the feature vector of a single image with 200  200 sizes within CUReT dataset. It calculates principal curvature information in the same size with two original images by using the derivative information obtained for each of r Table 7 Classification accuracy based on the Brodatz dataset. The results of CMR and PC [41], and Guo et al. are from the Ref. [13]. The bolded values represent the best classification results. Methods

Feature size

Brodatz (%)

GDF-HOG Eig(Hess)-HOG Eig(Hess)-CoHOG GM-CoHOG CMR [41] PC [41] Joint_Sort [13]

128 128 1176 1535 4480 4480 11,200

88.75 85.91 98.21 97.25 93.71 92.83 86.71

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx Table 8 Classification accuracy based on the UIUC dataset. The results of CMR and PC [41], and Guo et al. are from the Ref. [13]. The bolded values represent the best classification results. Methods

Feature size

UIUC (%)

GDF-HOG Eig(Hess)-HOG Eig(Hess)-CoHOG GM-CoHOG CMR [41] PC [41] Joint_Sort [13]

128 128 1536 1176 1000 1000 2500

88.63 85.71 96.82 98.41 93.03 90.69 92.73

Table 9 Classification accuracies (%) of LBP, RTRID, V2-RTRID and the proposed CoHOG algorithms on Brodatz album (rotated and noisy texture dataset from Brodatz album). The bolded values represent the best classification results. Methods

SNR

Average accuracy (%)

LBP RTRID V2-RTRID Eig(Hess)-CoHOG GM-CoHOG

4 4 4 4 4

86.50 89.70 91.4 93.54 96.08

values. For four different r values, a total of eight basic curvature data is calculated. Eventually, for each training set image, a feature vector of 8  40,000 size is calculated. When the number of training images for each class is taken 20 for CUReT dataset, training information in size of 160  40,000 shall be given to K-means algorithm. When the histogram of this information is prompted to separate into a set of 40 textons with K-means algorithm, significant time requirements will occur. Training image histograms obtained in the training phase by such processes and labeled as 40 textons are used in the testing phase. Texton dictionary-based methods are time-consuming and awkward with the K-means algorithm and they also require sufficient training texture images to construct texton dictionary. When considering the computational cost of the each texture class in the datasets, it will be understood how slow the texton dictionary-based methods work. Note that the above-mentioned texton-dictionary based methods do not provide us with information about the computation time. Therefore, any computational comparison could not be made with our algorithms. However, K-means and any other similar algorithms are not used in the four novel texture classification algorithms we proposed. The obtained feature vectors are directly given to the NN classifier. Especially owing to the principal curvature information of each pixel computed by eigenvalues in Eig(Hess)-CoHOG, time performance of the algorithm is improved algorithm through considerably minimizing feature vector size. Although proposed algorithms have lower dimensional feature vector, they generally have higher classification performance. 6.2. Comparison with several descriptors In this experiment we make a brief comparison between novel CoHOG algorithms and the several rotation invariant descriptors including rapid-transform based rotation invariant descriptor (RTRID) [11], V2-RTRID [11] and LBP [18]. The RTRID and V2-RTRID are based on the local circular neighborhood and the local feature vector is obtained by means of Rapid-transform. This experiment is conducted on the Brodatz album. This set consists of 16 Brodatz texture classes, and each texture class includes seventeen rotation images of size 128  128 rotated by different angles (0°, 10°, 20°, . . . , 160°). The single pattern (16, 2) is used for the three neighborhood based methods LBP, RTRID, V2-RTRID. To evaluate the

13

robustness of methods under the condition of additive Gaussian noise, we tested the performance of classifying texture images which were added Gaussian noise with zero mean and a variance dependent on specific texture to obtain required signal-to-noise ratio (SNR). Table 9 lists the comparative results. It can be seen that the proposed CoHOG algorithm achieves excellent classification performance under rotation conditions. From Table 9, it can be seen that single descriptors RTRID and V2-RTRID obtain better classification accuracy than LBP method. Their good performance should be contributed to feature selection processing. The RTRID method obtains poor performance due to its global rotation invariance. The textures of the same texture class have large intra-class dissimilarity, thus the directions selected by Radon transform are no longer stable. In many image processing applications we need to deal with noisy texture images. As a result, robustness to noise is considered as one of the most significant factors to assess texture classification methods. It is obvious that LBP method has a good result under ideal condition; however it is not as robust as the other methods under noise and rotation conditions. Our CoHOG algorithms also have the best correct classification rates of 93.54% and 95.08% among them. Due to usage of Hessian and principal curvatures, novel CoHOG descriptors have a good average classification rate. The experimental results show that higher order directional derivatives can obviously improve classification accuracy. So it is important to take into account the higher order directional derivatives (i.e. principal curvatures) for the improvement of accuracy. 7. Conclusion Designing an effective and robust feature extraction algorithm for texture classification under non-ideal conditions is a challenging task. In this paper we have proposed rotation invariant feature extraction algorithms which are robust to illumination and pose variations. The developed four novel feature extraction algorithms are based on the strengthening of the calculation of the gradient orientations in the original HOG and CoHOG algorithms. The classical HOG and CoHOG algorithms directly use the gradient information, which will discard the useful information in texture images. Therefore, we have presented four novel extensions of HOG and CoHOG, the key idea is to make efficient and reasonable use of derivative information in the local texture patterns, especially in the nonuniform patterns. In four novel algorithms we developed, Gaussian derivative filters, eigenvalues of the Hessian matrix and Gaussian–mean curvatures are used. Thanks to robust and powerful mathematical methods, rotation invariant and qualified texture features are obtained. Very high classification performance is obtained on well-known and widely used texture datasets without increasing feature vector dimension of the algorithms. Moreover, in the Eig(Hess)-CoHOG algorithm that we developed, higher classification performance is obtained by reducing the size of feature vector of the original CoHOG algorithm to the ratio of 23.43%. Rotation invariant classification experiments are carried out on the Brodatz dataset and promising results are obtained from experiments. Finally, the superiority of the proposed algorithms is analyzed by a comprehensive comparison with the state-of-the-art the texture classification methods. Appendix A The image behavior in a local neighborhood of a point can be expressed by the Taylor expansion. Therefore, the local neighborhood can be simplified as an expansion of functions. The Taylor expansion of a local neighborhood of a point ~ x0 is defined in Eq. (A1),

1* hÞ ¼ f ð~ h þ h T r2 f ð~ h þ rð~ hÞ x0 þ ~ x0 Þ þ rf ð~ x0 Þ~ x0 Þ~ f ð~ 2!

ðA1Þ

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

14

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx

hÞ is a residual. where rð~ Differential geometry, in general, allows extracting properties and features of points by means of first and second-order terms of the Taylor expansion. While first order terms in Taylor expansion contain information about gradient distribution in a local neighborhood, second-order terms include information about the local shape. G.1. First fundamental form (gradient distribution) Given a point p, the first-order terms of the Taylor expansion correspond to its so-called Jacobian matrix is defined in Eq. (A2),



@p @x @p @y

! ¼

1 0

fx

! ðA2Þ

0 1 fy

where first partial derivatives fx and fy are estimated applying Gaussian derivative kernels on the image. The first fundamental form I for the point p on an image f is defined in Eq. (A3): 2

T

I ¼JJ ¼

!

1 þ fx

f xf y

f yf x

1 þ fy

ðA3Þ

2

G.2. Second fundamental form (Hessian matrix) The normal vector N to the point p on an image f is defined in Eq. (A4),

1 f x 1 C

¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi B N ¼

@ f y A @p

2 2 

@p

þ f 1 þ f @x @y x y 1 @p @x

0

 @p @y

ðA4Þ

Second-order terms of the Taylor expansion conform the Hessian matrix H, which is similar to the second fundamental form. Eq. (A2) is derived from the normal curvature estimation and it is calculated using the second-order partial derivatives and the normal vector N for each point p. So the second fundamental form for images is described in Eq. (A5),

0 H¼@

@2 p N @x2

@2 p N @x@y

@2 p N @x@y

@2 p N @y2

1 A¼

ð0 0

f xx ÞN ð0 0

f xy ÞN

!

ð0 0 f xy ÞN ð0 0 f yy ÞN ! f xx f xy 1 1 ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffi H 2 2 f f detðIÞ yx yy 1 þ fx þ fy

ðA5Þ

where second partial derivatives fxx, fxy, fyy are estimated applying Gaussian derivative kernels on the image and det (I) is the determinant of the first fundamental form. References [1] L. Liu, P. Fieguth, D. Clausi, G. Kuang, Sorted random projections for robust rotation-invariant texture classification, Pattern Recogn. 45 (2012) 2405–2418. [2] Q. Ji, J. Engel, E. Craine, Texture analysis for classification of cervix lesions, IEEE Trans. Med. Imag. 19 (2000) 1144–1149. [3] H. Anys, D.-C. He, Evaluation of textural and multipolarization radar features for crop classification, IEEE Trans. Geosci. Remote Sens. 33 (1995) 1170–1181. [4] K. Hanbay, M.F. Talu, Segmentation of SAR images using improved artificial bee colony algorithm and neutrosophic set, Appl. Soft Comput. 21 (2014) 433–443. [5] D. Tsai, T. Huang, Automated surface inspection for statistical textures, Image Vis. Comput. 21 (2003) 307–323. [6] W. Jia, D.S. Huang, D. Zhang, Palmprint verification based on robust line orientation code, Pattern Recogn. 41 (2008) 1504–1513. [7] X.-F. Wang, D.-S. Huang, J.-X. Du, H. Xu, L. Heutte, Classification of plant leaf images with complicated background, Appl. Math. Comput. 205 (2008) 916– 926. [8] Y.-Y. Wan, J.-X. Du, D.-S. Huang, Z. Chi, Y.-M. Cheung, X.-F. Wang, G.-J. Zhang, Bark texture feature extraction based on statistical texture analysis, in: ISIMP: Proceedings of the International Symposium on Intelligent Multimedia, Video and Speech Processing, Hong Kong, China, pp. 482–485.

[9] X. Wang, K.K. Paliwal, Feature extraction and dimensionality reduction algorithms and their applications in vowel recognition, Pattern Recogn. 36 (2003) 2429–2439. [10] R. Maani, S. Kalra, Y.-H. Yang, Noise robust rotation invariant features for texture classification, Pattern Recogn. 46 (2013) 2103–2116. [11] C. Li, J. Li, D. Gao, B. Fu, Rapid-transform based rotation invariant descriptor for texture classification under non-ideal conditions, Pattern Recogn. 47 (2014) 313–325. [12] L. Liu, L. Zhao, Y. Long, G. Kuang, P. Fieguth, Extended local binary patterns for texture classification, Image Vis. Comput. 30 (2012) 86–99. [13] Z. Guo, Q. Li, L. Zhang, J. You, D. Zhang, W. Liu, Is local dominant orientation necessary for the classification of rotation invariant texture?, Neurocomputing 116 (2013) 182–191 [14] X. Sun, J. Wang, M.F.H. She, L. Kong, Scale invariant texture classification via sparse representation, Neurocomputing 122 (2013) 338–348. [15] S. Lazebnik, C. Schmid, J. Ponce, A sparse texture representation using local affine regions, IEEE Trans. Pattern Anal. Mach. Intell. 27 (8) (2005) 1265–1278. [16] J. Zhang, M. Marszalek, S. Lazebnik, C. Schmid, Local features and kernels for classification of texture and object categories: a comprehensive study, Int. J. Comput. Vis. 73 (2) (2007) 213–238. [17] E. Hayman, B. Caputo, M. Fritz, J. Eklundh, On the significance of real-world conditions for material classification, in: Computer Vision-ECCV, Springer, 2004, pp. 253–266. [18] T. Ojala, M. Pietikainen, T. Maenpaa, Multiresolution gray-scale and rotation invariant texture classification with local binary patterns, IEEE Trans. Pattern Anal. Mach. Intell. 24 (2002) 971–987. [19] M. Varma, A. Zisserman, A statistical approach to texture classification from single images, Int. J. Comput. Vis. 62 (2005) 61–81. [20] M. Varma, R. Garg, Locally invariant fractal features for statistical texture classification, in: ICCV 2007, IEEE International Conference on Computer Vision, 2007, pp. 1–8. [21] M. Varma, A. Zisserman, A statistical approach to material classification using image patches exemplars, IEEE Trans. Pattern Anal. Mach. Intell. 31 (2009) 2032–2047. [22] Y. Xu, H. Ji, C. Fermuller, Viewpoint invariant texture description using fractal analysis, Int. J. Comput. Vis. 83 (1) (2009) 85–100. [23] Y. Xu, X. Yang, H. Ling, H. Ji, A new texture descriptor using multifractal analysis in multi-orientation wavelet pyramid, in: CVPR 2010, IEEE Conference on Computer Vision and Pattern Recognition, 2010, pp. 161–168. [24] J.D.B. Nelson, Fused Lasso and rotation invariant autoregressive models for texture classification, Pattern Recogn. Lett. 34 (2013) 2166–2172. [25] J. Zhang, T. Tan, Brief review of invariant texture analysis methods, Pattern Recogn. 35 (2002) 735–747. [26] L. Davis, S. Johns, J. Aggarwal, Texture analysis using generalized cooccurrence matrices, IEEE Trans. Pattern Anal. Mach. Intell. 1 (3) (1979) 251–259. [27] L.S. Davis, Polarograms: a new tool for image texture analysis, Pattern Recogn. 13 (1981) 219–223. [28] D. Chetverikov, Experiments in the rotation-invariant texture discrimination using anisotropy features, in: Proceedings of the PRIP, 1982, pp. 1071–1074. [29] F. Cohen, Z. Fan, M. Patel, Classification of rotated and scaled textured images using gaussian markov random field models, IEEE Trans. Pattern Anal. Mach. Intell. 13 (1991) 192–202. [30] J. Chen, A. Kundu, Rotation and gray scale transform invariant texture identification using wavelet decomposition and hidden markov model, IEEE Trans. Pattern Anal. Mach. Intell. 16 (1994) 208–214. [31] C. Pun, M. Lee, Log-polar wavelet energy signatures for rotation and scale invariant texture classification, IEEE Trans. Pattern Anal. Mach. Intell. 25 (2003) 590–603. [32] R. Porter, N. Canagarajah, Robust rotation-invariant texture classification: wavelet, Gabor filter and GMRF based schemes, in: IEE Proceedings of the Vision, Image and Signal Processing, 1997, vol. 144, pp. 180–188. [33] R. Manthalkar, P. Biswas, B. Chatterji, Rotation and scale invariant texture features using discrete wavelet packet transform, Pattern Recogn. Lett. 24 (2003) 2455–2462. [34] K. Jafari-Khouzani, H. Soltanian-Zadeh, Rotation-invariant multiresolution texture analysis using radon and wavelet transforms, IEEE Trans. Image Process. 14 (2005) 783–795. [35] K. Jafari-Khouzani, H. Soltanian-Zadeh, Radon transform orientation estimation for rotation invariant texture analysis, IEEE Trans. Pattern Anal. Mach. Intell. 27 (2005) 1004–1008. [36] S. Arivazhagan, L. Ganesan, S. Priyal, Texture classification using Gabor wavelets based rotation invariant features, Pattern Recogn. Lett. 27 (2006) 1976–1982. [37] P. Cui, J. Li, Q. Pan, H. Zhang, Rotation and scaling invariant texture classification based on radon transform and multiscale analysis, Pattern Recogn. Lett. 27 (2006) 408–413. [38] E. Ves, D. Acevedo, A. Ruedin, X. Benavent, A statistical model for magnitudes and angles of wavelet frame coefficients and its application to texture retrieval, Pattern Recogn. 47 (2014) 2925–2939. [39] Y. Ge, D. Yang, J. Lu, B. Li, X. Zhang, Active appearance models using statistical characteristics of Gabor based texture representation, J. Vis. Commun. Image Represent 24 (2013) 627–634. [40] M. Mellor, B. Hong, M. Brady, Locally rotation, contrast, and scale invariant descriptors for texture analysis, IEEE Trans. Pattern Anal. Mach. Intell. 30 (2008) 52–61.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004

K. Hanbay et al. / Computer Vision and Image Understanding xxx (2014) xxx–xxx [41] J. Zhang, H. Zhao, J. Liang, Continuous rotation invariant local descriptors for texton dictionary-based texture classification, Comput. Vis. Image Und. 117 (2013) 56–75. [42] C. Lopez-Molina, B.D. Baets, H. Bustince, Generating fuzzy edge images from gradient magnitudes, Comput. Vis. Image Und. 115 (2011) 1571–1580. [43] N. Dalal, B. Triggs. Histograms of oriented gradients for human detection, in: CVPR 2005. IEEE Conference on Computer Vision and Pattern Recognition, 2009, pp. 886–993. [44] M.F. Talu, M. Gül, N. Alpaslan, B. Yigitcan, Calculation of melatonin and resveratrol effects on steatosis hepatis using soft computing methods, Comput. Methods Prog. Biomed. 111 (2013) 498–506. [45] T. Watanabe, S. Ito, K. Yokoi, Co-occurrence histograms of oriented gradients for pedestrian detection, Lect. Notes Comput. Sci. 5414 (2009) 37–47. [46] S. Maji, A. Berg, J. Malik, Classification using intersection kernel support vector machines is efficient, in: CVPR 2008, IEEE Conference on Computer Vision and Pattern Recognition, 2008, pp. 1–8. [47] P. Sabzmeydani, G. Mori, Detecting pedestrians by learning shapelet features, in: CVPR 2007, IEEE Conference on Computer Vision and Pattern Recognition, 2007, pp. 1–8. [48] S. Phung, A. Bouzerdoum, A new image feature for fast detection of people in images, Int. J. Inf. Syst. Sci. 3 (3) (2007) 383–391. [49] T. Do, E. Kijak, Face recognition using co-occurrence histograms of oriented gradients, in: IEEE Proceedings of the International Conference on Acoustics, Speech and Signal Processing, 2012, pp. 1301–1304. [50] T. Kozakaya, S. Ito, S. Kubota, O. Yamaguchi, Cat face detection with two heterogeneous features, in: ICIP 2009, IEEE International Conference on Image Processing, 2009, pp. 1213–1216. [51] J. Xu, Q. Wu, J. Zhang, Z. Tang, Object detection based on co-occurrence GMuLBP features, in: Proceedings of ICME 2012, IEEE International Conference for Multimedia and Expo, 2012, pp. 943–948. [52] X. Tizon, Algorithms for the Analysis of 3D Magnetic Resonance Angiography Images, Ph.D. Thesis, Uppsala, Sweden, 2004, ISBN 91-576-6700-4. [53] P. Orlowski, M. Orkisz, Efficient computation of Hessian-based enhancement filters for tubular structures in 3D images, IRBM 30 (2009) 128–132. [54] E. Moghimirad, S.H. Rezatofighi, H.S. Zadeh, Retinal vessel segmentation using a multi-scale medialness function, Comput. Biol. Med. 42 (2012) 50–60. [55] A. Sefidpour, N. Bouguila, Spatial color image segmentation based on finite non-Gaussian mixture models, Exp. Syst. Appl. 39 (2012) 8993–9001. [56] R. Lakemond, C. Fookes, S. Sridharan, Affine adaptation of local image features using the Hessian matrix, in: IEEE Proceedings of the International Conference

[57] [58] [59]

[60]

[61]

[62] [63] [64] [65]

[66]

[67] [68] [69] [70] [71]

[72]

15

on Advanced Video and Signal Based Surveillance, Genoa, Italy, 2009, pp. 496– 501. Y. Pang, Y. Yuan, X. Li, J. Pan, Efficient HOG human detection, Signal Process. 91 (4) (2011) 773–781. A. Albiol, D. Monzo, A. Martin, J. Sastre, A. Albiol, Face recognition using HOGEBGM, Pattern Recogn. Lett. 29 (10) (2008) 1537–1543. M. Kaâniche, F. Bre´mond, Recognizing gestures by learning local motion signatures of HOG descriptors, IEEE Trans. Pattern Anal. Mach. Intell. 34 (11) (2012) 2247–2258. P.-Y. Chen, C.-C. Huang, C.-Y. Lien, Y.-H. Tsai, An efficient hardware implementation of HOG feature extraction for human detection, IEEE Trans. Intell. Transp. Syst. 15 (2) (2014) 656–662. Y. Pang, H. Yan, Y. Yuan, K. Wang, Robust CoHOG feature extraction in humancentered image/video management system, IEEE Trans. Syst., Man Cybernet., Part B: Cybernet. 42 (2) (2012) 458–468. W.T. Freeman, E.H. Adelson, The design and use of steerable filters, IEEE Trans. Pattern Anal. Mach. Intell. 13 (9) (1991) 891–906. B.S. Morse, Lecture 11: Differential Geometry, Brigham Young University, 2000. M. Do Carmo, Differential Geometry of Curves and Surfaces, Prentice-Hall, Englewood Cliffs, NJ, 1976. R. Su, C. Sun, T.D. Pham, Junction detection for linear structures based on Hessian, correlation and shape information, Pattern Recogn. 45 (2012) 3695– 3706. M. Feuerstein, B. Glocker, T. Kitasaka, Y. Nakamura, S. Iwano, K. Mori, Mediastinal atlas creation from 3-D chest computed tomography images: application to automated detection and station mapping of lymph nodes, Med. Image Anal. 16 (2012) 63–74. R. Duda, P. Hart, D. Stork, Pattern Classification, Citeseer, 2001. B. Caputo, E. Hayman, M. Fritz, J. Eklundh, Classifying materials in the real world, Image Vis. Comput. 28 (2010) 150–163. K. Dana, B. Van Ginneken, S. Nayar, J. Koenderink, Reflectance and texture of real-world surfaces, ACM Trans. Graph. (TOG) 18 (1999) 1–34. P. Mallikarjuna, M. Fritz, A. Targhi, E. Hayman, B. Caputo, J. Eklundh, The kthtips and kth-tips2 databases, 2006. B. Caputo, E. Hayman, P. Mallikarjuna, Class-specific material categorisation, in: ICCV 2005, Tenth IEEE International Conference on Computer Vision, 2005, vol. 2, IEEE, pp. 1597–1604. P. Brodatz, Textures: A Photographic Album for Artists and Designers, vol. 66, Dover, New York, 1966.

Please cite this article in press as: K. Hanbay et al., Continuous rotation invariant features for gradient-based texture classification, Comput. Vis. Image Understand. (2014), http://dx.doi.org/10.1016/j.cviu.2014.10.004