J. Vis. Commun. Image R. 64 (2019) 102615
Contents lists available at ScienceDirect
J. Vis. Commun. Image R. journal homepage: www.elsevier.com/locate/jvci
Local energy oriented pattern for image indexing and retrieval q G.M. Galshetwar a,⇑, L.M. Waghmare b, A.B. Gonde a, S. Murala c a
Centre of Excellence in Signal and Image Processing (COESIP), Department of ECE, SGGSIET, Nanded, Maharashtra, India Department of Instrumentation Engineering, SGGSIET, Nanded, Maharashtra, India c Computer Vision and Pattern Recognition Laboratory, Department of EE, IIT Ropar, India b
a r t i c l e
i n f o
Article history: Received 6 December 2017 Revised 16 May 2019 Accepted 22 August 2019 Available online 26 August 2019 Keywords: Local Binary Patterns (LBP) Local Mesh Patterns (LMeP) Local Directional Mask Maximum Edge Patterns (LDMaMEP)
a b s t r a c t A novel image indexing algorithm for Content Based Image Retrieval (CBIR) using Local Energy Oriented Patterns (LEOP) is proposed in this paper. LEOP encodes pixel level energy orientations to find minute spatial features of an image whereas existing methods use neighborhood relationship. LEOP maps four pixel progression orientations to find top two maximum energy changes for each reference pixel in the image i.e. for each reference 3 3 grid, two more 3 3 grids out of four pixel progression are extracted. Finally, LEOP encodes the relationship among pixels of three 3 3 local grids extracted. LEOP is applied on four different image databases named MESSIDOR, VIA/I-ELCAP, COREL and ImageNet Database using traditional CBIR framework. To test the robustness of proposed feature descriptor the experiment is extended to a learning based CBIR approach on COREL database. The LEOP outperformed state-of-the-art methods in both traditional as well as learning environments and hence it is a strong descriptor. Ó 2019 Elsevier Inc. All rights reserved.
1. Introduction In recent years there has been technological developments in the field of image acquiring devices as well as evolution of social media, that lead to creation of enormous amount of image data all over the internet in digital form let it be due to natural photography, wildlife photography etc. Also, there has been technological revolution in the field of medical science in near past, which has led to noticeable changes in diagnosis taking help of medical imaging techniques. These revolutions have made a great impact on our day-to-day life. Automatically huge image databases are being created in raw form. Images captured by medical imaging techniques are mostly used once and later they are stored as trash in the systems. 1.1. Motivation In last few years, there has been a rapid growth in the severe and critical diseases in India and all over the world. Because of this growth, there is increasing need of expert medical services in urban as well as in remote places especially in developing countries. In areas where general clinical practices are present, we q
This paper has been recommended for acceptance by ‘Zicheng Liu’.
⇑ Corresponding author.
E-mail address:
[email protected] (G.M. Galshetwar). https://doi.org/10.1016/j.jvcir.2019.102615 1047-3203/Ó 2019 Elsevier Inc. All rights reserved.
can provide them a technological solution which can assist those clinics in bringing expert medical services to their help. Bio-medical imaging has emerged as a very useful technological development in medical diagnostic field. It is used to create visual representation of different interior parts of body that are used for medical analysis and diagnosis. There are different types of Biomedical imaging techniques such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI) and Fundus Imaging etc. With these techniques in hand, engineers can provide further technological solutions to medical field. Content Based Image Retrieval (CBIR) is one of such technological solutions in which engineering and medical streams can work hand-in-hand. CBIR works in two stages first is the feature extraction stage in which features are extracted from the images in database and feature vector database is created. In the second stage, similarity matching is done in which distance of query image feature is measured from each of the images in feature database. The detailed literature survey about CBIR is given in [1–8]. 1.2. Related work Color is the significant feature of an image, if it is maintained semantically intact and in a perceptually oriented way. CBIR system considering color as most significant feature is proposed in [9]. The concept of color histogram is introduced in [10] to find the relationship between the color distributions of pixels and
2
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
spatial correlation between pairs of color. In [11] authors have introduced the concept of the color correlogram. Another algorithm for image retrieval is proposed in [12] in which authors used first three central moments of each color space called mean, standard deviation and skewness. In the above proposed algorithms the retrieval accuracy of feature descriptors is less, especially on large database because of insufficiency in discrimination power. Along with color feature, the shape feature of an image is also widely used. After doing segmentation of the image into region, the shape feature of each region is used for feature extraction. In [13] authors used Fourier descriptor to collect shape description. However, this method is very sensitive to the noise and invariant to the geometric transformation. To overcome these limitations, the modified Fourier descriptor is proposed in [14], which overcomes the limitations of the existing Fourier descriptor. Along with color and shape, another important characteristic of an image is texture, so texture based feature descriptor also plays an important role in the field of CBIR. For texture investigation (texture arrangement and texture based image retrieval), Discrete Wavelet Transform (DWT) [15], Gabor transform [16], Rotated wavelet filters [17], Dual-Tree Rotated Complex Wavelet Filter (DT-CWT) [18] based feature are used. Rotation invariant complex wavelet filters are proposed in [19] for texture based image retrieval to address the problems present in DWT (limited directional information). Local Binary Pattern (LBP) is introduced in [20] for texture classification. Rotational invariant LBP is proposed in [21] texture classification. In [22] authors proposed the Completed LBP (CLBP) for texture classification. To address the problem present in LBP feature descriptor (LBP gives all directional first order derivatives), in [23] authors presented the local derivative pattern for pattern recognition. Local maximum edge binary patterns are presented in [24] for image retrieval and object tracking. In [25] authors proposed a new CBIR system in which the feature vector of CH (color histogram) and SOT (spatial orientation tree) is constructed and stored for similarity matching. Using binary wavelet transform and local binary pattern Directional Binary Wavelet Pattern (DBWP) is proposed in [26]. Local Tetra Pattern (LTrP) presented in [27] calculates the first-order derivatives in vertical and horizontal direction and based on these directions the relationship between the referenced pixel and its neighbors is encoded. Motif Co-Occurrence Matrix (MCM) proposed in [28] divides the whole image in to non-overlapping 2x2 grid patterns to obtain the motif-transformed image which later constructs MCM. In [29] authors combined the color, texture and MCM feature to construct feature vector, used for CBIR purpose. The Modified Color MCM is presented in [30] that obtains the inter-correlation among the RG-B color planes for CBIR; this is absent in color MCM. Another CBIR method Directional Local Ternary Pattern (LTP) is proposed in [31] that gives directional information with respect to reference pixels. Local Ternary Co-occurrence Pattern (LTCoP) proposed in [32] integrates the concept of Local ternary pattern and cooccurrence matrix for CT and MRI image retrieval, which encodes the co-occurrence between ternary images. Directional Local Motif XoR Patterns (DLMXoRP) [33] calculated motif using flexible 1 3 grids to construct motif image, later XoR operations are applied on motif images. In [51] authors have proposed sparse embedded hashing framework for image retrieval. They first generate sparse representations in a data-driven way and later, the projection matrix is learned by employing matrix factorization technique to approximate the Euclidean structures of the original data. In [52] authors have designed multi-scale trajectory-pooled 3D convolutional descriptors. In this approach they calculate multi-scale dense trajectories from the input video and perform trajectory pooling on feature maps of 3D Convolutional Neural Network (3D CNN). In [55] authors proposed a multi-graph learning based medical image retrieval technique in which, the query category
prediction is done by formulating query into a multi-graph structure with a set of selected subjects in the database to learn the relevance between the query subject and the existing subject categories based on which a set of subjects in the database are selected as candidate retrieval results. Later, the relationship between these candidates and the query is further learned with a new multi-graph which is used to rank the candidates. The affective image retrieval via multi-graph learning is proposed in [53] where, authors used combination of extracted generic features and features derived from elements-of-art are extracted as lowlevel features. Later, they constructed single graph for each kind of feature to test the retrieval performance. Local Gaussian Difference Extrema Pattern (LGDEP) is proposed in [54] for face recognition. In this method authors calculated prominent edges from three Gaussian scaled images. Volumetric Local Directional Triplet Patterns (VLDTP) for biomedical image retrieval is proposed in [55], where in authors considered directional information of 3D image to construct directional local triplet pattern. In this proposed method, we are calculating amount of pixel progression in four gradient directions to indicate maximum changes in those specific directions. Our proposed pattern encodes the relationship among pixels of three 3 3 local patches, out of which two 3 3 local patches are extracted by moving the reference 3 3 local patch in two gradient directions which give first two maximum pixel progressions out of four pixel progression directions. 1.3. Main contribution We have proposed a novel pattern to encode the gradient directional information among 3 3 local blocks of image extracted from image. First 3 3 reference block is extracted by considering each pixel as center pixel. Later by moving this local block in two gradient directions giving maximum pixel progression, we extract two more 3 3 local blocks. We encode the comparison of the pixel intensity of each pixel in first local block with each corresponding pixel in two extracted directional local blocks. Arrangement of paper is as follows: Section 2 gives introduction to some existing local patterns which inspired us for our approach. Section 3 gives detail description of proposed methodology. The experimental results analysis of proposed feature descriptor is depicted in Section 4. Section 5 gives concluding remarks. 2. Local patterns 2.1. Local Binary Pattern (LBP) A pattern very much useful for texture classification is presented in [20] popularly known as LBP, which carries out pixel intensity comparison of center pixel with the neighbor pixels. The comparison is encoded using (1).
LBPP;R ¼
P X
2ðn1Þ f 1 ðIn Ic Þ
ð1Þ
n¼1
where,
1 xP0 where In and Ic indicate pixel intensity of 0 else neighbor and center pixel respectively. P gives total number of surrounding neighbors and R is the neighborhood radius. f I ðxÞ ¼
2.2. Local Mesh Pattern (LMeP) Local mesh pattern popularly known as LMeP is proposed in [36]. For each pixel in an image the relationship among its
3
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
surrounding neighbors is encoded to compute LMeP and is carried out using (2).
LMePkN;R
¼
N X
2
ðn1Þ
f 1 ðIa j
R
Ii j R Þ
ð2Þ
n¼1
3.1. Gradient calculation The horizontal (Gx ) and vertical gradient (Gy ) calculation at each pixel is done using (3) and (4)
Gx ¼
@I @x
ð3Þ
Gy ¼
@I @y
ð4Þ
a ¼ 1 þ mod ððn þ N þ k 1Þ; NÞ N 8 k ¼ 1; 2; :::; 2 where LMeP index is given by k and mod ðA; BÞ returns the remainder for fA=Bg operation. P gives number of surrounding neighbors and R is the neighborhood radius.
Here I is the input image of size ðX; YÞ. Using horizontal and vertical directions from (3) and (4) gradient magnitude (Gxy ) and gradient direction (aðx; yÞ) are calculated at each pixel using (5) and (6).
ffi qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi Gxy ¼ G2 þ G2 x y
2.3. Local Directional Mask Maximum Edge Pattern (LDMaMEP) In [37] maximum directional edge information in an image is encoded to construct LDMaMEP. Here they first applied eight directional masks to obtain directional edges in the images, which are later used for computations of MEP (maximum edge patterns) and MEPP (maximum edge position patterns). The histograms of MEP and MEPPs are combined to form the final feature vectors and that are later used to do similarity matching between query image and database images.
3. Local energy oriented pattern The descriptors proposed in [20,36,37] inspired us to propose a novel pattern Local Energy Oriented Pattern (LEOP). Fig. 1 gives detail block diagram representation of the proposed system framework. In the first stage, LEOP calculates gradient magnitude and gradient direction at each pixel and quantize all gradient directions n o in four gradient directions þ900 ; þ450 ; 00 ; 450 to indicate maximum changes in those specific directions. Further, in the second stage, we select two directions out of four gradient directions, showing first maximum and second maximum changes at each pixel. In the third stage, for a given reference pixel, we extract reference local 3 3 patch and two directional 3 3 local patches considering first and second maximum directions selected previously in second stage. In fourth stage, LEOP encodes the relationship among pixels of three 3 3 local patches extracted in third stage.
aðx; yÞ ¼ tan1
ð5Þ
Gy Gx
ð6Þ
Here ðx; yÞ gives location of current pixel in an input image of size ðX; YÞ. 3.2. 1.1 Direction quantization All the gradient directions calculated from (6) are quantized and stored into four directional maps by computing maximum directional change at each pixel using (7). For computing maximum directional change, we are considering map of size ð2 N 1 þ 1Þ ð2 N 2 þ 1Þ.
IaL ðx; yÞ ¼
N1 X
N2 X Gxy ðx; yÞ f ðaL ; aU ; aðx; yÞÞ 2
where
f 2 ða; b; cÞ ¼
1 if a 6 c < b 0
else
8 aL ¼ þ90; þ45; 0; 45;
aU ¼ aL 45;
In (7) IaL is the directional map showing amount of pixel progression in gradient direction aL . For experimental purpose, we
Maximized (Quantized ( ) Magnitude) MAP
3-D Grid Extracon
Image Database Gradient Magnitude
Direcon 1 Direcon 2 Direcon 3
Gradient Direction
ð7Þ
x¼N 1 y¼N2
1st Maxima Grid 1st Maxima Reference Grid 2nd Maxima
Extrema Encoding
Paern Calculaon
2nd Maxima Grid
Direcon 4 Histogram
Query Image Retrieval Result Fig. 1. System block diagram of proposed framework.
Similarity Measurement
Feature Vector
4
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
Input Image (I)
59
45
31
44
43
37
900
-450
-450
900
-450
-450
00
57
43
22
5
15
20
31
00
450
450
450
-450
00
00
22
55
23
10
1
15
11
32
-450
900
-450
-450
900
450
450
16
34
3
3
2
4
13
27
00
450
-450
900
450
450
450
Gy
31
2
1
2
3
14
28
00
00
-450
-450
00
450
450
Gx
30
6
2
2
4
15
26
00
00
900
-450
450
450
450
36
30
28
29
29
30
29
00
-450
900
900
900
900
450
54
39
32
47
28
17
76
51
39
30
44
31
24
43
32
33
33
32
31
31
30
29
31
29
27
28
30
30
31
31
28
17
22
29
28
29
29
26
14
29
28
29
28
25
Gx2
Gxy
( x, y )
(33 31)
Vertical Difference Around Center Pixel
90
Gxy Map
1
tan
900
I (29 29) x Horizontal Difference Around Center Pixel
I y
G y2
13
Gx
Gy
Map
{ 450 , 900 ,00 , 450 }
93
68
20
Quantized
Magnitude Gxy Map
450
450
00
1
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
0
1
1
0
1
1
0
0
0
0
0
0
0
0
0
1
1
1
0
0
0
1
0
0
0
0
1
1
0
0
0
0
1
0
0
0
1
0
0
1
0
0
0
0
0
0
0
1
1
0
0
0
0
0
0
0
1
0
1
1
0
0
0
0
0
0
1
0
0
0
0
1
0
0
1
1
1
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
1
1
1
1
0
0
1
0
0
0
0
1
1
0
0
0
0
0
1
0
0
0
0
0
0
0
0
1
1
1
1
1
0
0
0
0
0
0
0
0
1
0
0
0
0
0
1
1
1
1
0
0
0
0
0
0
0
1
1
0
0
0
0
0
0
0
1
0
0
0
0
0
0
45
0
0
0
Maps
450
93
0
0
31
0
0
0
0
0
0
0
0
0
0
0
0
44
43
0
0
0
0
0
0
0
0
43
22
5
0
0
0
57
0
0
0
0
20
31
0
0
0
0
15
0
0
0
23
0
0
15
0
0
0
0
0
0
0
11
32
0
0
0
0
0
0
0
55
0
10
1
0
0
0
0
0
0
2
0
0
0
0
3
0
0
4
13
27
34
0
0
0
0
0
0
0
0
3
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
14
28
31
2
0
0
3
0
0
0
0
1
2
0
0
0
0
0
2
0
0
0
0
0
0
0
0
4
15
26
30
6
0
0
0
0
0
0
0
0
2
0
0
0
0
0
28
29
29
30
0
0
0
0
0
0
0
29
36
0
0
0
0
0
0
0
30
0
0
0
0
0
Maximized (
90 1 0
45
0
0
0
0
0
0
37
0
59
45
0
Gxy ) MAP
450 4
00 3
2
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
282
x
x
x
x
x
x
276
x
x
x
x
x
x
287
x
x
x
x
x
x
310
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
Maxima( )
arg max( 900 , 450 ,00 , 450 ) i.e.max(1,2,3, 4)
90
1
1st Maxima ( 1 )
0
2
450
3
4
00
1
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
4
x
x
x
x
x
x
3
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
x
Directional Block ( I1 )
450
2nd Maxima ( 2 )
Reference Block ( I 2 )
2
Directional Block ( I 3 )
31
28
17
33
33
32
32
31
22
29
26
14
29
31
29
29
27
16
28
25
13
30
31
31
31
28
17
Gradient Directions (γ)
Binary Pattern Binary Pattern
1 0
if ( I 2 ( g i )
I1 ( g i )& I 3 ( g i )
or I 2 ( g i )
I1 ( g i )& I 3 ( g i ))
Elsewhere
Binary Weights
1
1
1
1
2
4
0
1
1
128
256
8
0
1
1
64
32
16
{1+2+4+8+16+32+256=319} 319
Fig. 2. Pattern calculation of proposed method.
selected N 1 ¼ N 2 ¼ 3. We particularly selected 7 7 neighboring size local patch for our operations. As the distance between reference patch center pixel and directional patch center pixel is 2 as shown in Fig. 2. So, we need to consider local patches with center pixel at distance 2 in all the four directions; now the neighboring size becomes 7 7.
3.3. Maxima computation We now have four directional maps giving maximum pixel progression at each pixel in the image in four directions. Using (8) we record first and second maximum gradient direction for every pixel location ðx; yÞ.
5
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615 k
ck ðx; yÞ ¼ arg max ðIaL ¼900 ðx; yÞ; IaL ¼450 ðx; yÞ; IaL ¼00 ðx; yÞ; IaL ¼450 ðx; yÞÞ a
3.6. Similarity matching
L
ð8Þ where k ¼ 1; 2 th
In (8) maxk ðA; B; C; DÞ gives k maximum value among A; B; C and D. This stage gives us top two maximum pixel progression directions which allows us encoding the maximum energy changes of the image in the specified direction. This step makes our pattern a strong encoder which is rotation invariant as well as illuminance invariant. This will improve the matching of critical changes in the image resulting in improved retrieval accuracy. This complex approach of encoding pixel progressions made our pattern superior compared to other simpler encoding approaches in literature.
d1 distance measure: The calculated histograms are stored as feature vectors of each image, we get feature vector database created out of image database jDBj denoted as F DBj ¼ ðF DBj1 ; F DBj2 ; :::; F DBjn Þ where, j ¼ 1; 2; :::jDBj and query image feature vector is denoted as F q ¼ ðF q1 ; F q2 ; :::; F qn Þ. Our primary aim is for a given query image; n best similar images are to be selected. For this purpose, we select n top matches to query image by calculating the distance of query image from images in given database jDBj. For similarity matching, d1 similarity distance metric is used which is computed using (13):
Dðq; DBÞ ¼
3.4. Pattern calculation
n F DBji F qi X i¼1
The directional maxima computations help us to extract first and second maxima sub-grids for the given reference grid. The reference 3 3 grid is extracted from original image using (9)
Iref ¼ Iðx þ t; y þ tÞ
ð9Þ
where t ¼ 1 : 1 Two gradient directional blocks are extracted in (10) from original image considering the reference grid Iref .
Ick ¼ Iðx m þ t; y þ n þ tÞ
ð10Þ
th
th
where F DBji is i feature of j image in the database jDBj and n is feature vector length. To illustrate why we preferred d1 distance measure, we compared retrieval results using d1 distance measure and other standard distance measures. The other distance measures compared are given in (14)–(16) as below: Manhattan (L1 or city-block) distance measure:
DðQ ; DBÞ ¼
N X f DBp;q f Q q
ð14Þ
q¼1
where for each ck ; m and n take values as given below
8 ck ¼ 1 ! m ¼ 2 > > >
ck ¼ 3 ! m ¼ 0 > > : ck ¼ 4 ! m ¼ 2
ð13Þ
1 þ F DBji þ F qi
9 n ¼ 0> > > n¼2=
Euclidean (L2) distance measure:
n ¼ 2> > > ; n¼2
DðQ ; DBÞ ¼
N 2 X f DBp;q f Q q
!1=2 ð15Þ
q¼1
The final LEOP is calculated using (11)
LEOPðx; yÞ ¼
P X
Canberra distance measure:
i
2 f 3 ðIref ðs; tÞ; Ic1 ðs; tÞ; Ic2 ðs; tÞÞ
ð11Þ
i¼0
where
f 3 ða; b; cÞ ¼
1 if a < c&b OR a > c&b 0 else
s; t ¼ 1 : 3 where Iref ; Ic1 ; Ic2 indicate reference, first and second maximum gradient directional grids respectively. P denotes number of surrounding neighbors and R is neighborhood radius. The detail procedure for pattern calculation is described in pictorial form in Fig. 2.
N f DBp;q f Q q X DðQ ; DBÞ ¼ q¼1 f DBp;q þ f Q q
ð16Þ
In traditional CBIR systems choosing the distance measure is as crucial as the robustness of the feature extractor. The evaluation measures are highly affected by the choice of different distance measures [44]. So we tested performance of our feature descriptor using different standard distance measures and the retrieval result analysis is given in experiment #6. 3.7. Proposed approach
3.5. Histogram calculation Input: Grayscale Image, Output: Feature Vector After pattern calculation from (11), histogram of the LEOP map of size ðX; YÞ is computed using (12)
HistðlÞ ¼
XX i2X
f 4 ðLEOPði; jÞ; lÞ
l 2 ½0; 511
ð12Þ
j2Y
where
f 4 ða; bÞ ¼
1 if a ¼ b 0
else
This histogram of the LEOP map is stored as feature vector for the image which is later used for similarity matching and distance calculation purpose.
1. Load input image 2. Calculate gradient magnitude and gradient directions at each pixel in image using Eqs. (3)–(6). 3. Quantize gradient directions in four directions n o þ900 ; þ450 ; 00 ; 450 using Eq. (7). 4. Calculate maximum pixel progression at each pixel using Eq. (8). 5. Find first and second maximum pixel progression at each pixel. 6. Code LEOP maps using Eq. (9)–(11) 7. Calculate LEOP histogram using Eq. (12).
6
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
4. Results and discussions We performed experiments on Four different databases namely MESSIDOR [47] database of retinal images, VIA/I-ELCAP [48] database of lung CT Images, COREL 5 k as well as COREL 10 k [49] natural image database and ImageNet [50] database of miscellaneous objects. In the first part, we carried out experimentations using traditional CBIR systems and in the second part, we carried our experimentations using learning based CBIR system. Abbreviations used for analysis of results are given below: Abbreviations DT-CWT + GGD-Dual tree complex wavelet transform [40] LBP: Local binary patterns [34] GLBP: LBP with Gabor transform [47] LDP: Local derivative patterns [23] GLDP: LDP with Gabor transform [23] LTP: Local ternary patterns [34] GLTP: LTP with Gabor transform [34] DLEP: Directional local extrema patterns [41] LEPSEG: Local edge patterns for segmentation [38] LEPINV: Rotational invariant local edge patterns [38] BLK_LBP: Block based LBP [39] CS_LBP: Circular symmetric local binary patterns [27] LBPu2: LBP with uniform patterns [20] LTCoP: Local ternary co-occurrence pattern [32] GLTCoP: LTCoP with Gabor transform [32] GLCM1: Gray level co-occurrence matrix1 [42] INTH: Intensity histogram [42] LTrP: Local tetra patterns [27] SS3D: Spherical symmetric 3-D local ternary patterns [35] ES-LBP: Edgy salient local binary pattern [43] VLDTP: Volumetric local directional triplet patterns for biomedical image retrieval [55]
patients of diabetic retinopathy. These images are divided into four groups based on the severity of disease. These images are available in three sizes 1440 960, 2240 1488 and 2304 1536. All the images are annotated with retinopathy grades as given in the Table 1. The specifications of retinopathy grading are based on number of microaneurysms, haemorrhages and the sign of neovascularisation. Images in which above abnormalities are absent, are considered as normal images. The results of the experiment are given in Table 2. The proposed method is compared with other methods in terms of group wise retrieval precision (ARP) as well as average retrieval precision of each method applied. Fig. 3. gives performance comparison of proposed method with other existing methods in terms of Group wise and Average precision. ARP has improved significantly compared to the previous methods. The plot of group wise ARP and Average precision is given in Fig. 3. Among the earlier published methods ES-LBP could get highest ARP with the feature vector length of 6 256 and our proposed descriptor has feature vector length of 1 512 i.e. with less computational complexity for retrieval we could get remarkable improvement in ARP.
Table 1 Number of images in each retinopathy grading group with specifications. Retinopathy Grading Groups
No. of Images
Specification
Group0 Group1 Group2
546 153 247
Group3 Total
254 1200
(Normal) (0 < mA 5) and (H = 0) (5
Table 2 Comparison of different methods in terms of group wise and average arp (%).
The performance evaluation in each experiment is done by obtaining precision, average retrieval precision (ARP), recall and average retrieval rate (ARR) which are calculated using (17)-(20).
ARP ¼
d 1X PðIi Þjn610 d i¼1
Recall; R ¼
ARR ¼
NR \ NRT nRT
NR \ NRT nR
d 1X RðIi ÞjnP10 d i¼1
ð17Þ
ð18Þ
Group0
Group1
Group2
Group3
Average
Complex Wavelet LBP DT-CWT + GGD VLDTP Wavelet ES-LBP LEOP
66.70 65.79 64.00 70.03 70.70 72.05 86.70
34.32 36.90 35.00 34.06 33.68 33.67 73.59
42.75 44.86 40.00 41.94 46.72 47.61 75.37
45.24 43.89 56.00 51.58 48.33 49.92 72.00
53.08 53.15 53.70 55.73 56.28 57.82 79.58
ð19Þ Complex Wavelet Wavelet
ð20Þ
where NR is set of all relevant images in the database, N RT is set of all retrieved images from database, (N R \ N RT ) gives total number of relevant images retrieved. Also nR is total number of relevant images in database, nRT is total number of retrieved images from the database and Ii is ith query image and d is total number of images in database.
LBP ES-LBP
DT-CWT+GGD LEOP
VLDTP
90
80
70
ARP (%)
Precision; P ¼
Method
60
50
40
4.1. Experiments by traditional CBIR system 30
4.1.1. Experiment 1: Performance analysis on MESSIDOR retinal image database In experiment #1, we applied our proposed method to MESSIDOR database which consists of 1200 retinal images captured from
Group0
Group1
Group2
Group3
Average
GROUPS
Fig. 3. Performance comparison of proposed method with other existing methods in terms of group wise and average precision on MESSIDOR database.
7
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
LTP
LBP
LTCoP
GLDP
GLTP
GLBP
GLTCoP
LEOP
95
95
90
90
85
85
80
80
75
75
ARP (%)
ARP (%)
LDP
70 65
LBPu2
LTPu2
LDPu2
GLBPu2
GLDPu2
LTCoPu2
GLTCoPu2
LEOP
GLTPu2
70 65
60
60
55
55
50
50
45
45
40
40 35
35 10
20
30
40
50
60
70
80
90
10
100
20
30
40
No. of top matches considered
50
60
70
80
90
100
No. of top matches considered
(a)
(b)
Fig. 4. (a)–(b) performance comparison of proposed method with other existing methods in terms of ARP on VIA/I-ELCAP-CT database.
4.1.2. Experiment 2: Performance analysis on VIA/I-ELCAP CT image database Experiment #2 is performed on VIA/I-ELCAP CT database which is a publicly available database jointly created by vision and image analysis group (VIA) and international early lung cancer program (I-ELCAP). These CT images are of 512 512 resolutions, are recorded in digital imaging and communications in medicine (DIACOM) format. We used 1000 such images of CT scans which are in total 10 scans of 100 images in each scan. Fig. 4(a)–(b) compares performance of proposed approach with existing approaches in terms of ARP on VIA/I-ELCAP CT database. Fig. 5(a)–(b) gives compares performance of proposed approach with existing approaches in terms of ARR on VIA/I-ELCAP CT database. In case of VIA/I-ELCAP CT database GLTCoP could get ARP of 89.41% and LEOP achieved 95.44% ARP, there is noticeable improvement in ARP. Pixel level energy mapping and local level feature encoding approach of our feature descriptor could fetch such a noticeable improvement in the results. 4.1.3. Experiment 3: Performance analysis on COREL-5K image database In experiment 3, the COREL-5K database is used for measurement of the retrieval accuracy of proposed algorithm in terms of ARP and ARR. COREL-5K contains fifty different types of images like animals, sports, bikes, etc., and each category contains 100 differ-
LBP
LTCoP
GLDP
GLTP
LDP
LTP
GLBP
GLTCoP
ent images (50x100). Fig. 6(a)–(b) compares performance of proposed approach with existing approaches in terms of ARP and ARR on COREL5K database. 4.1.4. Experiment 4: Performance analysis on COREL-10K image database In experiment 4, the Corel-10K database is used for measurement of the retrieval accuracy of proposed algorithm in terms of ARP, ARR. Corel-10K contains hundred different types of images like animals, sports, bikes, etc., and each category contains 100 different images (100 100). Fig. 7(a)–(b) shows comparative analysis of performance of proposed approach with existing approaches in terms of ARP and ARR on COREL10K database. 4.1.5. Experiment 5: Performance analysis on ImageNet database In experiment #5 we tested our pattern on ImageNet database [50] which is publicly available large database containing about 15 million images of different categories. For experimentation purpose we have used 80 classes with 1001 images in each class i.e. total 80,080 images. ImageNet database consists of set of images of different objects such as bags, balls, cow etc. Performance analysis of our proposed method on ImageNet database is done in terms of ARP and ARR. Fig. 8(a)–(b) gives comparison between performance analysis of proposed and existing approaches in terms of ARP and ARR on ImageNet database. The proposed descriptor is
LEOP
LDPu2
GLBPu2
GLTPu2
GLDPu2
LTPu2
LTCoPu2
GLTCoPu2
LEOP
LBPu2
61
62 55
57
49
52 47
ARR (%)
ARR (%)
43
37
31
42
37 32 27
25
22 19
17 13
12 7
7 10
20
30
40
50
60
70
No. of top matches Considered
(a)
80
90
100
10
20
30
40
50
60
70
80
No. of top matches Considered
(b)
Fig. 5. (a)–(b) performance comparison of proposed method with other existing methods in terms of ARR on VIA/I-ELCAP-CT database.
90
100
8
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
CS_LBP
LEPSEG
LEPINV
BLK_LBP
LBP
DLEP
LTP
CS_LBP
LEOP
56
51
LEPINV
BLK_LBP
LBP
DLEP
LTP
LOEP
21
46
19
41
17
ARR (%)
ARP (%)
LEPSEG
23
36
15 13
31
11 26
9 21
7 16
5
11
3 10
20
30
40
50 60 No. of top matches considered
70
80
90
100
10
20
30
40
50 60 No. of top matches considered
(a)
70
80
90
100
(b)
Fig. 6. (a)–(b) performance of proposed method with other existing methods in terms of ARP and ARR on COREL5K database.
CS_LBP
LEPSEG
LEPINV
BLK_LBP
LBP
DLEP
LTP
LEOP
CS_LBP LBP
LEPSEG DLEP
LEPINV LTP
BLK_LBP LEOP
20
50
18
45
16 14
35
ARR (%)
ARP (%)
40
30
12 10
25 8
20
6
15
4
10
2
10
20
30
40
50
60
70
80
90
100
10
20
30
40
50
60
70
80
90
100
No. of top matches considered
No. of top matches considered
(a)
(b)
Fig. 7. (a)–(b) gives performance comparison of proposed method with other existing methods in terms of ARP and ARR on COREL10K database.
GLBP
LBP
LDP
LTP
LTrP
GLBP
LEOP
Average Retrieval Rate (ARR) (%)
Average Retrieval Precision (ARP) (%)
LBP
LDP
LTP
LTrP
LEOP
30
40 35 30 25 20 15 10
25
20
15
10
5
5
0 100
200
300
400
500
600
700
800
900
1000
No. of top matches considered
(a)
0 100
200
300
400
500
600
700
800
900
1000
No. of top matches considered
(b)
Fig. 8. (a)–(b) gives performance analysis of proposed method with other existing methods in terms of ARP and ARR on ImageNet database.
able to find prominent edges in different pre-assumed directions. Pixel level tracking of intensity changes has effectively collected all the minute underlying texture information of image and hence the strong encoding of spatial information. The proposed descrip-
tor which is rotational invarient as well as illuminance independent, has achieved ARP of 31.02% for top 100 matches which is remarkable improvement in retrieval results compared to other existing methods, having ARP below 10%.
9
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
4.1.6. Experiment 6: Performance analysis of proposed method using different distance matrices In this experiment we tested performance of our proposed method using different distance measures named as L1-distance measure, L2-distance measure, Canberra distance measure and d1 distance measure. The d1 distance measure is improved form of Canberra distance measure i.e. in denominator 1 is added to avoid divide by zero condition. Performance of the proposed method with different distance measures is shown in Table 3. This comparison shows clearly that selection of d1 distance measure will leave positive impact on retrieval results.
Input: Image Database, Output: Retrieval Results
4.2. Experiments by learning based CBIR system In Traditional CBIR systems choosing the distance measure is as crucial as the robustness of the feature extractor. The evaluation measures are highly affected by the choice of different distance measures [44]. In above four experiments we applied traditional CBIR system on handcrafted features and the analysis of the results is done. We extended our work to designing of learning based CBIR system by using ANN to check the robustness of the proposed feature descriptor. To do so, we applied the extracted features from proposed feature descriptor to Artificial Neural Network (ANN). The parametric classifier like ANN can classify many complex patterns effectively and is popularly used in many machine learning applications for classification purpose [45,46,56,57]. 4.2.1. Experiment 1: Performance analysis on COREL-10K image database using learning based CBIR framework Here, we used bi-layer feed-forward-neural-network, tansigmoid activation function at hidden layers and softmax activation function at output layer. The hidden layer neurons are equal to 2/3rd of input layer neurons. For experimentation purpose we used COREL10K dataset. We trained ANN using stochastic gradient descend back propagation algorithm for that we divided the dataset images into training (80%), testing (10%) and validation (10%) sets. On the trained network we applied traditional CBIR framework. For a given query image we carried out classification using
Table 3 Performance of proposed method using different distance measures in terms of groupwise and average arp (%). Distance metrics
Group 0
Group 1
Group 2
Group 3
Average
L1 L2 Canberra d1 Distance
86.48 82.56 86.48 86.70
73.33 69.41 73.33 73.59
74.23 64.47 75.61 75.37
71.92 66.35 71.45 72.00
79.20 73.73 79.38 79.58
LBP
LTrP
LTP
trained network and decided top N-significant probable classes and their probabilities. Based on the top N-significant probabilities we selected number of images from respective probable class randomly i.e. probability of class decides the number of images to be selected from that class. On those selected set of images, we applied CBIR framework to retrieve top matches for given query image. The results for this learning-based approach are shown in Fig. 9. From the results it can be clearly seen that the proposed energy-based learning approach outperforms the existing stateof-art methods. Algorithm for learning based CBIR is given below:
SS3D
1. Divide the input Images into training (80%), testing (10%) and validation (10%) randomly sets for all classes and extract features. 2. Train the ANN using stochastic gradient descend back propagation algorithm. 3. Input a query image. 4. Extract features for query image. 5. Classify the query image using trained ANN and record ANN scores. 6. Sort the ANN scores in descending order to decide top N Probable classes (for this experiment we selected N = 3). 7. Collect images from top N probable classes randomly (i.e. probability of class decides the number of images to be selected from that class). 8. Apply traditional CBIR framework on collected images to get retrieval results.
We have conducted experiments in two different ways, on four different publicly available databases. In the first approach, we directly used the handcrafted features for retrieval purpose. In the second approach we used learning based retrieval technique. The rigorous experimentations are carried out to prove the ability of the proposed descriptor to encode the underlying texture information of different images such as rotated images, illuminance affected images etc. This became possible, as the proposed descriptor is able to find prominent edges in different pre-assumed directions. Pixel level tracking of intensity changes has effectively collected all the minute underlying texture information of the image and later local level encoding of all the collected information to create a compact feature vector has resulted into a strong and robust feature descriptor. The performance analysis of all the 6 experiments is given in Figs. 3–9 which clearly shows that there is a significant improvement in the results, compared to state-of-the-art methods. Table 4 Table 4 Feature vector length of query image for different methods.
LEOP
100
95
ARP (%)
90
85
80
75
70
65 1
2
3
4
5
6
7
8
9
10
No. of top matches considered
Fig. 9. Gives performance comparison of proposed method with other existing methods in terms of ARP on COREL10K database.
Method
Feature Vector Length
LBP LTP LDP DLEP GLBP_u2 GLDP VLDTP ES-LBP BLK_LBP DT-CWT GLMEBP GLBP LTrP SS3D LEOP
256 2 256 4 256 4 512 12 59 12 256 1 1024 6 256 16 256 104 3 4 512 12 256 13 59 10 256 512
10
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615
Table 5 Running time of different methods for a given query image.
Acknowledgment
Method
GLDP
GLBP
LTP
LBP
LDP
LEOP
Running Time (Seconds)
3.54
0.77
0.63
0.35
0.28
0.09
shows the comparison of feature vector length of image for all the compared methods. LEOP has feature vector length of 512 only, while other methods have larger feature vector length, due to this LEOP has less computational complexity in retrieval and we could get remarkable improvement in results with reduced computational complexity. We have chosen the Feature Vector Length (FVL) for the analysis of computational complexity because FVL is independent of hardware (PC) configuration. We tried to reduce the complexity by reducing the number of comparisons to be done for given query image in our novel learning based CBIR approach. In traditional CBIR approach we obtain top best matches for query image from the entire dataset (i.e.10,000 in this case of COREL 10 K database) whereas by applying ANN learning framework we first obtained a reduced set of classified images (i.e. few hundred images) which is later used for obtaining top best matches for a given query image. Because of less computational complexity, which is further reduced in learning based approach, the proposed method is well suited for online applications with smaller number of comparisons. By keenly observing both the ways of experimentations carried out in this paper i.e. traditional as well as learning based CBIR approach, the proposed feature descriptor has proved itself very much strong and robust in different environments. We tried to apply the proposed algorithm on a bigger database (i.e. ImageNet database) and the results were 31% only for top 100 retrieved images. Even other compared methods in literature could not do well on larger database. The actual size of ImageNet database is around 15 million images, but we tested our approach on 80,080 images only (total 80 class having 1001 images in each class). The performance may degrade if we go for retrieval from millions of stored images using traditional CBIR method. That’s why we employed learning-based retrieval technique to reduce number of comparisons based on the classification probabilities. The next step may be adopting deep learningbased approach for retrieval on huge databases having millions of images. The running time for a given query image from ImageNet database is shown in Table 5. The running time depends on many factors such as computer hardware specification, number of images in database, programming skills of individual etc. We computed running time on a system with intel(R) core (TM) i7-4770 CPU @ 3.4 GHz, 8 GB RAM, 64-bit windows 7 operating system. 5. Conclusion Local energy oriented pattern encodes the relationship of pixels from 3-D grid which is extracted from image considering the first maxima and second maxima of pixel progression. The pixel level mapping of energy changes and local level spatial encoding of information between three 3 3 extracted sub grids enabled the proposed descriptor to show remarkable improvement in the performance. Proposed pattern does not directly encode local neighborhood relationship, but it encodes spatial information by monitoring pixel level energy orientations to find minute underlying texture information of an image. Our experimentations in different environments i.e. traditional CBIR and learning based CBIR, have proved the proposed LEOP as a robust feature descriptor. The significant growth in experimental results verifies the effectiveness of the proposed descriptor in image indexing and retrieval.
We would like to convey sincere thanks to Mr. Prashant W. Patil, Mr. Akshay A. Dudhane (Research Scholars), Computer Vision and Pattern Recognition (CVPR) laboratory, IIT Ropar, Punjab, India and Mr. Kuldeep M. Biradar (Research Scholar) MNIT Jaipur, Rajasthan, India, for their valuable technical discussions during this work. We would like to express our gratitude towards the anonymous reviewers whose expert insights has helped to improve the quality of this manuscript. Declaration of Competing Interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. References [1] T. Dharani, I.L. Aroquiaraj, A survey on content based image retrieval, in: 2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering (PRIME), IEEE, 2013, pp. 485–490. [2] Y. Mistry, D.T. Ingole, Survey on content based image retrieval systems, Int. J. Innovative Res. Comput. Commun. Eng. 1 (8) (2013) 1828–1836. [3] M. Rehman, M. Iqbal, M. Sharif, M. Raza, Content based image retrieval: survey, World Appl. Sci. J. 19 (3) (2012) 404–412. [4] N. Singhai, S.K. Shandilya, A survey on: content based image retrieval systems, Int. J. Comput. Appl. 4 (2) (2010) 22–26. [5] G. Deep, L. Kaur, S. Gupta, Biomedical image indexing and retrieval descriptors: a comparative study, Procedia Comput. Sci. 1 (85) (2016 Jan) 954–961. [6] R.C. Veltkamp M. Tanase, Content-Based Image Retrieval Systems: A Survey, Dept. Computing Science, Utrecht University, Utrecht, The Netherlands, Tech. Rep., vol. 2, 2002, pp. 1–62. [7] Y. Rui, T.S. Huang, S.F. Chang, Image retrieval: Current techniques, promising directions, and open issues, J. Vis. Commun. Image Represent. 10 (1) (1999) 39–62. [8] A.W. Smeulders, M. Worring, S. Santini, A. Gupta, R. Jain, Content-based image retrieval at the end of the early years, IEEE Trans. Pattern Anal. Mach. Intell. 22 (12) (2000) 1349–1380. [9] M.J. Swain, D.H. Ballard, Color indexing, Int. J. Comput. Vision 7 (1) (1991) 11– 32. [10] M.J. Swain, D.H. Ballard, Indexing via color histograms, in: Active Perception and Robot Vision, Springer, Berlin, Heidelberg, 1992, pp. 261–273. [11] M.A. Stricker, M. Orengo, Similarity of color images, Storage and Retrieval for Image and Video Databases III, vol. 2420, International Society for Optics and Photonics, 1995, pp. 381–393. [12] J. Huang, S.R. Kumar, M. Mitra, Combining supervised learning with color correlograms for content-based image retrievalProceedings of the Fifth ACM International Conference on Multimedia, ACM, 1997, pp. 325–334. [13] E. Persoon, K.S. Fu, Shape discrimination using Fourier descriptors, IEEE Trans. Syst., Man, Cybern. 7 (3) (1977) 170–179. [14] Y. Rui, A.C. She, T.S. Huang, Modified Fourier descriptors for shape representation-a practical approach, in: Proc of First International Workshop on Image Databases and Multi Media Search, Citeseer, 1996, pp. 22–23. [15] A. Ahmadian, A. Mostafa, An efficient texture classification algorithm using Gabor wavelet, in: 2003. Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, IEEE, 2003, pp. 930–933. [16] B.S. Manjunath, W.Y. Ma, Texture features for browsing and retrieval of image data, IEEE Trans. Pattern Anal. Mach. Intell. 18 (8) (1996 Aug) 837–842. [17] M. Kokare, P.K. Biswas, B.N. Chatterji, Texture image retrieval using rotated wavelet filters, Pattern Recogn. Lett. 28 (10) (2007 Jul 15) 1240–1249. [18] M. Kokare, P.K. Biswas, B.N. Chatterji, Texture image retrieval using new rotated complex wavelet filters, IEEE Trans. Syst., Man, Cybern., Part B (Cybern.) 35 (6) (2005) 1168–1178. [19] M. Kokare, P.K. Biswas, B.N. Chatterji, Rotation-invariant texture image retrieval using rotated complex wavelet filters, IEEE Trans. Syst., Man, Cybern., Part B (Cybern.) 36 (6) (2006) 1273–1282. [20] T. Ojala, M. Pietikäinen, D. Harwood, A comparative study of texture measures with classification based on featured distributions, Pattern Recogn. 29 (1) (1996) 51–59. [21] Z. Guo, L. Zhang, D. Zhang, Rotation invariant texture classification using LBP variance (LBPV) with global matching, Pattern Recogn. 43 (3) (2010) 706–719. [22] Z. Guo, L. Zhang, D. Zhang, A completed modeling of local binary pattern operator for texture classification, IEEE Trans. Image Process. 19 (6) (2010) 1657–1663. [23] B. Zhang, Y. Gao, S. Zhao, J. Liu, Local derivative pattern versus local binary pattern: face recognition with high-order local pattern descriptor, IEEE Trans. Image Process. 19 (2) (2010 Feb) 533–544. [24] M. Subrahmanyam, R.P. Maheshwari, R. Balasubramanian, Local maximum edge binary patterns: a new descriptor for image retrieval and object tracking, Signal Process. 92 (6) (2012 Jun 1) 1467–1479.
G.M. Galshetwar et al. / J. Vis. Commun. Image R. 64 (2019) 102615 [25] M. Subrahmanyam, R.P. Maheshwari, R. Balasubramanian, Expert system design using wavelet and color vocabulary trees for image retrieval, Expert Syst. Appl. 39 (5) (2012) 5104–5114. [26] S. Murala, R.P. Maheshwari, R. Balasubramanian, Directional binary wavelet patterns for biomedical image indexing and retrieval, J. Med. Syst. 36 (5) (2012) 2865–2879. [27] S. Murala, R.P. Maheshwari, R. Balasubramanian, Local tetra patterns: a new feature descriptor for content-based image retrieval, IEEE Trans. Image Process. 21 (5) (2012) 2874–2886. [28] N. Jhanwar, S. Chaudhuri, G. Seetharaman, B. Zavidovique, Content based image retrieval using motif cooccurrence matrix, Image Vis. Comput. 22 (14) (2004) 1211–1220. [29] C.H. Lin, R.T. Chen, Y.K. Chan, A smart content-based image retrieval system based on color and texture feature, Image Vis. Comput. 27 (6) (2009) 658–665. [30] M. Subrahmanyam, Q.J. Wu, R.P. Maheshwari, R. Balasubramanian, Modified color motif co-occurrence matrix for image indexing and retrieval, Comput. Electr. Eng. 39 (3) (2013) 762–774. [31] S.K. Vipparthi, S.K. Nagar, Directional local ternary patterns for multimedia image indexing and retrieval, Int. J. Signal Imag. Syst. Eng. 8 (3) (2015) 137– 145. [32] S. Murala, Q.J. Wu, Local ternary co-occurrence patterns: a new feature descriptor for MRI and CT image retrieval, Neurocomputing 7 (119) (2013) 399–412. [33] S.K. Vipparthi, S.K. Nagar, Expert image retrieval system using directional local motif XoR patterns, Expert Syst. Appl. 41 (17) (2014) 8016–8026. [34] X. Tan, B. Triggs, Enhanced local texture feature sets for face recognition under difficult lighting conditions, IEEE Trans. Image Process. 19 (6) (2010) 1635– 1650. [35] S. Murala, Q.J. Wu, Spherical symmetric 3D local ternary patterns for natural, texture and biomedical image indexing and retrieval, Neurocomputing. 3 (149) (2015) 1502–1514. [36] S. Murala, Q.J. Wu, Local mesh patterns versus local binary patterns: biomedical image indexing and retrieval, IEEE J. Biomed. Health. Inf. 18 (3) (2014) 929–938. [37] S.K. Vipparthi, S. Murala, A.B. Gonde, Q.J. Wu, Local directional mask maximum edge patterns for image retrieval and face recognition, IET Comput. Vision 10 (3) (2016) 182–192. [38] M. Heikkilä, M. Pietikäinen, C. Schmid, Description of interest regions with local binary patterns, Pattern Recogn. 42 (3) (2009) 425–436. [39] C.H. Yao, S.Y. Chen, Retrieval of translated, rotated and scaled color textures, Pattern Recogn. 36 (4) (2003) 913–929. [40] C.G. Baby, D.A. Chandy, Content-based retinal image retrieval using dual-tree complex wavelet transform, in: 2013 International Conference on Signal Processing Image Processing & Pattern Recognition (ICSIPR), IEEE, 2013, pp. 195–199. [41] S. Murala, R.P. Maheshwari, R. Balasubramanian, Directional local extrema patterns: a new descriptor for content based image retrieval, Int. J. Multimedia Inform. Ret. 1 (3) (2012 Oct 1) 191–203.
11
[42] L. Sorensen, S.B. Shaker, M. De Bruijne, Quantitative analysis of pulmonary emphysema using local binary patterns, IEEE Trans. Med. Imaging 29 (2) (2010) 559–569. [43] G.M. Galshetwar, L.M. Waghmare, A.B. Gonde, S. Murala, Edgy salient local binary patterns in inter-plane relationship for image retrieval in Diabetic Retinopathy, Procedia Comput. Sci. 31 (115) (2017) 440–447. [44] M. Kokare, B.N. Chatterji, P.K. Biswas, Comparison of similarity metrics for texture image retrieval, in: TENCON 2003. Conference on Convergent Technologies for the Asia-Pacific Region, IEEE, 2003, pp. 571–575. [45] K. Ashizawa, T. Ishida, H. MacMahon, C.J. Vyborny, S. Katsuragawa, K. Doi, Artificial neural networks in chest radiography: application to the differential diagnosis of interstitial lung disease, Acad. Radiol. 1; 6(1) (1999) 2-9. [46] M. Anthimopoulos, S. Christodoulidis, L. Ebner, A. Christe, S. Mougiakakou, Lung pattern classification for interstitial lung diseases using a deep convolutional neural network, IEEE Trans. Med. Imaging 35 (5) (2016) 1207– 1216. [47] MESSIDOR Database Available at . [48] VIA/I-ELCAP-CT Lung Image Database, Available from . [49] Corel-10K Image Database Available at . [50] J. Deng, W. Dong, R. Socher, L.J. Li, K. Li, L. Fei-Fei, Imagenet: A large-scale hierarchical image database, in: IEEE Conference on Computer Vision and Pattern Recognition, 2009. CVPR 2009, IEEE, 2009, pp. 248–255. [51] G. Ding, J. Zhou, Y. Guo, Z. Lin, S. Zhao, J. Han, Large-scale image retrieval with sparse embedded hashing, Neurocomputing 27 (257) (2017) 24–36. [52] X. Lu, H. Yao, S. Zhao, X. Sun, S. Zhang, Action recognition with multi-scale trajectory-pooled 3D convolutional descriptors, Multimedia Tools Appl. 1–7 (2017). [53] S. Zhao, H. Yao, Y. Yang, Y. Zhang, Affective image retrieval via multi-graph learning, in: Proceedings of the 22nd ACM international conference on Multimedia, ACM, 2014, pp. 1025–1028. [54] K.M. Biradar, V. Kesana, K.B. Rakhonde, A. Sahu, A.B. Gonde, S. Murala, Local Gaussian difference extrema pattern: a new feature extractor for face recognition, in: 2017 Fourth International Conference on Image Information Processing (ICIIP), IEEE, 2017, pp. 1–5. [55] A.B. Gonde, P.W. Patil, G.M. Galshetwar, L.M. Waghmare, Volumetric local directional triplet patterns for biomedical image retrieval, in: 2017 Fourth International Conference on Image Information Processing (ICIIP), IEEE, 2017, pp. 1–6. [56] A.A. Dudhane, S.N. Talbar, Multi-scale directional mask pattern for medical image classification and retrieval, in: Proceedings of 2nd International Conference on Computer Vision & Image Processing, Springer, 2018, pp. 345–357. [57] A. Dudhane, G. Shingadkar, P. Sanghavi, B. Jankharia, S. Talbar, Interstitial lung disease classification using feed forward neural networks, ICCASP 2017 in Procedings of Advances in Intel-ligent Systems Research, vol. 137, 2017, pp. 515–521.