Click to open the HelpDesk interface
AECE - Front page banner

Menu:


FACTS & FIGURES

JCR Impact Factor: 0.700
JCR 5-Year IF: 0.700
SCOPUS CiteScore: 1.8
Issues per year: 4
Current issue: Nov 2024
Next issue: Feb 2025
Avg review time: 54 days
Avg accept to publ: 60 days
APC: 300 EUR


PUBLISHER

Stefan cel Mare
University of Suceava
Faculty of Electrical Engineering and
Computer Science
13, Universitatii Street
Suceava - 720229
ROMANIA

Print ISSN: 1582-7445
Online ISSN: 1844-7600
WorldCat: 643243560
doi: 10.4316/AECE


TRAFFIC STATS

3,016,453 unique visits
1,170,170 downloads
Since November 1, 2009



Robots online now
bingbot


SCOPUS CiteScore

SCOPUS CiteScore


SJR SCImago RANK

SCImago Journal & Country Rank




TEXT LINKS

Anycast DNS Hosting
MOST RECENT ISSUES

 Volume 24 (2024)
 
     »   Issue 4 / 2024
 
     »   Issue 3 / 2024
 
     »   Issue 2 / 2024
 
     »   Issue 1 / 2024
 
 
 Volume 23 (2023)
 
     »   Issue 4 / 2023
 
     »   Issue 3 / 2023
 
     »   Issue 2 / 2023
 
     »   Issue 1 / 2023
 
 
 Volume 22 (2022)
 
     »   Issue 4 / 2022
 
     »   Issue 3 / 2022
 
     »   Issue 2 / 2022
 
     »   Issue 1 / 2022
 
 
 Volume 21 (2021)
 
     »   Issue 4 / 2021
 
     »   Issue 3 / 2021
 
     »   Issue 2 / 2021
 
     »   Issue 1 / 2021
 
 
  View all issues  




SAMPLE ARTICLES

A Novel Control Approach Utilizing Neural Network for Efficient Microgrid Operation with Solar PV and Energy Storage Systems, JABBARI, A., KHAN, H., MUSHTAQ, D., SARWAR, M., DURAIBI, S., ALMALKI, K. J., AHMED, W., SIDDIQUI, A. S.
Issue 3/2024

AbstractPlus

UAV-Assisted Cooperative NOMA System with the nth Best Relay Selection, UMAKOGLU, I., NAMDAR, M., BASGUMUS, A.
Issue 3/2023

AbstractPlus

Droop Control Algorithm Design for Power Balancing in Island Inverter Based Microgrid, DRAGOUN, J., VINS, M., TALLA, J., BLAHNIK, V.
Issue 4/2022

AbstractPlus

Deep Learning Based DNS Tunneling Detection and Blocking System, ALTUNCU, M. A., GULAGIZ, F. K., OZCAN, H., BAYIR, O. F., GEZGIN, A., NIYAZOV, A., CAVUSLU, M. A., SAHIN, S.
Issue 3/2021

AbstractPlus

Reduction in Total Harmonic Distortion in Induction Motor Drives with High-Performance FPGA Controller, SUMAM, M. J., SHINY, G.
Issue 1/2022

AbstractPlus

Performance Analysis of Ryu-POX Controller in Different Tree-Based SDN Topologies, CABARKAPA, D., RANCIC, D.
Issue 3/2021

AbstractPlus




LATEST NEWS

2024-Jun-20
Clarivate Analytics published the InCites Journal Citations Report for 2023. The InCites JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.700 (0.700 without Journal self-cites), and the InCites JCR 5-Year Impact Factor is 0.600.

2023-Jun-28
Clarivate Analytics published the InCites Journal Citations Report for 2022. The InCites JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.800 (0.700 without Journal self-cites), and the InCites JCR 5-Year Impact Factor is 1.000.

2023-Jun-05
SCOPUS published the CiteScore for 2022, computed by using an improved methodology, counting the citations received in 2019-2022 and dividing the sum by the number of papers published in the same time frame. The CiteScore of Advances in Electrical and Computer Engineering for 2022 is 2.0. For "General Computer Science" we rank #134/233 and for "Electrical and Electronic Engineering" we rank #478/738.

2022-Jun-28
Clarivate Analytics published the InCites Journal Citations Report for 2021. The InCites JCR Impact Factor of Advances in Electrical and Computer Engineering is 0.825 (0.722 without Journal self-cites), and the InCites JCR 5-Year Impact Factor is 0.752.

2022-Jun-16
SCOPUS published the CiteScore for 2021, computed by using an improved methodology, counting the citations received in 2018-2021 and dividing the sum by the number of papers published in the same time frame. The CiteScore of Advances in Electrical and Computer Engineering for 2021 is 2.5, the same as for 2020 but better than all our previous results.

Read More »


    
 

  2/2020 - 12

Generation of Visual Patterns from BoVW for Image Retrieval using modified Similarity Score Fusion

ARULMOZHI, P. See more information about ARULMOZHI, P. on SCOPUS See more information about ARULMOZHI, P. on IEEExplore See more information about ARULMOZHI, P. on Web of Science, ABIRAMI, M. See more information about ABIRAMI, M. on SCOPUS See more information about ABIRAMI, M. on SCOPUS See more information about ABIRAMI, M. on Web of Science
 
Extra paper information in View the paper record and citations in Google Scholar View the paper record and similar papers in Microsoft Bing View the paper record and similar papers in Semantic Scholar the AI-powered research tool
Click to see author's profile in See more information about the author on SCOPUS SCOPUS, See more information about the author on IEEE Xplore IEEE Xplore, See more information about the author on Web of Science Web of Science

Download PDF pdficon (1,519 KB) | Citation | Downloads: 975 | Views: 2,125

Author keywords
feature extraction, image fusion, image matching, image representation, supervised learning

References keywords
image(36), retrieval(19), visual(14), vision(13), recognition(12), pattern(11), cvpr(11), words(9), fusion(8), classification(8)
Blue keywords are present in both the references section and the paper title.

About this article
Date of Publication: 2020-05-31
Volume 20, Issue 2, Year 2020, On page(s): 101 - 112
ISSN: 1582-7445, e-ISSN: 1844-7600
Digital Object Identifier: 10.4316/AECE.2020.02012
Web of Science Accession Number: 000537943500012
SCOPUS ID: 85087452294

Abstract
Quick view
Full text preview
The Bag of Visual Words (BoVW) turns up to be an efficient method to represent images for Content Based Image Retrieval (CBIR). Despite their significant usage, the traditional BoVW method has low discriminative power and fails to provide spatial information, which increases the false positive images and reduces the precision values. To address the first issue, a novel way of identifying a set of visual words unique for each category, named as Visual Patterns (VP) is proposed. Also, the weight for the respective VPs and a new way of score calculations for similarity matching with the database images are proposed. Then, to address the second issue of enhancing the spatial information, late fusion of Gabor filter features along with VP is proposed. As a consequence, VP provides better discriminative power and Gabor filtering, taking advantage of its complementary clue, provides spatial information. Hence, it helps to reduce the false matches and improves the precision values. Experiments are carried out on the popular datasets, namely, Caltech 256, Oxford 5K and Inria Holidays datasets along with Flickr 1M dataset. The proposed method is compared with other BoVW based models and proved that the MAP value is improved 0.50 times from the basic BoVW model.


References | Cited By  «-- Click to see who has cited this paper

[1] Y. Rui, T.S. Huang, and S. F. Chang, "Image retrieval: Current techniques, promising directions, and open issues," Journal of visual communication and image representation, vol.10, no. 1, pp. 39-62, 1999.
[CrossRef] [Web of Science Times Cited 1198] [SCOPUS Times Cited 1589]


[2] R. Datta, D. Joshi, J. Li and J. Z. Wang, "Image retrieval: Ideas, influences, and trends of the new age," ACM Computing Surveys (Csur), vol.40, no.2, pp.5, 2008.
[CrossRef] [Web of Science Times Cited 1729] [SCOPUS Times Cited 2661]


[3] M. Rehman, M. Iqbal, M. Sharif and M. Raza, "Content based image retrieval: survey," World Applied Sciences Journal, vol.19, no. 3, pp. 404-412, 2012.
[CrossRef] [SCOPUS Times Cited 48]


[4] A. Alzubi, A. Amira and N. Ramzan, "Semantic content-based image retrieval: A comprehensive study," Journal of Visual Communication and Image Representation, vol.32, pp. 20-54, July 2015.
[CrossRef] [Web of Science Times Cited 133] [SCOPUS Times Cited 181]


[5] W. Zhou, H. Li and Q. Tian, "Recent advance in content-based image retrieval: A literature survey," arXiv preprint arXiv: 1706.06064 2017.

[6] J. Sivic and A. Zisserman, "Video Google: a text retrieval approach to object matching in videos," Proceedings Ninth IEEE International Conference on Computer Vision, vol.2, pp. 1470-1477, 2003.
[CrossRef] [Web of Science Times Cited 3561] [SCOPUS Times Cited 5634]


[7] G. Scurka, L. Fan, C. R. Dance and C. Brey, "Visual categorization with bags of keypoints," In Workshop on statistical learning in computer vision, ECCV, vol. 1, no. 1-22, pp. 1-2, 2004.

[8] D. Nister and H. Stewenius, "Scalable recognition with a vocabulary tree," In 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06), vol. 2, pp. 2161-2168, 2006.
[CrossRef] [SCOPUS Times Cited 3239]


[9] H. Jegou, M. Douze, C. Schmid and P. Perez, "Aggregating local descriptors into a compact image representation," In CVPR 2010-23rd IEEE Conference on Computer Vision & Pattern Recognition, pp. 3304-3311,2010.
[CrossRef] [Web of Science Times Cited 1653] [SCOPUS Times Cited 2273]


[10] Y. Mu, J. Sun, T. X. Han, L. F. Cheong and S. Yan, "Randomized locality sensitive vocabularies for bag-of-features model," In European Conference on Computer Vision, pp. 748-761, 2010.
[CrossRef] [SCOPUS Times Cited 19]


[11] Q. Huang et al., "Building contextual visual vocabulary for large-scale image applications," In Proceedings of the 18th ACM international conference on Multimedia, ACM, pp. 501-510, 2010.
[CrossRef] [Web of Science Times Cited 73] [SCOPUS Times Cited 109]


[12] N. Ali, K. Bajwa, R. Sablatnig and Z. Mehmood, "Image retrieval by addition of spatial information based on histograms of triangular regions," Computers & Electrical Engineering, vol.54, pp. 539-550, 2016.
[CrossRef] [Web of Science Times Cited 57] [SCOPUS Times Cited 67]


[13] W. Li, P. Dong, B. Xiao and L. Zhou, "Object recognition based on the region of interest and optimal bag of words model," Neurocomputing, vol. 172, pp. 271-280, 2016.
[CrossRef]


[14] A. K. Tiwari, V. Kanhangad and R. B. Pachori, "Histogram refinement for texture descriptor based image retrieval," Signal Processing: Image Communication, vol.53, pp.73-85, 2017.
[CrossRef] [Web of Science Times Cited 28] [SCOPUS Times Cited 38]


[15] A. Bolovinou, I. Pratikakis and S. Perantonis, "Bag of spatio-visual words for context inference in scene classification," Pattern Recognition, vol.46, no.3 pp. 1039-1053, 2013.
[CrossRef]


[16] Z. Zhou, Q. M. Jonathan Wu and X. Sun, "Encoding multiple contextual clues for partial-duplicate image retrieval," Pattern Recognition Letters, vol.109, pp.18-26, 2018.
[CrossRef]


[17] R. Wang, K. Ding, J. Yang and L. Xue, "A novel method for image classification based on bag of visual words," Journal of Visual Communication and Image Representation, vol. 40, pp. 24-33, 2016.
[CrossRef] [Web of Science Times Cited 18] [SCOPUS Times Cited 24]


[18] W. C. Lin, C. F. Tsai, Z. Y. Chen and S. W. Ke, "Keypoint selection for efficient bag-of-words feature generation and effective image classification," Information Sciences, vol. 329, pp. 33-51, 2016.
[CrossRef] [Web of Science Times Cited 31] [SCOPUS Times Cited 38]


[19] R. Khan, C. Barat, D. Muselet and C. Ducottet, "Spatial histograms of soft pairwise similar patches to improve the bag-of-visual-words model," Computer Vision and Image Understanding, vol. 132. pp. 102-112, 2015,
[CrossRef] [Web of Science Times Cited 32] [SCOPUS Times Cited 37]


[20] E. Gavves, C. G. M. Snoek and A. W. M. Smeulders, "Visual synonyms for landmark image retrieval," Computer Vision and Image Understanding, vol.116, no. 2, pp. 238-249, 2012.
[CrossRef] [Web of Science Times Cited 12] [SCOPUS Times Cited 14]


[21] F. B. Silva, R. Werneck, S. Goldenstein, S. Tobbone and R. Torres, "Graph-based bag-of-words for classification," Pattern Recognition, vol.74, pp. 266-285, 2018.
[CrossRef] [Web of Science Times Cited 47] [SCOPUS Times Cited 51]


[22] I. Dimitrovski, D. Kocev, S. Loskovska and S. Dzeroski, "Improving bag-of-visual-words image retrieval with predictive clustering trees," Information Sciences, vol. 329, pp. 851-865, 2016.
[CrossRef] [Web of Science Times Cited 36] [SCOPUS Times Cited 40]


[23] C. G. M. Snoek, M. Worring and A. W. M. Smeulders, "Early versus late fusion in semantic video analysis," In Proceedings of the 13th annual ACM international conference on Multimedia, pp. 399-402, 2005.
[CrossRef] [SCOPUS Times Cited 665]


[24] P. A. S. Kimura, J. M. B. Cavalcanti, P. S. Saraiva, R. S. Torres and M. A. Goncalves, "Evaluating retrieval effectiveness of descriptors for searching in large image databases," Journal of information and data management vol.2, no. 3, 2011.

[25] M. A. Herraez, J. Domingo and F. J. Ferri, "Combining similarity measures in content-based image retrieval," Pattern Recognition Letters, vol.29, no. 16, pp. 2174-2181, 2008.
[CrossRef]


[26] H. J. Escalante, C. A. Hernadez, L. E. Sucar and M. Montes, "Late fusion of heterogeneous methods for multimedia image retrieval," In Proceedings of the 1st ACM international conference on Multimedia information retrieval, pp. 172-179, 2008.
[CrossRef]


[27] L. Piras and G. Giacinto, "Information fusion in content based image retrieval: A comprehensive overview," Information Fusion, vol. 37, pp. 50-60, 2017.
[CrossRef] [Web of Science Times Cited 85] [SCOPUS Times Cited 121]


[28] Y. Xu and Y. Lu, "Adaptive weighted fusion: A novel fusion approach for image classification," Neurocomputing, vol.168, pp. 566-574, 2015.
[CrossRef] [Web of Science Times Cited 62] [SCOPUS Times Cited 67]


[29] C. Jun, J. Shao, X. Xu, D. Ouyang and L. Gao, "Exploiting score distribution for heterogenous feature fusion in image classification," Neurocomputing, vol. 253, pp. 70-76, 2017.
[CrossRef]


[30] Z. Liang et al., "Query-adaptive late fusion for image search and person re-identification," In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1741-1750, 2015.
[CrossRef] [SCOPUS Times Cited 285]


[31] G. Griffin, A. Holub and P. Perona, "Caltech-256 object category dataset," 2007.

[32] J. Philbin, O. Chum, M. Isard, J. Sivic and A. Zisserman, "Object retrieval with large vocabularies and fast spatial matching," In 2007 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1-8, 2007.
[CrossRef] [Web of Science Times Cited 1515] [SCOPUS Times Cited 2585]


[33] H. Jegou, M. Douze and C. Schmid, "Hamming embedding and weak geometric consistency for large scale image search," In European conference on computer vision, Springer, Berlin, pp. 304-317, 2008.
[CrossRef] [SCOPUS Times Cited 1235]


[34] H. Jegou, M. Douze and C. Schmid, "Improving bag-of-features for large scale image search," International journal of computer vision, vol. 87, no. 3, pp. 316-336, 2010.

[35] J. M. Dos Santos, E. S. De Moura, A. S. De Silva and R. S. Torres, "A signature-based bag of visual words method for image indexing and search," Pattern Recognition Letters, vol. 65, pp. 1-7, 2015.
[CrossRef] [Web of Science Times Cited 13] [SCOPUS Times Cited 14]


[36] F. Perronnin et al., "Aggregating local image descriptors into compact codes," IEEE transactions on pattern analysis and machine intelligence, vol. 34, no. 9, pp. 1704-1716, 2011.
[CrossRef]


[37] Y. Zhang, Z. Jia and T. Chen, "Image retrieval with geometry-preserving visual phrases," In CVPR 2011, pp. 809-816, 2011.
[CrossRef] [SCOPUS Times Cited 274]


[38] C. Yang, C. Wang, Z. Li and L. Zhang, "Spatial-bag-of-features," IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3352-3359, 2010.
[CrossRef] [Web of Science Times Cited 122] [SCOPUS Times Cited 221]


[39] G. Amato, F. Falchi and L. Vadicamo, "Aggregating binary local descriptors for image retrieval," Multimedia Tools and Applications, vol. 77, no. 5, pp. 5385-5415, 2018.
[CrossRef] [Web of Science Times Cited 5] [SCOPUS Times Cited 10]


[40] R. Wang, K. Ding, J. Yang and L. Xue, "A novel method for image classification based on bag of visual words," Journal of Visual Communication and Image Representation, vol.40 pp. 24-33, 2016.
[CrossRef] [Web of Science Times Cited 18] [SCOPUS Times Cited 24]


[41] J. Yang, K. Yu, Y. Gong and T.S. Huang, "Linear spatial pyramid matching using sparse coding for image classification," CVPR, vol. 1, no. 2, pp. 6, 2009.
[CrossRef]


[42] M. D. Zeiler and R. Fergus, "Visualizing and understanding convolutional networks," In European conference on computer vision, pp.818-833, 2014.



References Weight

Web of Science® Citations for all references: 10,428 TCR
SCOPUS® Citations for all references: 21,563 TCR

Web of Science® Average Citations per reference: 243 ACR
SCOPUS® Average Citations per reference: 501 ACR

TCR = Total Citations for References / ACR = Average Citations per Reference

We introduced in 2010 - for the first time in scientific publishing, the term "References Weight", as a quantitative indication of the quality ... Read more

Citations for references updated on 2024-11-26 05:46 in 243 seconds.




Note1: Web of Science® is a registered trademark of Clarivate Analytics.
Note2: SCOPUS® is a registered trademark of Elsevier B.V.
Disclaimer: All queries to the respective databases were made by using the DOI record of every reference (where available). Due to technical problems beyond our control, the information is not always accurate. Please use the CrossRef link to visit the respective publisher site.

Copyright ©2001-2024
Faculty of Electrical Engineering and Computer Science
Stefan cel Mare University of Suceava, Romania


All rights reserved: Advances in Electrical and Computer Engineering is a registered trademark of the Stefan cel Mare University of Suceava. No part of this publication may be reproduced, stored in a retrieval system, photocopied, recorded or archived, without the written permission from the Editor. When authors submit their papers for publication, they agree that the copyright for their article be transferred to the Faculty of Electrical Engineering and Computer Science, Stefan cel Mare University of Suceava, Romania, if and only if the articles are accepted for publication. The copyright covers the exclusive rights to reproduce and distribute the article, including reprints and translations.

Permission for other use: The copyright owner's consent does not extend to copying for general distribution, for promotion, for creating new works, or for resale. Specific written permission must be obtained from the Editor for such copying. Direct linking to files hosted on this website is strictly prohibited.

Disclaimer: Whilst every effort is made by the publishers and editorial board to see that no inaccurate or misleading data, opinions or statements appear in this journal, they wish to make it clear that all information and opinions formulated in the articles, as well as linguistic accuracy, are the sole responsibility of the author.




Website loading speed and performance optimization powered by: 


DNS Made Easy