1
|
Automatic Number Plate Recognition:A Detailed Survey of Relevant Algorithms. SENSORS 2021; 21:s21093028. [PMID: 33925845 PMCID: PMC8123416 DOI: 10.3390/s21093028] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/17/2021] [Revised: 04/16/2021] [Accepted: 04/21/2021] [Indexed: 11/19/2022]
Abstract
Technologies and services towards smart-vehicles and Intelligent-Transportation-Systems (ITS), continues to revolutionize many aspects of human life. This paper presents a detailed survey of current techniques and advancements in Automatic-Number-Plate-Recognition (ANPR) systems, with a comprehensive performance comparison of various real-time tested and simulated algorithms, including those involving computer vision (CV). ANPR technology has the ability to detect and recognize vehicles by their number-plates using recognition techniques. Even with the best algorithms, a successful ANPR system deployment may require additional hardware to maximize its accuracy. The number plate condition, non-standardized formats, complex scenes, camera quality, camera mount position, tolerance to distortion, motion-blur, contrast problems, reflections, processing and memory limitations, environmental conditions, indoor/outdoor or day/night shots, software-tools or other hardware-based constraint may undermine its performance. This inconsistency, challenging environments and other complexities make ANPR an interesting field for researchers. The Internet-of-Things is beginning to shape future of many industries and is paving new ways for ITS. ANPR can be well utilized by integrating with RFID-systems, GPS, Android platforms and other similar technologies. Deep-Learning techniques are widely utilized in CV field for better detection rates. This research aims to advance the state-of-knowledge in ITS (ANPR) built on CV algorithms; by citing relevant prior work, analyzing and presenting a survey of extraction, segmentation and recognition techniques whilst providing guidelines on future trends in this area.
Collapse
|
2
|
Yang J, Li S, Wang Z, Dong H, Wang J, Tang S. Using Deep Learning to Detect Defects in Manufacturing: A Comprehensive Survey and Current Challenges. MATERIALS 2020; 13:ma13245755. [PMID: 33339413 PMCID: PMC7766692 DOI: 10.3390/ma13245755] [Citation(s) in RCA: 53] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/03/2020] [Revised: 12/05/2020] [Accepted: 12/07/2020] [Indexed: 12/18/2022]
Abstract
The detection of product defects is essential in quality control in manufacturing. This study surveys stateoftheart deep-learning methods in defect detection. First, we classify the defects of products, such as electronic components, pipes, welded parts, and textile materials, into categories. Second, recent mainstream techniques and deep-learning methods for defects are reviewed with their characteristics, strengths, and shortcomings described. Third, we summarize and analyze the application of ultrasonic testing, filtering, deep learning, machine vision, and other technologies used for defect detection, by focusing on three aspects, namely method and experimental results. To further understand the difficulties in the field of defect detection, we investigate the functions and characteristics of existing equipment used for defect detection. The core ideas and codes of studies related to high precision, high positioning, rapid detection, small object, complex background, occluded object detection and object association, are summarized. Lastly, we outline the current achievements and limitations of the existing methods, along with the current research challenges, to assist the research community on defect detection in setting a further agenda for future studies.
Collapse
Affiliation(s)
- Jing Yang
- School of Mechanical Engineering, Guizhou University, Guiyang 550025, China; (J.Y.); (Z.W.); (H.D.); (J.W.)
- Guizhou Provincial Key Laboratory of Public Big Data, Guizhou University, Guiyang 550025, China
| | - Shaobo Li
- School of Mechanical Engineering, Guizhou University, Guiyang 550025, China; (J.Y.); (Z.W.); (H.D.); (J.W.)
- Guizhou Provincial Key Laboratory of Public Big Data, Guizhou University, Guiyang 550025, China
- Key Laboratory of Advanced Manufacturing Technology of Ministry of Education, Guizhou University, Guiyang 550025, China;
- Correspondence:
| | - Zheng Wang
- School of Mechanical Engineering, Guizhou University, Guiyang 550025, China; (J.Y.); (Z.W.); (H.D.); (J.W.)
| | - Hao Dong
- School of Mechanical Engineering, Guizhou University, Guiyang 550025, China; (J.Y.); (Z.W.); (H.D.); (J.W.)
| | - Jun Wang
- School of Mechanical Engineering, Guizhou University, Guiyang 550025, China; (J.Y.); (Z.W.); (H.D.); (J.W.)
| | - Shihao Tang
- Key Laboratory of Advanced Manufacturing Technology of Ministry of Education, Guizhou University, Guiyang 550025, China;
| |
Collapse
|
3
|
Chang LB, Borenstein E, Zhang W, Geman S. Maximum likelihood features for generative image models. Ann Appl Stat 2017. [DOI: 10.1214/17-aoas1025] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
4
|
Wu T, Zhu SC. Learning Near-Optimal Cost-Sensitive Decision Policy for Object Detection. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2015; 37:1013-1027. [PMID: 26353325 DOI: 10.1109/tpami.2014.2359653] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Many popular object detectors, such as AdaBoost, SVM and deformable part-based models (DPM), compute additive scoring functions at a large number of windows in an image pyramid, thus computational efficiency is an important consideration in real time applications besides accuracy. In this paper, a decision policy refers to a sequence of two-sided thresholds to execute early reject and early accept based on the cumulative scores at each step. We formulate an empirical risk function as the weighted sum of the cost of computation and the loss of false alarm and missing detection. Then a policy is said to be cost-sensitive and optimal if it minimizes the risk function. While the risk function is complex due to high-order correlations among the two-sided thresholds, we find that its upper bound can be optimized by dynamic programming efficiently. We show that the upper bound is very tight empirically and thus the resulting policy is said to be near-optimal. In experiments, we show that the decision policy outperforms state-of-the-art cascade methods significantly, in several popular detection tasks and benchmarks, in terms of computational efficiency with similar accuracy of detection.
Collapse
|
5
|
Chakroborty S, Patil MA. Real-time arrhythmia classification for large databases. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2015; 2014:1448-51. [PMID: 25570241 DOI: 10.1109/embc.2014.6943873] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
In this paper we introduce a coarse-to-fine arrhythmia classification technique that can be used for efficient processing of large Electrocardiogram (ECG) records. This technique reduces time-complexity of arrhythmia classification by reducing size of the beats as well as by quantizing the number of beats using Multi-Section Vector Quantization (MSVQ) without compromising on the accuracy of the classification. The proposed solution is tested on MIT-BIH arrhythmia database. This work achieves a highest computational speed-up factor of 2.2:1 in comparison with standard arrhythmia classification technique with marginal loss (<;1%) in classification accuracy. The clinical application of this technique enhances physician's throughput by factor of 2x while processing large ECG records from Holter system.
Collapse
|
6
|
Zhu Q, Li Z, Liu J, Fan Z, Yu L, Chen Y. Improved minimum squared error algorithm with applications to face recognition. PLoS One 2013; 8:e70370. [PMID: 23936418 PMCID: PMC3735590 DOI: 10.1371/journal.pone.0070370] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2013] [Accepted: 06/17/2013] [Indexed: 11/18/2022] Open
Abstract
Minimum squared error based classification (MSEC) method establishes a unique classification model for all the test samples. However, this classification model may be not optimal for each test sample. This paper proposes an improved MSEC (IMSEC) method, which is tailored for each test sample. The proposed method first roughly identifies the possible classes of the test sample, and then establishes a minimum squared error (MSE) model based on the training samples from these possible classes of the test sample. We apply our method to face recognition. The experimental results on several datasets show that IMSEC outperforms MSEC and the other state-of-the-art methods in terms of accuracy.
Collapse
Affiliation(s)
- Qi Zhu
- Bio-Computing Center, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
- School of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai, China
- Key Laboratory of Network Oriented Intelligent Computation, Shenzhen, China
| | - Zhengming Li
- Bio-Computing Center, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
- Key Laboratory of Network Oriented Intelligent Computation, Shenzhen, China
- Guangdong Industrial Training Center, Guangdong Polytechnic Normal University, Guangzhou, China
| | - Jinxing Liu
- College of Information and Communication Technology, Qufu Normal University, Rizhao, China
| | - Zizhu Fan
- Bio-Computing Center, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
- School of Basic Science, East China Jiaotong University, Nanchang, China
| | - Lei Yu
- School of Urban Planning and Management, Harbin Institute of Technology Shenzhen Graduate School, Shenzhen, China
| | - Yan Chen
- Shenzhen Sunwin Intelligent Co., Ltd., Shenzhen, China
- * E-mail:
| |
Collapse
|
7
|
Sung J, Kim PJ, Ma S, Funk CC, Magis AT, Wang Y, Hood L, Geman D, Price ND. Multi-study integration of brain cancer transcriptomes reveals organ-level molecular signatures. PLoS Comput Biol 2013; 9:e1003148. [PMID: 23935471 PMCID: PMC3723500 DOI: 10.1371/journal.pcbi.1003148] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2012] [Accepted: 06/05/2013] [Indexed: 12/23/2022] Open
Abstract
We utilized abundant transcriptomic data for the primary classes of brain cancers to study the feasibility of separating all of these diseases simultaneously based on molecular data alone. These signatures were based on a new method reported herein – Identification of Structured Signatures and Classifiers (ISSAC) – that resulted in a brain cancer marker panel of 44 unique genes. Many of these genes have established relevance to the brain cancers examined herein, with others having known roles in cancer biology. Analyses on large-scale data from multiple sources must deal with significant challenges associated with heterogeneity between different published studies, for it was observed that the variation among individual studies often had a larger effect on the transcriptome than did phenotype differences, as is typical. For this reason, we restricted ourselves to studying only cases where we had at least two independent studies performed for each phenotype, and also reprocessed all the raw data from the studies using a unified pre-processing pipeline. We found that learning signatures across multiple datasets greatly enhanced reproducibility and accuracy in predictive performance on truly independent validation sets, even when keeping the size of the training set the same. This was most likely due to the meta-signature encompassing more of the heterogeneity across different sources and conditions, while amplifying signal from the repeated global characteristics of the phenotype. When molecular signatures of brain cancers were constructed from all currently available microarray data, 90% phenotype prediction accuracy, or the accuracy of identifying a particular brain cancer from the background of all phenotypes, was found. Looking forward, we discuss our approach in the context of the eventual development of organ-specific molecular signatures from peripheral fluids such as the blood. From a multi-study, integrated transcriptomic dataset, we identified a marker panel for differentiating major human brain cancers at the gene-expression level. The ISSAC molecular signatures for brain cancers, composed of 44 unique genes, are based on comparing expression levels of pairs of genes, and phenotype prediction follows a diagnostic hierarchy. We found that sufficient dataset integration across multiple studies greatly enhanced diagnostic performance on truly independent validation sets, whereas signatures learned from only one dataset typically led to high error rate. Molecular signatures of brain cancers, when obtained using all currently available gene-expression data, achieved 90% phenotype prediction accuracy. Thus, our integrative approach holds significant promise for developing organ-level, comprehensive, molecular signatures of disease.
Collapse
Affiliation(s)
- Jaeyun Sung
- Institute for Systems Biology, Seattle, Washington, United States of America
- Department of Chemical and Biomolecular Engineering, University of Illinois, Urbana, Illinois, United States of America
| | - Pan-Jun Kim
- Asia Pacific Center for Theoretical Physics, Pohang, Gyeongbuk, Republic of Korea
- Department of Physics, POSTECH, Pohang, Gyeongbuk, Republic of Korea
| | - Shuyi Ma
- Institute for Systems Biology, Seattle, Washington, United States of America
- Department of Chemical and Biomolecular Engineering, University of Illinois, Urbana, Illinois, United States of America
| | - Cory C. Funk
- Institute for Systems Biology, Seattle, Washington, United States of America
| | - Andrew T. Magis
- Institute for Systems Biology, Seattle, Washington, United States of America
- Center for Biophysics and Computational Biology, University of Illinois, Urbana, Illinois, United States of America
| | - Yuliang Wang
- Institute for Systems Biology, Seattle, Washington, United States of America
- Department of Chemical and Biomolecular Engineering, University of Illinois, Urbana, Illinois, United States of America
| | - Leroy Hood
- Institute for Systems Biology, Seattle, Washington, United States of America
| | - Donald Geman
- Institute for Computational Medicine, Johns Hopkins University, Baltimore, Maryland, United States of America
- Department of Applied Mathematics and Statistics, Johns Hopkins University, Baltimore, Maryland, United States of America
| | - Nathan D. Price
- Institute for Systems Biology, Seattle, Washington, United States of America
- * E-mail:
| |
Collapse
|
8
|
Barinova O, Lempitsky V, Kholi P. On detection of multiple object instances using Hough transforms. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2012; 34:1773-1784. [PMID: 22450818 DOI: 10.1109/tpami.2012.79] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
Hough transform-based methods for detecting multiple objects use nonmaxima suppression or mode seeking to locate and distinguish peaks in Hough images. Such postprocessing requires the tuning of many parameters and is often fragile, especially when objects are located spatially close to each other. In this paper, we develop a new probabilistic framework for object detection which is related to the Hough transform. It shares the simplicity and wide applicability of the Hough transform but, at the same time, bypasses the problem of multiple peak identification in Hough images and permits detection of multiple objects without invoking nonmaximum suppression heuristics. Our experiments demonstrate that this method results in a significant improvement in detection accuracy both for the classical task of straight line detection and for a more modern category-level (pedestrian) detection problem.
Collapse
Affiliation(s)
- Olga Barinova
- Lomonosov Moscow State University, Molodezhnaya str. 111, 119296 Moscow, Russia.
| | | | | |
Collapse
|
9
|
Hinterstoisser S, Cagniart C, Ilic S, Sturm P, Navab N, Fua P, Lepetit V. Gradient response maps for real-time detection of textureless objects. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2012; 34:876-888. [PMID: 22442120 DOI: 10.1109/tpami.2011.206] [Citation(s) in RCA: 78] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
We present a method for real-time 3D object instance detection that does not require a time-consuming training stage, and can handle untextured objects. At its core, our approach is a novel image representation for template matching designed to be robust to small image transformations. This robustness is based on spread image gradient orientations and allows us to test only a small subset of all possible pixel locations when parsing the image, and to represent a 3D object with a limited set of templates. In addition, we demonstrate that if a dense depth sensor is available we can extend our approach for an even better performance also taking 3D surface normal orientations into account. We show how to take advantage of the architecture of modern computers to build an efficient but very discriminant representation of the input images that can be used to consider thousands of templates in real time. We demonstrate in many experiments on real data that our method is much faster and more robust with respect to background clutter than current state-of-the-art methods.
Collapse
Affiliation(s)
- Stefan Hinterstoisser
- Department of Computer Aided Medical Procedures (CAMP), Technische Universität München, Garching bei München 85478, Germany.
| | | | | | | | | | | | | |
Collapse
|
10
|
Efficient Scale and Rotation Invariant Object Detection Based on HOGs and Evolutionary Optimization Techniques. ACTA ACUST UNITED AC 2012. [DOI: 10.1007/978-3-642-33179-4_22] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register]
|
11
|
Hough Regions for Joining Instance Localization and Segmentation. COMPUTER VISION – ECCV 2012 2012. [DOI: 10.1007/978-3-642-33712-3_19] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/25/2023]
|
12
|
Yang K, Wang M, Hua XS, Yan S, Zhang HJ. Assemble new object detector with few examples. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2011; 20:3341-3349. [PMID: 21632300 DOI: 10.1109/tip.2011.2158231] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
Learning a satisfactory object detector generally requires sufficient training data to cover the most variations of the object. In this paper, we show that the performance of object detector is severely degraded when training examples are limited. We propose an approach to handle this issue by exploring a set of pretrained auxiliary detectors for other categories. By mining the global and local relationships between the target object category and auxiliary objects, a robust detector can be learned with very few training examples. We adopt the deformable part model proposed by Felzenszwalb and simultaneously explore the root and part filters in the auxiliary object detectors under the guidance of the few training examples from the target object category. An iterative solution is introduced for such a process. The extensive experiments on the PASCAL VOC 2007 challenge data set show the encouraging performance of the new detector assembled from those related auxiliary detectors.
Collapse
Affiliation(s)
- Kuiyuan Yang
- Department of Automation, the University of Science and Technology of China, Hefei 230027, China
| | | | | | | | | |
Collapse
|
13
|
Bart E, Welling M, Perona P. Unsupervised organization of image collections: taxonomies and beyond. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2011; 33:2302-2315. [PMID: 21519098 DOI: 10.1109/tpami.2011.79] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
We introduce a nonparametric Bayesian model, called TAX, which can organize image collections into a tree-shaped taxonomy without supervision. The model is inspired by the Nested Chinese Restaurant Process (NCRP) and associates each image with a path through the taxonomy. Similar images share initial segments of their paths and thus share some aspects of their representation. Each internal node in the taxonomy represents information that is common to multiple images. We explore the properties of the taxonomy through experiments on a large (~10(4)) image collection with a number of users trying to locate quickly a given image. We find that the main benefits are easier navigation through image collections and reduced description length. A natural question is whether a taxonomy is the optimal form of organization for natural images. Our experiments indicate that although taxonomies can organize images in a useful manner, more elaborate structures may be even better suited for this task.
Collapse
Affiliation(s)
- Evgeniy Bart
- Palo Alto Research Center, Palo Alto, CA 94304, USA.
| | | | | |
Collapse
|
14
|
|
15
|
Zhu LL, Chen Y, Yuille A. Learning a hierarchical deformable template for rapid deformable object parsing. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2010; 32:1029-1043. [PMID: 20431129 DOI: 10.1109/tpami.2009.65] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/29/2023]
Abstract
In this paper, we address the tasks of detecting, segmenting, parsing, and matching deformable objects. We use a novel probabilistic object model that we call a hierarchical deformable template (HDT). The HDT represents the object by state variables defined over a hierarchy (with typically five levels). The hierarchy is built recursively by composing elementary structures to form more complex structures. A probability distribution--a parameterized exponential model--is defined over the hierarchy to quantify the variability in shape and appearance of the object at multiple scales. To perform inference--to estimate the most probable states of the hierarchy for an input image--we use a bottom-up algorithm called compositional inference. This algorithm is an approximate version of dynamic programming where approximations are made (e.g., pruning) to ensure that the algorithm is fast while maintaining high performance. We adapt the structure-perceptron algorithm to estimate the parameters of the HDT in a discriminative manner (simultaneously estimating the appearance and shape parameters). More precisely, we specify an exponential distribution for the HDT using a dictionary of potentials, which capture the appearance and shape cues. This dictionary can be large and so does not require handcrafting the potentials. Instead, structure-perceptron assigns weights to the potentials so that less important potentials receive small weights (this is like a "soft" form of feature selection). Finally, we provide experimental evaluation of HDTs on different visual tasks, including detection, segmentation, matching (alignment), and parsing. We show that HDTs achieve state-of-the-art performance for these different tasks when evaluated on data sets with groundtruth (and when compared to alternative algorithms, which are typically specialized to each task).
Collapse
Affiliation(s)
- Long Leo Zhu
- Massachusetts Institute of Technology, Cambridge, MA 02139, USA.
| | | | | |
Collapse
|
16
|
|
17
|
Cortés L, Amit Y. Efficient annotation of vesicle dynamics video microscopy. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2008; 30:1998-2010. [PMID: 18787247 DOI: 10.1109/tpami.2008.84] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2023]
Abstract
We describe an algorithm for the efficient annotation of events of interest in video microscopy. The specific application involves the detection and tracking of multiple p ossibly overlapping vesicles in total internal reflection fluorescent microscopy images. A st atistical model for the dynamic image data of vesicle configurations allows us to properly weight various hypotheses online. The goal is to find the most likely trajectories given a sequence of images. The computational challenge is addressed by defining a sequence of coarse-to-fine tests, derived from the statistical model, to quickly eliminate most candidate positions at each time frame. The computational load of the tests is initially very low and gradually in creases as the false positives become more difficult to eliminate. Only at the last step, state variables are estimated from a complete time- dependent model. Processing time thus mainly depends on the number of vesicles in the image and not on image size.
Collapse
Affiliation(s)
- Leandro Cortés
- Department of Computer Science, University of Chicago, 1100 E. 58th St., Chicago, IL 60637, USA.
| | | |
Collapse
|
18
|
|
19
|
Gavrila DM. A Bayesian, exemplar-based approach to hierarchical shape matching. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2007; 29:1408-21. [PMID: 17568144 DOI: 10.1109/tpami.2007.1062] [Citation(s) in RCA: 30] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/15/2023]
Abstract
This paper presents a novel probabilistic approach to hierarchical, exemplar-based shape matching. No feature correspondence is needed among exemplars, just a suitable pairwise similarity measure. The approach uses a template tree to efficiently represent and match the variety of shape exemplars. The tree is generated offline by a bottom-up clustering approach using stochastic optimization. Online matching involves a simultaneous coarse-to-fine approach over the template tree and over the transformation parameters. The main contribution of this paper is a Bayesian model to estimate the a posteriori probability of the object class, after a certain match at a node of the tree. This model takes into account object scale and saliency and allows for a principled setting of the matching thresholds such that unpromising paths in the tree traversal process are eliminated early on. The proposed approach was tested in a variety of application domains. Here, results are presented on one of the more challenging domains: real-time pedestrian detection from a moving vehicle. A significant speed-up is obtained when comparing the proposed probabilistic matching approach with a manually tuned nonprobabilistic variant, both utilizing the same template tree structure.
Collapse
Affiliation(s)
- Dariu M Gavrila
- Machine Perception Department of DainlerChrysler R&D, Ulm, Germany.
| |
Collapse
|
20
|
Allassonnière S, Amit Y, Trouvé A. Towards a coherent statistical framework for dense deformable template estimation. J R Stat Soc Series B Stat Methodol 2007. [DOI: 10.1111/j.1467-9868.2007.00574.x] [Citation(s) in RCA: 112] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
21
|
|
22
|
|
23
|
|