1
|
Kaur S, Bawa S, Kumar R. A survey of mono- and multi-lingual character recognition using deep and shallow architectures: indic and non-indic scripts. Artif Intell Rev 2020. [DOI: 10.1007/s10462-019-09720-9] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
2
|
Sun S, Liu H, Meng J, Chen CLP, Yang Y. Substructural Regularization With Data-Sensitive Granularity for Sequence Transfer Learning. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2018; 29:2545-2557. [PMID: 28504948 DOI: 10.1109/tnnls.2016.2638321] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/07/2023]
Abstract
Sequence transfer learning is of interest in both academia and industry with the emergence of numerous new text domains from Twitter and other social media tools. In this paper, we put forward the data-sensitive granularity for transfer learning, and then, a novel substructural regularization transfer learning model (STLM) is proposed to preserve target domain features at substructural granularity in the light of the condition of labeled data set size. Our model is underpinned by hidden Markov model and regularization theory, where the substructural representation can be integrated as a penalty after measuring the dissimilarity of substructures between target domain and STLM with relative entropy. STLM can achieve the competing goals of preserving the target domain substructure and utilizing the observations from both the target and source domains simultaneously. The estimation of STLM is very efficient since an analytical solution can be derived as a necessary and sufficient condition. The relative usability of substructures to act as regularization parameters and the time complexity of STLM are also analyzed and discussed. Comprehensive experiments of part-of-speech tagging with both Brown and Twitter corpora fully justify that our model can make improvements on all the combinations of source and target domains.
Collapse
|
3
|
Sun S, Yun J, Lin H, Zhang N, Abraham A, Liu H. Granular transfer learning using type-2 fuzzy HMM for text sequence recognition. Neurocomputing 2016. [DOI: 10.1016/j.neucom.2016.05.077] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
4
|
Jiang Z, Ding X, Peng L, Liu C. Exploring More Representative States of Hidden Markov Model in Optical Character Recognition: A Clustering-Based Model Pre-Training Approach. INT J PATTERN RECOGN 2015. [DOI: 10.1142/s0218001415500147] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
Hidden Markov Model (HMM) is an effective method to describe sequential signals in many applications. As to model estimation issue, common training algorithm only focuses on the optimization of model parameters. However, model structure influences system performance as well. Although some structure optimization methods are proposed, they are usually implemented as an independent module before parameter optimization. In this paper, the clustering feature of states in HMM is discussed through comparing the mechanism of Quadratic Discriminant Function (QDF) classifier and HMM. Then, through the clustering effect of Viterbi training and Baum–Welch training, a novel clustering-based model pre-training approach is proposed. It can optimize model parameters and model structure by turns, until the representative states of all models are explored. Finally, the proposed approach is evaluated on two typical OCR applications, printed and handwritten Arabic text line recognition. And it is compared with some other optimization methods. The improvement of character recognition performance proves the proposed approach can make more precise state allocation. And the representative states are benefit to HMM decoding.
Collapse
Affiliation(s)
- Zhiwei Jiang
- State Key Laboratory of Intelligent Technology and Systems, Tsinghua University, Beijing 100084, P. R. China
- Tsinghua National Laboratory for Information Science and Technology, Tsinghua University, Beijing 100084, P. R. China
- Department of Electronic Engineering, Tsinghua University, Beijing 100084, P. R. China
| | - Xiaoqing Ding
- State Key Laboratory of Intelligent Technology and Systems, Tsinghua University, Beijing 100084, P. R. China
- Tsinghua National Laboratory for Information Science and Technology, Tsinghua University, Beijing 100084, P. R. China
- Department of Electronic Engineering, Tsinghua University, Beijing 100084, P. R. China
| | - Liangrui Peng
- State Key Laboratory of Intelligent Technology and Systems, Tsinghua University, Beijing 100084, P. R. China
- Tsinghua National Laboratory for Information Science and Technology, Tsinghua University, Beijing 100084, P. R. China
- Department of Electronic Engineering, Tsinghua University, Beijing 100084, P. R. China
| | - Changsong Liu
- State Key Laboratory of Intelligent Technology and Systems, Tsinghua University, Beijing 100084, P. R. China
- Tsinghua National Laboratory for Information Science and Technology, Tsinghua University, Beijing 100084, P. R. China
- Department of Electronic Engineering, Tsinghua University, Beijing 100084, P. R. China
| |
Collapse
|