Today, high accuracy of character recognition is attainable using Neural Network for problems with relatively small number of categories. But for large categories, like Chinese characters, it is difficult to reach the neural network convergence because of the `local minima problem' and a large number of calculation. Studies are being done to solve the problem by splitting the neural network into some small modules. The effectiveness of the combination of Learning Vector Quantization (LVQ) and Back Propagation (BP) has been reported. LVQ is used for rough classification and BP is used for fine recognition. It is difficult to obtain high accuracy for rough classification by LVQ itself. In this paper, to deal with this problem, we propose Hierarchical Learning Vector Quantization (HLVQ). HLVQ divides categories in feature space hierarchically in learning procedure. The adjacent feature spaces overlap each other near the borders. HLVQ possesses both classification speed and accuracy due to the hierarchical architecture and the overlapping technique. In the experiment using ETL9B, the largest database of handwritten character in Japan, (includes 3036 categories, 607,200 samples), the effectiveness of HLVQ was verified.
|出版ステータス||Published - 1997 1月 1|
|イベント||Proceedings of the 1997 4th International Conference on Document Analysis and Recognition, ICDAR'97. Part 1 (of 2) - Ulm, Ger|
継続期間: 1997 8月 18 → 1997 8月 20
|Other||Proceedings of the 1997 4th International Conference on Document Analysis and Recognition, ICDAR'97. Part 1 (of 2)|
|Period||97/8/18 → 97/8/20|
ASJC Scopus subject areas
- コンピュータ ビジョンおよびパターン認識