Distilling GRU with Data Augmentation for Unconstrained Handwritten Text Recognition

2018 
Handwritten texts with various styles, such as horizontal, overlapping, vertical, and multi-lines texts, are commonly observed in the community. However, most existing handwriting recognition methods only concentrate on one specific kind of text style. In this paper, we focus on the problem of new unconstrained handwritten text recognition and propose distilling gated recurrent unit (GRU) with a new data augmentation technology to model the complex sequential dynamic of unconstrained handwriting text of various styles. The proposed data augmentation method can synthesize realistic handwritten text datasets including horizontal, vertical, overlap, right-down, screw-rotation, and multi-line situation, which render our framework robust for general purposes. The recommended distilling GRU can not only accelerate the training speed through the distilling stage but also maintain the original recognition accuracy. Experiments on our synthesized handwritten test sets show that the proposed multi-layer GRU performs well on the unconstrained handwriting text recognition problem. On the ICDAR2013 handwritten text recognition benchmark dataset, the proposed framework demonstrates comparable performance with state-of-the-art techniques.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    27
    References
    4
    Citations
    NaN
    KQI
    []