Word Prediction Model using an elaborated RNN integrated with N-gram for efficient Text Input

Author:

Ikegami Yukino1,Tsuruta Setsuo2,Kutics Andrea3,Damiani Ernesto4,Knauf Rainer5

Affiliation:

1. IO Inc., Tokyo, Japan

2. Tokyo Denki University

3. International Christian University

4. Khalifa University of Science and Technology

5. Ilmenau University of Technology

Abstract

Abstract Smartphone users are beyond two billion worldwide. Heavy users of the texting application rely on input prediction to reduce typing effort. In languages based on the Roman alphabet, many techniques are available. However, Japanese text is based on multiple character sets such as Kanji, Hiragana and Katakana. For its time intensive input, next word prediction an open challenge. To tackle this, a hybrid language model is proposed. It integrates a Recurrent Neural Network (RNN) with an n-gram model. RNNs are powerful models for learning long sequences for next word prediction. N-gram models are best at current word completion. Our RNN language model predicts next words. According the “price” of the performance gain paid by a higher time complexity, our model best deploys on a client-server architecture. Heavily-loaded RNN-LM deploys on the server while the n-gram model on the client. It consists of an input layer equipped with word embedding, an output layer, and hidden layers connected with LSTMs (Long Short-Term Memories). Training is done via BPTT (Back Propagation Through Time). For robust training, BPTT is elaborated by learning rate refinement and gradient norm scaling. To avoid overfitting, the dropout technique is applied except for LSTM. Due to synergetic elaboration, it shows 10% lower perplexity than Zaremba’s excellent conventional models in our experiment. Our model has been incorporated into IME (Input Method Editor) we call Flick. In our experiment, Flick outperforms Mozc (Google Japanese Input) by 16% in time and 34% in the number of key strokes.

Publisher

Research Square Platform LLC

Reference62 articles.

1. Alex B (2005) An unsupervised system for identifying English inclusions in German text. Proc. of the ACL Student Research Workshop, Association for Computational Linguistics, pp. 133–138.

2. Alex B, Dubey A, Keller F (2007) Using foreign inclusion detection to improve parsing performance. Proc. of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, EMNLP-CoNLL, pp. 151–160.

3. Alsharif O, Ouyang T, Beaufays F, Zhai S, Breuel T, Schalkwyk J (2015) Long short term memory neural network for keyboard gesture decoding. Proc. of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2015, pp. 2076–2080.

4. An evolutionary algorithm that constructs recurrent neural networks;Angeline PJ;IEEE Transactions on Neural Networks,1994

5. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers;Ateniese A;International Journal of Security and Networks,2015

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3