Fast ML-based next-word prediction for hybrid languages

Yukino Ikegami, Setsuo Tsuruta, Andrea Kutics, Ernesto Damiani, Rainer Knauf

    Research output: Contribution to journalReview articlepeer-review

    1 Scopus citations

    Abstract

    Smartphone users are beyond two billion worldwide. Heavy users of the texting application rely on input prediction to reduce typing effort. In languages based on the Roman alphabet, many techniques are available. However, Japanese text is based on multiple character sets such as Kanji (Chinese-like word symbols), Hiragana and Katakana syllable sets. For its time/labor intensive input, next word prediction is crucial. It is still an open challenge. To tackle this, a hybrid language model is proposed. It integrates a Recurrent Neural Network (RNN) with an n-gram model. RNNs are powerful models for learning long sequences for next word prediction. N-gram models are best at current word completion. Our RNN language model (RNN-LM) predicts the next words. According the “price” of the performance gain paid by a higher time complexity, our model best deploys on a client-server architecture. Heavily-loaded RNN-LM deploys on the server while the n-gram model on the client. Our RNN-LM consists of an input layer equipped with word embedding, an output layer, and hidden layers connected with LSTMs (Long Short-Term Memories). Training is done via BPTT (Back Propagation Through Time). For robust training, BPTT is elaborated by learning rate refinement and gradient norm scaling. To avoid overfitting, the dropout technique is applied except for LSTM. Our novel model is compact (2 LSTMs, 650 units per layer), indeed. Due to synergetic elaboration, it shows 10 % lower perplexity than Zaremba's excellent conventional models in our Japanese text prediction experiment. Our model has been incorporated into IME (Input Method Editor) we call Flick. On the Japanese text input experiment, Flick outperforms Mozc (Google Japanese Input) by 16 % in time and 34 % in the number of keystrokes.

    Original languageBritish English
    Article number101064
    JournalInternet of Things (Netherlands)
    Volume25
    DOIs
    StatePublished - Apr 2024

    Keywords

    • Back propagation through time
    • Client-server model
    • Hybrid language model
    • Input method editor
    • Recurrent neural networks
    • Word prediction

    Fingerprint

    Dive into the research topics of 'Fast ML-based next-word prediction for hybrid languages'. Together they form a unique fingerprint.

    Cite this