Skip to content

TheanoLM is a recurrent neural network language modeling tool implemented using Theano

License

Notifications You must be signed in to change notification settings

senarvi/theanolm

Repository files navigation

TheanoLM

Introduction

TheanoLM is a recurrent neural network language modeling tool implemented using the Python library Theano. Theano allows the user to customize and extend the neural network very conveniently, still generating highly efficient code that can utilize multiple GPUs or CPUs for parallel computation. TheanoLM allows the user to specify an arbitrary network architecture. New layer types and optimization methods can be easily implemented.

TheanoLM can be used for rescoring n-best lists and Kaldi lattices, decoding HTK word lattices, and generating text. It can be called from command line or from a Python script.

Implementations of many currently popular layer types are provided, such as long short-term memory (LSTM), gated recurrent units (GRU), bidirectional recurrent networks, gated linear units (GLU), and highway networks are provided. Several different Stochastic Gradient Descent (SGD) based optimizers are implemented, including RMSProp, AdaGrad, ADADELTA, and Adam.

There are several features that are especially useful with very large vocabularies. The effective vocabulary size can be reduced by using a class model. TheanoLM supports also subword vocabularies create e.g. using Morfessor. In addition to the standard cross-entropy cost, one can use sampling based noise-contrastive estimation (NCE) or BlackOut.

About the project

TheanoLM is open source and licensed under the Apache License, Version 2.0. The source code is available on GitHub. Documentation can be read online on Read the Docs.

Author

About

TheanoLM is a recurrent neural network language modeling tool implemented using Theano

Resources

License

Stars

Watchers

Forks

Packages

No packages published