Skip to content

Gated Orthogonal Recurrent Unit implementation in tensorflow

License

Notifications You must be signed in to change notification settings

jingli9111/GORU-tensorflow

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GORU-tensorflow

Gated Orthogonal Recurrent Unit

This model combines gating mechanism and orthogonal RNN approach. It solves forgetting problem and long-term dependency.

If you find this work useful, please cite arXiv:1706.02761.

Installation

requires TensorFlow 1.2.0

Usage

To use GORU in your model, simply copy goru.py.

Then you can use GORU in the same way you use built-in LSTM:

from goru import GORUCell
cell = GORUCell(hidden_size, capacity, fft)

Args:

  • hidden_size: Integer.
  • capacity: Optional. Integer. Only works for tunable style.
  • fft: Optional. Bool. If True, GORU is set to FFT style. Default is True.

Example tasks for GORU

We put two examples toy tasks: copying task and denoise task

Copying Memory Task

python copying_task.py --model GORU

Denoise Task

python denoise_task.py --model GORU

About

Gated Orthogonal Recurrent Unit implementation in tensorflow

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages