Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper
-
Updated
Oct 30, 2021 - Python
Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper
Alternative approach for Adaptive Computation Time in TensorFlow
The implementation of Adaptive Computation function used in RNN "https://arxiv.org/pdf/1603.08983.pdf" and Universal Transformer Network "https://arxiv.org/pdf/1807.03819.pdf"
first attempt at description2code from 2016
Adaptive Computation Time (Graves, 2016, arXiv:1603.08983) wrapper for TensorFlow RNN cells.
Unofficial Implementation of Universal Transformer https://arxiv.org/abs/1807.03819
Adaptive Computation Time in Chainer
TensorFlow 2.X reimplementation of PonderNet: Learning to Ponder, Andrea Banino, Jan Balaguer, Charles Blundell.
A PyTorch implementation of adaptive computation time RNNs that's clean, idiomatic, and extensible.
GPTs trained with shakespeare dataset. Includes: small 10.8M GPT mimicking Andrej Karpathy's video lecture, Universal Transformer with Adaptive Computation Time
Add a description, image, and links to the adaptive-computation-time topic page so that developers can more easily learn about it.
To associate your repository with the adaptive-computation-time topic, visit your repo's landing page and select "manage topics."