implemented forward and backward pass of "vanilla RNN", "LSTM" and "GRU" from scratch.
-
Updated
Sep 6, 2023 - Python
implemented forward and backward pass of "vanilla RNN", "LSTM" and "GRU" from scratch.
Predictes the posible full word based on given partial text.
Performed supervised learning using Vanilla RNN, Long Short Term Memory (LSTM) and Gated Recurrent Unit (GRU) models to predict sentiment of IMDb movie reviews data.
Sentence Sequence Transduction Library (Seq to Seq) for text generation using sequential generative Vanilla RNN using numpy
This is the assignment of lecture Deep Learning and Neural Network from KIT. To goal is to build an LSTM network without Tensorflow, Keras, and PyTorch.
English to Urdu Neural Machine Translation (NMT) using a Vanilla RNN Encoder-Decoder in PyTorch. An empirical study on sequence-to-sequence bottlenecks, vanishing gradients, and linguistic challenges in SOV languages without using LSTMs or Attention.
Rnn (vanial, GRU and LSTM) from scratch
Add a description, image, and links to the vanilla-rnn topic page so that developers can more easily learn about it.
To associate your repository with the vanilla-rnn topic, visit your repo's landing page and select "manage topics."