Generative adversarial imitation learning to produce a proxy for the reward function present in dialogue.
-
Updated
Dec 25, 2020 - Jupyter Notebook
Generative adversarial imitation learning to produce a proxy for the reward function present in dialogue.
A Seq2Seq Attention chatbot deployed on Heroku
This repository is base on Pytorch Tutorial with some experiments and refined.
Some natural language processing networks from scratch in PyTorch for personal educational purposes.
Seq2Seq model that restores punctuation on English input text.
Seq2seq-attention house price prediction.
French to English neural machine translation trained on multi30k dataset.
I replicate and make the original Seq2Seq from PyTorch tutorials to be easy to use and adapt.
Experimentation of converting English to Pig Latin via a variety of Vanilla-seq2seq networks, Attention-mechanism based models and Transformer based Machine translation system.
Seq2Seq Neural Machine Translation Task, Completed as a part of CS462-NLP Coursework
Neural Machine Translation by Seq2Seq Model with Attention layer
A few approaches using sequence to sequence (seq2seq) architecture to solve semantice parsing problem
Sequence to sequence learning for GEC task using several deep models.
This repository contains the code for a speech to speech translation system created from scratch for digits translation from English to Tamil
Chatbot using Seq2Seq model using Tensorflow
基于Seq2Seq+Attention模型的Textsum文本自动摘要
Generates summary of a given news article. Used attention seq2seq encoder decoder model.
Sequence-to-sequence model implementations including RNN, CNN, Attention, and Transformers using PyTorch
load point forecast
Neural Machine Translation using LSTMs and Attention mechanism. Two approaches were implemented, models, one without out attention using repeat vector, and the other using encoder decoder architecture and attention mechanism.
Add a description, image, and links to the seq2seq-attn topic page so that developers can more easily learn about it.
To associate your repository with the seq2seq-attn topic, visit your repo's landing page and select "manage topics."