All Projects → campdav → Text Generation Using Bidirectional Lstm And Doc2vec Models

campdav / Text Generation Using Bidirectional Lstm And Doc2vec Models

Licence: apache-2.0
Text Generation using Bidirectional LSTM and Doc2Vec models

Projects that are alternatives of or similar to Text Generation Using Bidirectional Lstm And Doc2vec Models

Dslt
Deep Regression Tracking with Shrinkage Loss
Stars: ✭ 55 (-3.51%)
Mutual labels:  jupyter-notebook
Clr
Stars: ✭ 1,087 (+1807.02%)
Mutual labels:  jupyter-notebook
Pointseg
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
A Week In Wild Ai
360 view on ai/ml/dl applications
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Pybkb v2
Python scripts that help me be a successfull meteorologist. (Python 2) For Python 3, use: https://github.com/blaylockbk/pyBKB_v3
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Covidnet Ct
COVID-Net Open Source Initiative - Models and Data for COVID-19 Detection in Chest CT
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Imagenet
Trial on kaggle imagenet object localization by yolo v3 in google cloud
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Deep Belief Network Pytorch
This repository has implementation and tutorial for Deep Belief Network
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Codeforces Api
Tools for estimating problem difficulty, predictors rating trajectories, and tracking individual learning progress in algorithms.
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Niftidataset
pytorch dataset class for reading and transforming NIfTI files
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Endtoend Predictive Modeling Using Python
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Interpreting Decision Trees And Random Forests
Unwrapping decision trees and random forests to make them less of a black box
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Covid Nyc Dasymetric Map
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Tensorflow Machine Learning Projects
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Hypothesis Testing With Python
True difference or noise? 📊
Stars: ✭ 58 (+1.75%)
Mutual labels:  jupyter-notebook
Sccaf
Single-Cell Clustering Assessment Framework
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Amazon Sagemaker Safe Deployment Pipeline
Safe blue/green deployment of Amazon SageMaker endpoints using AWS CodePipeline, CodeBuild and CodeDeploy.
Stars: ✭ 56 (-1.75%)
Mutual labels:  jupyter-notebook
Polyaxon Examples
Code for polyaxon tutorials and examples
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Data Mining
Lecture slides and quizzes for Leskovec, Rajaraman, and Ullman's "Mining of Massive Datasets" Stanford course
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook
Cinemanet
Stars: ✭ 57 (+0%)
Mutual labels:  jupyter-notebook

Text-Generation-using-Bidirectional-LSTM-and-Doc2Vec-models

Text Generation using Bidirectional LSTM and Doc2Vec models

This repository contains jupyter notebooks and data, regarding this article published on Medium.

The purpose of this article is to discuss about text generation, using machine learning approaches, especially Recurrent Neural Networks (RNN) and Doc2vec.

Mainly, these approaches are using standard RNN such as LSTM (Long Short-Term Memory), and the are pretty fun to be experimented.

However, generated texts have a taste of unachievement. Generated sentences seem to be quite right, with correct grammar and syntax, as if the neural network was understanding correctly the structure of a sentence. But the whole new text does not have a great sense. And sometimes, has complete nonsense.

This result could come from the approach itself, using only LSTM to generate text, word by word.

In these notebooks, I will try to investigate a new way to generate sentences in a text generator solution. It does not mean that I will use something completely different from LTSM : I am not. I will use LTSM networks to generate sequences of words. However I will try to go further than a classic LSTM neural network and I will use an additional neural network (LSTM again), to select the best sentences in the text generation.

It will described :

  • how to train a neural network to generate sentences (i.e. sequences of words), based on existing novels. I will use a bidirectional LSTM Architecture to perform that.
  • how to train a doc2vec model to vectorize sentences,
  • how to train a neural network to select the best next sentence for a given paragraph (i.e. a sequence of sentences). I will also use a bidirectional LSTM architecture, in addition with a Doc2Vec model trained on the same target novels.

Content

This repo contains:

  • a data folder, with input files
  • three notebooks:
    • 1 generate sentence: train a bidirectional LSTM to generate sentences, word by word.
    • 2 select sentence: train a Doc2Vec model and a bidirectional LTSM to vectorize sentences  - 3 generate paragraph: combine all models to generate text
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].