All Projects → RubensZimbres → Repo 2017

RubensZimbres / Repo 2017

Python codes in Machine Learning, NLP, Deep Learning and Reinforcement Learning with Keras and Theano

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Repo 2017

Datastories Semeval2017 Task4
Deep-learning model presented in "DataStories at SemEval-2017 Task 4: Deep LSTM with Attention for Message-level and Topic-based Sentiment Analysis".
Stars: ✭ 184 (-83.62%)
Mutual labels:  sentiment-analysis, glove, keras-models
Text2vec
Fast vectorization, topic modeling, distances and GloVe word embeddings in R.
Stars: ✭ 715 (-36.33%)
Mutual labels:  natural-language-processing, word2vec, glove
Gpnd
Generative Probabilistic Novelty Detection with Adversarial Autoencoders
Stars: ✭ 112 (-90.03%)
Mutual labels:  generative-adversarial-network, anomaly-detection, autoencoder
Textclf
TextClf :基于Pytorch/Sklearn的文本分类框架,包括逻辑回归、SVM、TextCNN、TextRNN、TextRCNN、DRNN、DPCNN、Bert等多种模型,通过简单配置即可完成数据处理、模型训练、测试等过程。
Stars: ✭ 105 (-90.65%)
Mutual labels:  sentiment-analysis, word2vec, glove
sarcasm-detection-for-sentiment-analysis
Sarcasm Detection for Sentiment Analysis
Stars: ✭ 21 (-98.13%)
Mutual labels:  sentiment-analysis, word2vec, glove
Rectorch
rectorch is a pytorch-based framework for state-of-the-art top-N recommendation
Stars: ✭ 121 (-89.23%)
Mutual labels:  generative-adversarial-network, autoencoder, variational-autoencoder
Focal Frequency Loss
Focal Frequency Loss for Generative Models
Stars: ✭ 141 (-87.44%)
Mutual labels:  generative-adversarial-network, autoencoder, variational-autoencoder
Repo 2016
R, Python and Mathematica Codes in Machine Learning, Deep Learning, Artificial Intelligence, NLP and Geolocation
Stars: ✭ 103 (-90.83%)
Mutual labels:  natural-language-processing, word2vec, autoencoder
Awesome Tensorlayer
A curated list of dedicated resources and applications
Stars: ✭ 248 (-77.92%)
Mutual labels:  natural-language-processing, generative-adversarial-network, autoencoder
Magnitude
A fast, efficient universal vector embedding utility package.
Stars: ✭ 1,394 (+24.13%)
Mutual labels:  natural-language-processing, word2vec, glove
Natural Language Processing
Programming Assignments and Lectures for Stanford's CS 224: Natural Language Processing with Deep Learning
Stars: ✭ 377 (-66.43%)
Mutual labels:  natural-language-processing, word2vec, glove
Languagecrunch
LanguageCrunch NLP server docker image
Stars: ✭ 281 (-74.98%)
Mutual labels:  natural-language-processing, sentiment-analysis, word2vec
Ad examples
A collection of anomaly detection methods (iid/point-based, graph and time series) including active learning for anomaly detection/discovery, bayesian rule-mining, description for diversity/explanation/interpretability. Analysis of incorporating label feedback with ensemble and tree-based detectors. Includes adversarial attacks with Graph Convolutional Network.
Stars: ✭ 641 (-42.92%)
Mutual labels:  generative-adversarial-network, anomaly-detection, autoencoder
Anime Inpainting
An application tool of edge-connect, which can do anime inpainting and drawing. 动漫人物图片自动修复,去马赛克,填补,去瑕疵
Stars: ✭ 761 (-32.24%)
Mutual labels:  opencv, generative-adversarial-network
Getting Things Done With Pytorch
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.
Stars: ✭ 738 (-34.28%)
Mutual labels:  sentiment-analysis, anomaly-detection
Nlp In Practice
Starter code to solve real world text data problems. Includes: Gensim Word2Vec, phrase embeddings, Text Classification with Logistic Regression, word count with pyspark, simple text preprocessing, pre-trained embeddings and more.
Stars: ✭ 790 (-29.65%)
Mutual labels:  natural-language-processing, word2vec
Twitter sentiment analysis word2vec convnet
Twitter Sentiment Analysis with Gensim Word2Vec and Keras Convolutional Network
Stars: ✭ 24 (-97.86%)
Mutual labels:  sentiment-analysis, word2vec
Neurec
Next RecSys Library
Stars: ✭ 731 (-34.91%)
Mutual labels:  autoencoder, variational-autoencoder
Nlp With Ruby
Curated List: Practical Natural Language Processing done in Ruby
Stars: ✭ 907 (-19.23%)
Mutual labels:  natural-language-processing, sentiment-analysis
Servenet
Service Classification based on Service Description
Stars: ✭ 21 (-98.13%)
Mutual labels:  word2vec, glove

Python Codes in Data Science

Codes in NLP, Deep Learning, Reinforcement Learning and Artificial Intelligence

Welcome to my GitHub repo.

I am a Data Scientist and I code in R, Python and Wolfram Mathematica. Here you will find some Machine Learning, Deep Learning, Natural Language Processing and Artificial Intelligence models I developed.

Outputs of the models can be seen at my portfolio: https://drive.google.com/file/d/0B0RLknmL54khdjRQWVBKeTVxSHM/view?usp=sharing


Keras version used in models: keras==1.1.0

Autoencoder for Audio is a model where I compressed an audio file and used Autoencoder to reconstruct the audio file, for use in phoneme classification.

Collaborative Filtering is a Recommender System where the algorithm predicts a movie review based on genre of movie and similarity among people who watched the same movie.

Convolutional NN Lasagne is a Convolutional Neural Network model in Lasagne to solve the MNIST task.

Ensembled Machine Learning is a .py file where 7 Machine Learning algorithms are used in a classification task with 3 classes and all possible hyperparameters of each algorithm are adjusted. Iris dataset of scikit-learn.

GAN Generative Adversarial are models of Generative Adversarial Neural Networks.

Hyperparameter Tuning RL is a model where hyperparameters of Neural Networks are adjusted via Reinforcement Learning. According to a reward, hyperparameter tuning (environment) is changed through a policy (mechanization of knowledge) using the Boston Dataset. Hyperparameters tuned are: learning rate, epochs, decay, momentum, number of hidden layers and nodes and initial weights.

Keras Regularization L2 is a Neural Network model for regression made with Keras where a L2 regularization was applied to prevent overfitting.

Lasagne Neural Nets Regression is a Neural Network model based in Theano and Lasagne, that makes a linear regression with a continuous target variable and reaches 99.4% accuracy. It uses the DadosTeseLogit.csv sample file.

Lasagne Neural Nets + Weights is a Neural Network model based in Theano and Lasagne, where is possible to visualize weights between X1 and X2 to hidden layer. Can also be adapted to visualize weights between hidden layer and output. It uses the DadosTeseLogit.csv sample file.

Multinomial Regression is a regression model where target variable has 3 classes.

Neural Networks for Regression shows multiple solutions for a regression problem, solved with sklearn, Keras, Theano and Lasagne. It uses the Boston dataset sample file from sklearn and reaches more than 98% accuracy.

NLP + Naive Bayes Classifier is a model where movie reviews were labeled as positive and negative and the algorithm then classifies a totally new set of reviews using Logistic Regression, Decision Trees and Naive Bayes, reaching an accuracy of 92%.

NLP Anger Analysis is a Doc2Vec model associated with Word2Vec model to analyze level of anger using synonyms in consumer complaints of a U.S. retailer in Facebook posts.

NLP Consumer Complaint is a model where Facebook posts of a U.S. computer retailer were scraped, tokenized, lemmatized and applied Word2Vec. After that, t-SNE and Latent Dirichlet Allocation were developed in order to classify the arguments and weights of each keyword used by a consumer in his complaint. The code also analyzes frequency of words in 100 posts.

NLP Convolutional Neural Network is a Convolutional Neural Network for Text in order to classify movie reviews.

NLP Doc2Vec is a Natural Language Procesing file where cosine similarity among phrases is measured through Doc2Vec.

NLP Document Classification is a code for Document Classification according to Latent Dirichlet Allocation.

NLP Facebook Analysis analyzes Facebook posts regarding Word Frequency and Topic Modelling using LDA.

NLP Facebook Scrap is a Python code for scraping data from Facebook.

NLP - Latent Dirichlet Allocation is a Natural Language Processing model where a Wikipedia page on Statistical Inference is classified regarding topics, using Latent Dirichlet Allocation with Gensim, NLTK, t-SNE and K-Means.

NLP Probabilistic ANN is a Natural Language Processing model where sentences are vectorized by Gensim and a probabilistic Neural Network model is deveoped using Gensim, for sentiment analysis.

NLP Semantic Doc2Vec + Neural Network is a model where positive and negative movie reviews were extracted and semantically classified with NLTK and BeautifulSoup, then labeled as positive or negative. Text was then used as an input for the Neural Network model training. After training, new sentences are entered in the Keras Neural Network model and then classified. It uses the zip file.

NLP Sentiment Positive is a model that identifies website content as positive, neutral or negative using BeautifulSoup and NLTK libraries, plotting the results.

NLP Twitter Analysis ID # is a model that extracts posts from Twitter based in ID of user or Hashtag.

NLP Twitter Scrap is a model that scraps Twitter data and shows the cleaned text as output.

NLP Twitter Streaming is a model of analysis of real-time data from Twitter (under development).

NLP Twitter Streaming Mood is a model where the evolution of mood Twitter posts is measured during a period of time.

NLP Wikipedia Summarization is a Python code that summarizes any given page in a few sentences.

NLP Word Frequency is a model that calculates the frequency of nouns, verbs, words in Facebook posts.

Probabilistic Neural Network is a Probabilistic Neural Network for Time Series Prediction.

REAL-TIME Twitter Analysis is a model where Twitter streaming is extracted, words and sentences tokenized, word embeddings were created, topic modeling was made and classified using K-Means. Then, NLTK SentimentAnalyzer was used to classify each sentence of the streaming into positive, neutral or negative. Accumulated sum was used to generate the plot and the code loops each 1 second, collecting new tweets.

RESNET-2 is a Deep Residual Neural Network.

ROC Curve Multiclass is a .py file where Naive Bayes was used to solve the IRIS Dataset task and ROC curve of different classes are plotted.

SQUEEZENET is a simplified version of the AlexNet.

Stacked Machine Learning is a .py notebook where t-SNE, Principal Components Analysis and Factor Analysis were applied to reduce dimensionality of data. Classification performances were measured after applying K-Means.

Support Vector Regression is a SVM model for non linear regression in an artificial dataset.

Text-to-Speech is a .py file where Python speaks any given text and saves it as an audio .wav file.

Time Series ARIMA is a ARIMA model to forecast time series, with an error margin of 0.2%.

Time Series Prediction with Neural Networks - Keras is a Neural Network model to forecast time series, using Keras with an adaptive learning rate depending upon derivative of loss.

Variational Autoencoder is a VAE made with Keras.

Web Crawler is a code that scraps data from different URLs of a hotel website.

t-SNE Dimensionality Reduction is a t-SNE model for dimensionality reduction which is compared to Principal Components Analysis regarding its discriminatory power.

t-SNE PCA + Neural Networks is a model that compares performance or Neural Networks made after t-SNE, PCA and K-Means.

t-SNE PCA LDA embeddings is a model where t-SNE, Principal Components Analysis, Linear Discriminant Analysis and Random Forest embeddings are compared in a task to classify clusters of similar digits.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].