Mead BaselineDeep-Learning Model Exploration and Development for NLP
Stars: ✭ 238 (-37.2%)
NLP-paper🎨 🎨NLP 自然语言处理教程 🎨🎨 https://dataxujing.github.io/NLP-paper/
Stars: ✭ 23 (-93.93%)
Rcnn Text ClassificationTensorflow Implementation of "Recurrent Convolutional Neural Network for Text Classification" (AAAI 2015)
Stars: ✭ 127 (-66.49%)
Pytorch Seq2seqTutorials on implementing a few sequence-to-sequence (seq2seq) models with PyTorch and TorchText.
Stars: ✭ 3,418 (+801.85%)
Text classificationall kinds of text classification models and more with deep learning
Stars: ✭ 7,179 (+1794.2%)
RmdlRMDL: Random Multimodel Deep Learning for Classification
Stars: ✭ 375 (-1.06%)
EmbeddingEmbedding模型代码和学习笔记总结
Stars: ✭ 25 (-93.4%)
Keras Textclassification中文长文本分类、短句子分类、多标签分类、两句子相似度(Chinese Text Classification of Keras NLP, multi-label classify, or sentence classify, long or short),字词句向量嵌入层(embeddings)和网络层(graph)构建基类,FastText,TextCNN,CharCNN,TextRNN, RCNN, DCNN, DPCNN, VDCNN, CRNN, Bert, Xlnet, Albert, Attention, DeepMoji, HAN, 胶囊网络-CapsuleNet, Transformer-encode, Seq2seq, SWEM, LEAM, TextGCN
Stars: ✭ 914 (+141.16%)
Nlp TutorialsSimple implementations of NLP models. Tutorials are written in Chinese on my website https://mofanpy.com
Stars: ✭ 394 (+3.96%)
Malware ClassificationTowards Building an Intelligent Anti-Malware System: A Deep Learning Approach using Support Vector Machine for Malware Classification
Stars: ✭ 88 (-76.78%)
Image Caption GeneratorA neural network to generate captions for an image using CNN and RNN with BEAM Search.
Stars: ✭ 126 (-66.75%)
Rcnn Relation ExtractionTensorflow Implementation of Recurrent Convolutional Neural Network for Relation Extraction
Stars: ✭ 64 (-83.11%)
transformerA PyTorch Implementation of "Attention Is All You Need"
Stars: ✭ 28 (-92.61%)
Nlp researchNLP research:基于tensorflow的nlp深度学习项目,支持文本分类/句子匹配/序列标注/文本生成 四大任务
Stars: ✭ 141 (-62.8%)
Neural spEnd-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (+7.65%)
Nlp JourneyDocuments, papers and codes related to Natural Language Processing, including Topic Model, Word Embedding, Named Entity Recognition, Text Classificatin, Text Generation, Text Similarity, Machine Translation),etc. All codes are implemented intensorflow 2.0.
Stars: ✭ 1,290 (+240.37%)
KomputationKomputation is a neural network framework for the Java Virtual Machine written in Kotlin and CUDA C.
Stars: ✭ 295 (-22.16%)
Ner BertBERT-NER (nert-bert) with google bert https://github.com/google-research.
Stars: ✭ 339 (-10.55%)
seq2seq-pytorchSequence to Sequence Models in PyTorch
Stars: ✭ 41 (-89.18%)
TRAR-VQA[ICCV 2021] TRAR: Routing the Attention Spans in Transformers for Visual Question Answering -- Official Implementation
Stars: ✭ 49 (-87.07%)
tensorflow-chatbot-chinese網頁聊天機器人 | tensorflow implementation of seq2seq model with bahdanau attention and Word2Vec pretrained embedding
Stars: ✭ 50 (-86.81%)
german-sentimentA data set and model for german sentiment classification.
Stars: ✭ 37 (-90.24%)
Transfer Learning SuiteTransfer Learning Suite in Keras. Perform transfer learning using any built-in Keras image classification model easily!
Stars: ✭ 212 (-44.06%)
chatbot一个基于深度学习的中文聊天机器人,这里有详细的教程与代码,每份代码都有详细的注释,作为学习是美好的选择。A Chinese chatbot based on deep learning.
Stars: ✭ 94 (-75.2%)
Seq2Seq-chatbotTensorFlow Implementation of Twitter Chatbot
Stars: ✭ 18 (-95.25%)
Mobilenet V2A Complete and Simple Implementation of MobileNet-V2 in PyTorch
Stars: ✭ 206 (-45.65%)
kospeechOpen-Source Toolkit for End-to-End Korean Automatic Speech Recognition leveraging PyTorch and Hydra.
Stars: ✭ 456 (+20.32%)
Transformer Temporal TaggerCode and data form the paper BERT Got a Date: Introducing Transformers to Temporal Tagging
Stars: ✭ 55 (-85.49%)
CrabNetPredict materials properties using only the composition information!
Stars: ✭ 57 (-84.96%)
ConformerOfficial code for Conformer: Local Features Coupling Global Representations for Visual Recognition
Stars: ✭ 345 (-8.97%)
datastories-semeval2017-task6Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (-94.72%)
Cnn SvmAn Architecture Combining Convolutional Neural Network (CNN) and Linear Support Vector Machine (SVM) for Image Classification
Stars: ✭ 170 (-55.15%)
h-transformer-1dImplementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning
Stars: ✭ 121 (-68.07%)
verseagilityRamp up your custom natural language processing (NLP) task, allowing you to bring your own data, use your preferred frameworks and bring models into production.
Stars: ✭ 23 (-93.93%)
HRFormerThis is an official implementation of our NeurIPS 2021 paper "HRFormer: High-Resolution Transformer for Dense Prediction".
Stars: ✭ 357 (-5.8%)
RNNSearchAn implementation of attention-based neural machine translation using Pytorch
Stars: ✭ 43 (-88.65%)
dl-reluDeep Learning using Rectified Linear Units (ReLU)
Stars: ✭ 20 (-94.72%)
transformerNeutron: A pytorch based implementation of Transformer and its variants.
Stars: ✭ 60 (-84.17%)
Transformer TensorflowTensorFlow implementation of 'Attention Is All You Need (2017. 6)'
Stars: ✭ 319 (-15.83%)
deep-molecular-optimizationMolecular optimization by capturing chemist’s intuition using the Seq2Seq with attention and the Transformer
Stars: ✭ 60 (-84.17%)
visualizationa collection of visualization function
Stars: ✭ 189 (-50.13%)
dtsA Keras library for multi-step time-series forecasting.
Stars: ✭ 130 (-65.7%)
Encoder decoderFour styles of encoder decoder model by Python, Theano, Keras and Seq2Seq
Stars: ✭ 269 (-29.02%)