All Projects → p208p2002 → taipei-QA-BERT

p208p2002 / taipei-QA-BERT

Licence: other
台北QA問答機器人(使用BERT、ALBERT)

Programming Languages

python
139335 projects - #7 most used programming language
Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to taipei-QA-BERT

gender-unbiased BERT-based pronoun resolution
Source code for the ACL workshop paper and Kaggle competition by Google AI team
Stars: ✭ 42 (+10.53%)
Mutual labels:  bert-model
text-classification-transformers
Easy text classification for everyone : Bert based models via Huggingface transformers (KR / EN)
Stars: ✭ 32 (-15.79%)
Mutual labels:  bert-model
german-sentiment
A data set and model for german sentiment classification.
Stars: ✭ 37 (-2.63%)
Mutual labels:  bert-model
RECCON
This repository contains the dataset and the PyTorch implementations of the models from the paper Recognizing Emotion Cause in Conversations.
Stars: ✭ 126 (+231.58%)
Mutual labels:  bert-model
german-sentiment-lib
An easy to use python package for deep learning-based german sentiment classification.
Stars: ✭ 33 (-13.16%)
Mutual labels:  bert-model
Distill-BERT-Textgen
Research code for ACL 2020 paper: "Distilling Knowledge Learned in BERT for Text Generation".
Stars: ✭ 121 (+218.42%)
Mutual labels:  bert-model
R-BERT
Pytorch re-implementation of R-BERT model
Stars: ✭ 59 (+55.26%)
Mutual labels:  bert-model
Transformer Temporal Tagger
Code and data form the paper BERT Got a Date: Introducing Transformers to Temporal Tagging
Stars: ✭ 55 (+44.74%)
Mutual labels:  bert-model
Simple-HOHMM
High Order Hidden Markov Model for accurate sequence classification.
Stars: ✭ 21 (-44.74%)
Mutual labels:  sequence-classification
FinBERT-QA
Financial Domain Question Answering with pre-trained BERT Language Model
Stars: ✭ 70 (+84.21%)
Mutual labels:  bert-model
AiSpace
AiSpace: Better practices for deep learning model development and deployment For Tensorflow 2.0
Stars: ✭ 28 (-26.32%)
Mutual labels:  albert-chinese
Kashgari
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+5781.58%)
Mutual labels:  bert-model
SIGIR2021 Conure
One Person, One Model, One World: Learning Continual User Representation without Forgetting
Stars: ✭ 23 (-39.47%)
Mutual labels:  bert-model
PIE
Fast + Non-Autoregressive Grammatical Error Correction using BERT. Code and Pre-trained models for paper "Parallel Iterative Edit Models for Local Sequence Transduction": www.aclweb.org/anthology/D19-1435.pdf (EMNLP-IJCNLP 2019)
Stars: ✭ 164 (+331.58%)
Mutual labels:  bert-model
use-cases-of-bert
Use-cases of Hugging Face's BERT (e.g. paraphrase generation, unsupervised extractive summarization).
Stars: ✭ 18 (-52.63%)
Mutual labels:  bert-model
MSR2021-ProgramRepair
Code of our paper Applying CodeBERT for Automated Program Repair of Java Simple Bugs which is accepted to MSR 2021.
Stars: ✭ 26 (-31.58%)
Mutual labels:  bert-model
siamese-BERT-fake-news-detection-LIAR
Triple Branch BERT Siamese Network for fake news classification on LIAR-PLUS dataset in PyTorch
Stars: ✭ 96 (+152.63%)
Mutual labels:  bert-model
bert-movie-reviews-sentiment-classifier
Build a Movie Reviews Sentiment Classifier with Google's BERT Language Model
Stars: ✭ 12 (-68.42%)
Mutual labels:  bert-model
bert-tensorflow-pytorch-spacy-conversion
Instructions for how to convert a BERT Tensorflow model to work with HuggingFace's pytorch-transformers, and spaCy. This walk-through uses DeepPavlov's RuBERT as example.
Stars: ✭ 26 (-31.58%)
Mutual labels:  bert-model

台北QA問答機器人(with BERT or ALBERT)

問一個問題,告訴你應該去哪個單位處理這些問題

檔案說明

  • train.py : 模型訓練(BERT fine-tune)
  • predict.py : 提問預測
  • tutorial : 投影片檔案

中文Albert

  • 預設使用bert-base-chinese
  • 欲切換至albert-zh-tiny請將train.pypredict.py對應註解拿掉

    使用albert-zh-tiny需要對訓練參數進行微調,否則無法獲得好的效果

  • 更多albert-zh模型與用法請參閱p208p2002/albert-zh-for-pytorch-transformers

環境需求

  • python 3.6+
  • pytorch 1.3+
Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].