All Projects → cahya-wirawan → Indonesian Language Models

cahya-wirawan / Indonesian Language Models

Licence: mit
Indonesian Language Models and its Usage

Projects that are alternatives of or similar to Indonesian Language Models

Vietnamese Electra
Electra pre-trained model using Vietnamese corpus
Stars: ✭ 55 (-14.06%)
Mutual labels:  jupyter-notebook, language-model, transformer
Gpt2 French
GPT-2 French demo | Démo française de GPT-2
Stars: ✭ 47 (-26.56%)
Mutual labels:  jupyter-notebook, language-model, transformer
Neural sp
End-to-end ASR/LM implementation with PyTorch
Stars: ✭ 408 (+537.5%)
Mutual labels:  language-model, transformer
Tsai
Time series Timeseries Deep Learning Pytorch fastai - State-of-the-art Deep Learning with Time Series and Sequences in Pytorch / fastai
Stars: ✭ 407 (+535.94%)
Mutual labels:  jupyter-notebook, transformer
Nlp Paper
NLP Paper
Stars: ✭ 484 (+656.25%)
Mutual labels:  language-model, transformer
Bert Multitask Learning
BERT for Multitask Learning
Stars: ✭ 380 (+493.75%)
Mutual labels:  jupyter-notebook, transformer
Deepsvg
[NeurIPS 2020] Official code for the paper "DeepSVG: A Hierarchical Generative Network for Vector Graphics Animation". Includes a PyTorch library for deep learning with SVG data.
Stars: ✭ 403 (+529.69%)
Mutual labels:  jupyter-notebook, transformer
Bert Pytorch
Google AI 2018 BERT pytorch implementation
Stars: ✭ 4,642 (+7153.13%)
Mutual labels:  language-model, transformer
Demo Chinese Text Binary Classification With Bert
Stars: ✭ 276 (+331.25%)
Mutual labels:  jupyter-notebook, transformer
Kobert
Korean BERT pre-trained cased (KoBERT)
Stars: ✭ 591 (+823.44%)
Mutual labels:  jupyter-notebook, language-model
Awesome Bert Nlp
A curated list of NLP resources focused on BERT, attention mechanism, Transformer networks, and transfer learning.
Stars: ✭ 567 (+785.94%)
Mutual labels:  language-model, transformer
Getting Things Done With Pytorch
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.
Stars: ✭ 738 (+1053.13%)
Mutual labels:  jupyter-notebook, transformer
Question generation
Neural question generation using transformers
Stars: ✭ 356 (+456.25%)
Mutual labels:  jupyter-notebook, transformer
Azureml Bert
End-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
Stars: ✭ 342 (+434.38%)
Mutual labels:  jupyter-notebook, language-model
Deeplearning Nlp Models
A small, interpretable codebase containing the re-implementation of a few "deep" NLP models in PyTorch. Colab notebooks to run with GPUs. Models: word2vec, CNNs, transformer, gpt.
Stars: ✭ 64 (+0%)
Mutual labels:  jupyter-notebook, transformer
Dab
Data Augmentation by Backtranslation (DAB) ヽ( •_-)ᕗ
Stars: ✭ 294 (+359.38%)
Mutual labels:  jupyter-notebook, transformer
Pytorch Original Transformer
My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing otherwise seemingly hard concepts. Currently included IWSLT pretrained models.
Stars: ✭ 411 (+542.19%)
Mutual labels:  jupyter-notebook, transformer
MinTL
MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
Stars: ✭ 61 (-4.69%)
Mutual labels:  transformer, language-model
Transformer
Implementation of Transformer model (originally from Attention is All You Need) applied to Time Series.
Stars: ✭ 273 (+326.56%)
Mutual labels:  jupyter-notebook, transformer
Sentiment analysis fine grain
Multi-label Classification with BERT; Fine Grained Sentiment Analysis from AI challenger
Stars: ✭ 546 (+753.13%)
Mutual labels:  jupyter-notebook, language-model

Indonesian Language Models

The language model is a probability distribution over word sequences used to predict the next word based on previous sentences. This ability makes the language model the core component of modern natural language processing. We use it for many different tasks, such as speech recognition, conversational AI, information retrieval, sentiment analysis, or text summarization.

For this reason, many big companies are competing to build large and larger language models, such as Google BERT, Facebook RoBERTa, or OpenAI GPT3, with its massive number of parameters. Most of the time, they built only language models in English and some other European languages. Other countries with low resource languages have big challenges to catch up on this technology race.

Therefore the author tries to build some language models for Indonesian, started with ULMFiT in 2018. The first language model has been only trained with Indonesian Wikipedia, which is very small compared to other datasets used to train the English language model.

Universal Language Model Fine-tuning (ULMFiT)

Jeremy Howard and Sebastian Ruder proposed ULMFiT in early 2018 as a novel method for fine-tuning language models for inductive transfer learning. The language model ULMFiT for Indonesian has been trained as part of the author's project while learning FastAI. It achieved a perplexity of 27.67 on Indonesian Wikipedia.

Transformers

Ashish Vaswani et al. proposed Transfomer in the paper Attention Is All You Need. It is a novel architecture that aims to solve sequence-to-sequence tasks while handling long-range dependencies with ease.

At the time of writing (March 2021), there are already more than 50 different types of transformer-based language models (according to the model list at huggingface), such as BERT, GPT2, Longformer, or MT5, built by companies and individual contributors. The author built also several Indonesian transformer-based language models using Huggingface Transformers Library and hosted them in the Huggingfaces model hub.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].