All Projects → albertlauncher → Albert

albertlauncher / Albert

A fast and flexible keyboard launcher

Programming Languages

C++
36643 projects - #6 most used programming language
python
139335 projects - #7 most used programming language
CMake
9771 projects
c
50402 projects - #5 most used programming language

Projects that are alternatives of or similar to Albert

Albert zh
A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS, 海量中文预训练ALBERT模型
Stars: ✭ 3,500 (-41.84%)
Mutual labels:  albert
bert in a flask
A dockerized flask API, serving ALBERT and BERT predictions using TensorFlow 2.0.
Stars: ✭ 32 (-99.47%)
Mutual labels:  albert
CLUE pytorch
CLUE baseline pytorch CLUE的pytorch版本基线
Stars: ✭ 72 (-98.8%)
Mutual labels:  albert
keras-bert-ner
Keras solution of Chinese NER task using BiLSTM-CRF/BiGRU-CRF/IDCNN-CRF model with Pretrained Language Model: supporting BERT/RoBERTa/ALBERT
Stars: ✭ 7 (-99.88%)
Mutual labels:  albert
MobileQA
离线端阅读理解应用 QA for mobile, Android & iPhone
Stars: ✭ 49 (-99.19%)
Mutual labels:  albert
ALBERT-Pytorch
Pytorch Implementation of ALBERT(A Lite BERT for Self-supervised Learning of Language Representations)
Stars: ✭ 214 (-96.44%)
Mutual labels:  albert
Medi-CoQA
Conversational Question Answering on Clinical Text
Stars: ✭ 22 (-99.63%)
Mutual labels:  albert
tfbert
基于tensorflow1.x的预训练模型调用,支持单机多卡、梯度累积,XLA加速,混合精度。可灵活训练、验证、预测。
Stars: ✭ 54 (-99.1%)
Mutual labels:  albert
classifier multi label
multi-label,classifier,text classification,多标签文本分类,文本分类,BERT,ALBERT,multi-label-classification
Stars: ✭ 127 (-97.89%)
Mutual labels:  albert
classifier multi label seq2seq attention
multi-label,classifier,text classification,多标签文本分类,文本分类,BERT,ALBERT,multi-label-classification,seq2seq,attention,beam search
Stars: ✭ 26 (-99.57%)
Mutual labels:  albert
bert nli
A Natural Language Inference (NLI) model based on Transformers (BERT and ALBERT)
Stars: ✭ 97 (-98.39%)
Mutual labels:  albert
NLP-paper
🎨 🎨NLP 自然语言处理教程 🎨🎨 https://dataxujing.github.io/NLP-paper/
Stars: ✭ 23 (-99.62%)
Mutual labels:  albert
Transformer-QG-on-SQuAD
Implement Question Generator with SOTA pre-trained Language Models (RoBERTa, BERT, GPT, BART, T5, etc.)
Stars: ✭ 28 (-99.53%)
Mutual labels:  albert
ClusterTransformer
Topic clustering library built on Transformer embeddings and cosine similarity metrics.Compatible with all BERT base transformers from huggingface.
Stars: ✭ 36 (-99.4%)
Mutual labels:  albert
KB-ALBERT
KB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델
Stars: ✭ 215 (-96.43%)
Mutual labels:  albert
SupermarktConnector
Collecting product information from Dutch supermarkets: Albert Heijn and Jumbo using the Mobile API
Stars: ✭ 91 (-98.49%)
Mutual labels:  albert
Spark Nlp
State of the Art Natural Language Processing
Stars: ✭ 2,518 (-58.16%)
Mutual labels:  albert
Clue
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Stars: ✭ 2,425 (-59.7%)
Mutual labels:  albert
Chineseglue
Language Understanding Evaluation benchmark for Chinese: datasets, baselines, pre-trained models,corpus and leaderboard
Stars: ✭ 1,548 (-74.28%)
Mutual labels:  albert

Albert launcher Build Status

Telegram community chat Discord IRC IRC Web Client

Albert is a unified and efficient access to your machine. Technically it is a keyboard launcher written in C++/Qt. The plugin based architecture makes it extremely flexible and powerful. Plugins can be used to create frontends or functional extensions. Extensions can also be implemented using embedded Python modules or via a CGI approach in any language you want.

albert

Getting started

Check the website for nice pics and vidz. There you will also find instructions for installing, using, extending or troubleshooting Albert.

Support

If you like this software consider donating a 🍺. 👍

Have fun with albert, if you do not, tell me why.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].