All Projects → ma-ji → npo_classifier

ma-ji / npo_classifier

Licence: other
Automated coding using machine-learning and remapping the U.S. nonprofit sector: A guide and benchmark

Programming Languages

Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to npo classifier

Awesome Bert
bert nlp papers, applications and github resources, including the newst xlnet , BERT、XLNet 相关论文和 github 项目
Stars: ✭ 1,732 (+9522.22%)
Mutual labels:  bert
Texar
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project: http://casl-project.ai/
Stars: ✭ 2,236 (+12322.22%)
Mutual labels:  bert
Cool-NLPCV
Some Cool NLP and CV Repositories and Solutions (收集NLP中常见任务的开源解决方案、数据集、工具、学习资料等)
Stars: ✭ 143 (+694.44%)
Mutual labels:  bert
Clue
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Stars: ✭ 2,425 (+13372.22%)
Mutual labels:  bert
Kashgari
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Stars: ✭ 2,235 (+12316.67%)
Mutual labels:  bert
Pytorch Sentiment Analysis
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
Stars: ✭ 3,209 (+17727.78%)
Mutual labels:  bert
Fast Bert
Super easy library for BERT based NLP models
Stars: ✭ 1,678 (+9222.22%)
Mutual labels:  bert
AiSpace
AiSpace: Better practices for deep learning model development and deployment For Tensorflow 2.0
Stars: ✭ 28 (+55.56%)
Mutual labels:  bert
Keras Bert
Implementation of BERT that could load official pre-trained models for feature extraction and prediction
Stars: ✭ 2,264 (+12477.78%)
Mutual labels:  bert
protonet-bert-text-classification
finetune bert for small dataset text classification in a few-shot learning manner using ProtoNet
Stars: ✭ 28 (+55.56%)
Mutual labels:  bert
Awesome Sentence Embedding
A curated list of pretrained sentence and word embedding models
Stars: ✭ 1,973 (+10861.11%)
Mutual labels:  bert
Pycorrector
pycorrector is a toolkit for text error correction. 文本纠错,Kenlm,Seq2Seq_Attention,BERT,MacBERT,ELECTRA,ERNIE,Transformer等模型实现,开箱即用。
Stars: ✭ 2,857 (+15772.22%)
Mutual labels:  bert
Nlp Architect
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
Stars: ✭ 2,768 (+15277.78%)
Mutual labels:  bert
Mt Dnn
Multi-Task Deep Neural Networks for Natural Language Understanding
Stars: ✭ 1,871 (+10294.44%)
Mutual labels:  bert
GLUE-bert4keras
基于bert4keras的GLUE基准代码
Stars: ✭ 59 (+227.78%)
Mutual labels:  bert
Roberta zh
RoBERTa中文预训练模型: RoBERTa for Chinese
Stars: ✭ 1,953 (+10750%)
Mutual labels:  bert
Bertviz
Tool for visualizing attention in the Transformer model (BERT, GPT-2, Albert, XLNet, RoBERTa, CTRL, etc.)
Stars: ✭ 3,443 (+19027.78%)
Mutual labels:  bert
Bert-model-code-interpretation
解读tensorflow版本Bert中modeling.py数据流
Stars: ✭ 19 (+5.56%)
Mutual labels:  bert
spacy-sentence-bert
Sentence transformers models for SpaCy
Stars: ✭ 88 (+388.89%)
Mutual labels:  bert
Transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Stars: ✭ 55,742 (+309577.78%)
Mutual labels:  bert

Automated coding using machine-learning and remapping the U.S. nonprofit sector: A guide and benchmark

NVSQ DOI OPEN ACCESS PAPER

This research developed a machine-learning classifier that reliably automates the coding process using the National Taxonomy of Exempt Entities as a schema and remapped the U.S. nonprofit sector. I achieved 90% overall accuracy for classifying the nonprofits into nine broad categories and 88% for classifying them into 25 major groups. The intercoder reliabilities between algorithms and human coders measured by kappa statistics are in the "almost perfect" range of 0.80--1.00. The results suggest that a state-of-the-art machine-learning algorithm can approximate human coders and substantially improve researchers' productivity. I also reassigned multiple category codes to over 439 thousand nonprofits and discovered a considerable amount of organizational activities that were previously ignored. The classifier is an essential methodological prerequisite for large-N and Big Data analyses, and the remapped U.S. nonprofit sector can serve as an important instrument for asking or reexamining fundamental questions of nonprofit studies.

Useful resources

Folder structure

.
├── API
├── dataset
│   ├── UCF
│   │   ├── test
│   │   └── train
│   ├── intermediary
│   └── muy060_suppl_supplementary_appendix
├── output
│   ├── classification_results
│   ├── fig
│   └── result_dicts
├── reference
│   ├── algorithms
│   └── assign_NTEE
└── script
    ├── classification_algorithms
    ├── data_acquisition
    └── data_analysis

How to cite

Ma, J. (2021). Automated Coding Using Machine Learning and Remapping the U.S. Nonprofit Sector: A Guide and Benchmark. Nonprofit and Voluntary Sector Quarterly, 50(3), 662–687. https://doi.org/10.1177/0899764020968153

@article{MaAutomatedCodingUsing2021,
	title = {Automated {Coding} {Using} {Machine} {Learning} and {Remapping} the {U}.{S}. {Nonprofit} {Sector}: {A} {Guide} and {Benchmark}},
	volume = {50},
	issn = {0899-7640},
	shorttitle = {Automated {Coding} {Using} {Machine} {Learning} and {Remapping} the {U}.{S}. {Nonprofit} {Sector}},
	url = {https://doi.org/10.1177/0899764020968153},
	doi = {10.1177/0899764020968153},
	abstract = {This research developed a machine learning classifier that reliably automates the coding process using the National Taxonomy of Exempt Entities as a schema and remapped the U.S. nonprofit sector. I achieved 90\% overall accuracy for classifying the nonprofits into nine broad categories and 88\% for classifying them into 25 major groups. The intercoder reliabilities between algorithms and human coders measured by kappa statistics are in the “almost perfect” range of .80 to 1.00. The results suggest that a state-of-the-art machine learning algorithm can approximate human coders and substantially improve researchers’ productivity. I also reassigned multiple category codes to more than 439,000 nonprofits and discovered a considerable amount of organizational activities that were previously ignored. The classifier is an essential methodological prerequisite for large-N and Big Data analyses, and the remapped U.S. nonprofit sector can serve as an important instrument for asking or reexamining fundamental questions of nonprofit studies. The working directory with all data sets, source codes, and historical versions are available on GitHub (https://github.com/ma-ji/npo\_classifier).},
	language = {en},
	number = {3},
	urldate = {2021-05-22},
	journal = {Nonprofit and Voluntary Sector Quarterly},
	author = {Ma, Ji},
	month = jun,
	year = {2021},
	note = {Publisher: SAGE Publications Inc},
	keywords = {BERT, computational social science, machine learning, National Taxonomy of Exempt Entities, neural network, nonprofit organization},
	pages = {662--687}
}

Funding

This project was supported in part by the 2019-20 PRI Award and Stephen H. Spurr Centennial Fellowship from the LBJ School of Public Affairs and a Planet Texas 2050 grant from UT Austin.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].