All Projects → nlx-group → Wordnetembeddings

nlx-group / Wordnetembeddings

Licence: mit
Obtaining word embeddings from a WordNet ontology

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Wordnetembeddings

sembei
🍘 単語分割を経由しない単語埋め込み 🍘
Stars: ✭ 14 (-57.58%)
Mutual labels:  word-embeddings
Biosentvec
BioWordVec & BioSentVec: pre-trained embeddings for biomedical words and sentences
Stars: ✭ 308 (+833.33%)
Mutual labels:  word-embeddings
Meta
A Modern C++ Data Sciences Toolkit
Stars: ✭ 600 (+1718.18%)
Mutual labels:  word-embeddings
Text-Analysis
Explaining textual analysis tools in Python. Including Preprocessing, Skip Gram (word2vec), and Topic Modelling.
Stars: ✭ 48 (+45.45%)
Mutual labels:  word-embeddings
Hands On Deep Learning Algorithms With Python
Master Deep Learning Algorithms with Extensive Math by Implementing them using TensorFlow
Stars: ✭ 272 (+724.24%)
Mutual labels:  word-embeddings
Wego
Word Embeddings (e.g. Word2Vec) in Go!
Stars: ✭ 336 (+918.18%)
Mutual labels:  word-embeddings
SWDM
SIGIR 2017: Embedding-based query expansion for weighted sequential dependence retrieval model
Stars: ✭ 35 (+6.06%)
Mutual labels:  word-embeddings
Concise Ipython Notebooks For Deep Learning
Ipython Notebooks for solving problems like classification, segmentation, generation using latest Deep learning algorithms on different publicly available text and image data-sets.
Stars: ✭ 23 (-30.3%)
Mutual labels:  word-embeddings
Named Entity Recognition With Bidirectional Lstm Cnns
Named-Entity-Recognition-with-Bidirectional-LSTM-CNNs
Stars: ✭ 283 (+757.58%)
Mutual labels:  word-embeddings
Nlp Notebooks
A collection of notebooks for Natural Language Processing from NLP Town
Stars: ✭ 513 (+1454.55%)
Mutual labels:  word-embeddings
NLP-Stuff
Programs with word vectors, RNN, NLP stuff, etc
Stars: ✭ 19 (-42.42%)
Mutual labels:  word-embeddings
neuralnets-semantics
Word semantics Deep Learning with Vanilla Python, Keras, Theano, TensorFlow, PyTorch
Stars: ✭ 15 (-54.55%)
Mutual labels:  word-embeddings
Deep learning nlp
Keras, PyTorch, and NumPy Implementations of Deep Learning Architectures for NLP
Stars: ✭ 407 (+1133.33%)
Mutual labels:  word-embeddings
SPINE
Code for SPINE - Sparse Interpretable Neural Embeddings. Jhamtani H.*, Pruthi D.*, Subramanian A.*, Berg-Kirkpatrick T., Hovy E. AAAI 2018
Stars: ✭ 44 (+33.33%)
Mutual labels:  word-embeddings
Inltk
Natural Language Toolkit for Indic Languages aims to provide out of the box support for various NLP tasks that an application developer might need
Stars: ✭ 702 (+2027.27%)
Mutual labels:  word-embeddings
yelp comments classification nlp
Yelp round-10 review comments classification using deep learning (LSTM and CNN) and natural language processing.
Stars: ✭ 72 (+118.18%)
Mutual labels:  word-embeddings
Chakin
Simple downloader for pre-trained word vectors
Stars: ✭ 323 (+878.79%)
Mutual labels:  word-embeddings
Syntree2vec
An algorithm to augment syntactic hierarchy into word embeddings
Stars: ✭ 9 (-72.73%)
Mutual labels:  word-embeddings
Text2vec
Fast vectorization, topic modeling, distances and GloVe word embeddings in R.
Stars: ✭ 715 (+2066.67%)
Mutual labels:  word-embeddings
Bert Embedding
🔡 Token level embeddings from BERT model on mxnet and gluonnlp
Stars: ✭ 424 (+1184.85%)
Mutual labels:  word-embeddings

WordNet Embeddings

wnet2vec

Article

Saedi, Chakaveh, António Branco, João António Rodrigues and João Ricardo Silva, 2018, "WordNet Embeddings", In Proceedings, 3rd Workshop on Representation Learning for Natural Language Processing (RepL4NLP), 56th Annual Meeting of the Association for Computational Linguistics, 15-20 July 2018, Melbourne, Australia.

WordNet used in the above paper

Princeton WordNet 3.0

Test sets used in above paper

Please note that the semantic network to semantic space method presented in the above paper includes random-based subprocedures (e.g. selecting one word from a set of words with identical number of outgoing edges). The test scores may present slight fluctuations over different runs of the code.

SimLex-999

RG1965

WordSim-353-Similarity

WordSim-353-Relatedness

MEN

MTurk-771

Models

The best wnet2vec model we have obtained that was ran with 60,000 words using Princeton WordNet 3.0, referred in the article, is available for download here.

How to run wn2vec software

To provide input files to the software the following structure must exist:

|-- main.py
|-- data
|   |-- input
|   |   |-- language_wnet
|   |   |   |-- *wnet_files
|   |   |-- language_testset
|   |   |   |-- *testset_files
|   |-- output
|-- modules
|   |-- input_output.py
|   |-- sort_rank_remove.py
|   |-- vector_accuracy_checker.py
|   |-- vector_distance.py
|   |-- vector_generator.py

Where language is the language that you are using that must be indicated in main.py in the variable lang. If the language isn't supported by the current path routing in the code, which was mainly use for experiments, you may add the path to the directory in the files input_output.py, vector_generator.py and vector_accuracy_checker.py.

Various variables for the output of the model, such as embedding dimension, can be found in main.py.

To run the software, you will need the following packages:

  • Numpy
  • progressbar
  • keras
  • sklearn
  • scipy
  • gensim

Python3.5 was used for the experimentation.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].