All Projects → KazutoshiShinoda → SiameseCBOW

KazutoshiShinoda / SiameseCBOW

Licence: other
Implementation of Siamese CBOW using keras whose backend is tensorflow.

Programming Languages

python
139335 projects - #7 most used programming language
Jupyter Notebook
11667 projects

Projects that are alternatives of or similar to SiameseCBOW

Word2VecfJava
Word2VecfJava: Java implementation of Dependency-Based Word Embeddings and extensions
Stars: ✭ 14 (+0%)
Mutual labels:  word-embeddings
contextualLSTM
Contextual LSTM for NLP tasks like word prediction and word embedding creation for Deep Learning
Stars: ✭ 28 (+100%)
Mutual labels:  word-embeddings
Active-Explainable-Classification
A set of tools for leveraging pre-trained embeddings, active learning and model explainability for effecient document classification
Stars: ✭ 28 (+100%)
Mutual labels:  word-embeddings
wefe
WEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!
Stars: ✭ 164 (+1071.43%)
Mutual labels:  word-embeddings
word2vec-on-wikipedia
A pipeline for training word embeddings using word2vec on wikipedia corpus.
Stars: ✭ 68 (+385.71%)
Mutual labels:  word-embeddings
Arabic-Word-Embeddings-Word2vec
Arabic Word Embeddings Word2vec
Stars: ✭ 26 (+85.71%)
Mutual labels:  word-embeddings
overview-and-benchmark-of-traditional-and-deep-learning-models-in-text-classification
NLP tutorial
Stars: ✭ 41 (+192.86%)
Mutual labels:  word-embeddings
datastories-semeval2017-task6
Deep-learning model presented in "DataStories at SemEval-2017 Task 6: Siamese LSTM with Attention for Humorous Text Comparison".
Stars: ✭ 20 (+42.86%)
Mutual labels:  word-embeddings
dasem
Danish Semantic analysis
Stars: ✭ 17 (+21.43%)
Mutual labels:  word-embeddings
compress-fasttext
Tools for shrinking fastText models (in gensim format)
Stars: ✭ 124 (+785.71%)
Mutual labels:  word-embeddings
fuzzymax
Code for the paper: Don't Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors, ICLR 2019.
Stars: ✭ 43 (+207.14%)
Mutual labels:  word-embeddings
S-WMD
Code for Supervised Word Mover's Distance (SWMD)
Stars: ✭ 90 (+542.86%)
Mutual labels:  word-embeddings
pair2vec
pair2vec: Compositional Word-Pair Embeddings for Cross-Sentence Inference
Stars: ✭ 62 (+342.86%)
Mutual labels:  word-embeddings
sister
SImple SenTence EmbeddeR
Stars: ✭ 66 (+371.43%)
Mutual labels:  word-embeddings
JoSH
[KDD 2020] Hierarchical Topic Mining via Joint Spherical Tree and Text Embedding
Stars: ✭ 55 (+292.86%)
Mutual labels:  word-embeddings
two-stream-cnn
A two-stream convolutional neural network for learning abitrary similarity functions over two sets of training data
Stars: ✭ 24 (+71.43%)
Mutual labels:  word-embeddings
word-benchmarks
Benchmarks for intrinsic word embeddings evaluation.
Stars: ✭ 45 (+221.43%)
Mutual labels:  word-embeddings
SIFRank
The code of our paper "SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model"
Stars: ✭ 96 (+585.71%)
Mutual labels:  word-embeddings
robot-mind-meld
A little game powered by word vectors
Stars: ✭ 31 (+121.43%)
Mutual labels:  word-embeddings
MorphologicalPriorsForWordEmbeddings
Code for EMNLP 2016 paper: Morphological Priors for Probabilistic Word Embeddings
Stars: ✭ 53 (+278.57%)
Mutual labels:  word-embeddings

SiameseCBOW

Implementation of SiameseCBOW using python3.4, keras and tensorflow.

Environments

How to use

0. Installation

$ pip install -r requirements.txt

1. Customize src/load.py

At first, you should customize src/load.py so that you can load your data and preprocess it.

If you just want to test the codes, please comment out x, y = load(file) and set variable:x, y in main.py like this:

def main():

...

 import numpy as np
 x = [np.ones((10, input_length))] * (1 + n_positive + n_negative)
 y = np.ones((10, n_positive + n_negative))

...

 model.fit(x, y, epochs=1)

2. Set Hyper-parameters

If you want, please set Hyper-parameters such as embedding dimension in main.py.

3. Train&Save

Please execute this command at the git project directory:

$ python main.py -f <data_path>

and a pickle file of an embedding vector will be saved in ./save/.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].