All Projects → nlpub → russe

nlpub / russe

Licence: other
RUSSE: Russian Semantic Evaluation.

Programming Languages

HTML
75241 projects
SCSS
7915 projects
ruby
36898 projects - #4 most used programming language

Projects that are alternatives of or similar to russe

Tensorflow Tutorials
텐서플로우를 기초부터 응용까지 단계별로 연습할 수 있는 소스 코드를 제공합니다
Stars: ✭ 2,096 (+18954.55%)
Mutual labels:  word2vec
Sensegram
Making sense embedding out of word embeddings using graph-based word sense induction
Stars: ✭ 209 (+1800%)
Mutual labels:  word2vec
Book deeplearning in pytorch source
Stars: ✭ 236 (+2045.45%)
Mutual labels:  word2vec
Text Pairs Relation Classification
About Text Pairs (Sentence Level) Classification (Similarity Modeling) Based on Neural Network.
Stars: ✭ 182 (+1554.55%)
Mutual labels:  word2vec
Chameleon recsys
Source code of CHAMELEON - A Deep Learning Meta-Architecture for News Recommender Systems
Stars: ✭ 202 (+1736.36%)
Mutual labels:  word2vec
Stocksensation
基于情感字典和机器学习的股市舆情情感分类可视化Web
Stars: ✭ 215 (+1854.55%)
Mutual labels:  word2vec
Debiaswe
Remove problematic gender bias from word embeddings.
Stars: ✭ 175 (+1490.91%)
Mutual labels:  word2vec
Siamese-Sentence-Similarity
Keras and Tensorflow implementation of Siamese Recurrent Architectures for Learning Sentence Similarity
Stars: ✭ 47 (+327.27%)
Mutual labels:  semantic-similarity
Word2vec
Python interface to Google word2vec
Stars: ✭ 2,370 (+21445.45%)
Mutual labels:  word2vec
Koan
A word2vec negative sampling implementation with correct CBOW update.
Stars: ✭ 232 (+2009.09%)
Mutual labels:  word2vec
Nlp learning
结合python一起学习自然语言处理 (nlp): 语言模型、HMM、PCFG、Word2vec、完形填空式阅读理解任务、朴素贝叶斯分类器、TFIDF、PCA、SVD
Stars: ✭ 188 (+1609.09%)
Mutual labels:  word2vec
Shallowlearn
An experiment about re-implementing supervised learning models based on shallow neural network approaches (e.g. fastText) with some additional exclusive features and nice API. Written in Python and fully compatible with Scikit-learn.
Stars: ✭ 196 (+1681.82%)
Mutual labels:  word2vec
Practical 1
Oxford Deep NLP 2017 course - Practical 1: word2vec
Stars: ✭ 220 (+1900%)
Mutual labels:  word2vec
Sentence Similarity
对四种句子/文本相似度计算方法进行实验与比较
Stars: ✭ 181 (+1545.45%)
Mutual labels:  word2vec
Aravec
AraVec is a pre-trained distributed word representation (word embedding) open source project which aims to provide the Arabic NLP research community with free to use and powerful word embedding models.
Stars: ✭ 239 (+2072.73%)
Mutual labels:  word2vec
Splitter
A Pytorch implementation of "Splitter: Learning Node Representations that Capture Multiple Social Contexts" (WWW 2019).
Stars: ✭ 177 (+1509.09%)
Mutual labels:  word2vec
Gemsec
The TensorFlow reference implementation of 'GEMSEC: Graph Embedding with Self Clustering' (ASONAM 2019).
Stars: ✭ 210 (+1809.09%)
Mutual labels:  word2vec
Cukatify
Cukatify is a music social media project
Stars: ✭ 21 (+90.91%)
Mutual labels:  word2vec
Movietaster Open
A practical movie recommend project based on Item2vec.
Stars: ✭ 253 (+2200%)
Mutual labels:  word2vec
Cw2vec
cw2vec: Learning Chinese Word Embeddings with Stroke n-gram Information
Stars: ✭ 224 (+1936.36%)
Mutual labels:  word2vec

The First International Workshop on Russian Semantic Similarity Evaluation (RUSSE)

Motivation

A similarity measure is a numerical measure of the degree the two objects are alike. Usually, it quantifies similarity with a scalar in range [0; 1] or [0; ∞]. A semantic similarity measure is a specific similarity measure designed to quantify semantic relatedness of lexical units (e.g. nouns and multiword expressions). It yields high values for the pairs of words in a semantic relation (synonyms, hyponyms, associations or co-hyponyms) and zero values for all other pairs.

Semantic similarity measures proved to be useful for text processing applications, including text similarity, query expansion, question answering and word sense disambiguation. Such measures are practical because of the gap between lexical surface of the text and its meaning. Indeed, the same concept is often represented by different terms. Furthermore, these measures can be useful in linguistic and philological studies.

Measures of semantic similarity is an actively developing field of computational linguistics. Many methods were proposed and tested during last 20 years. Recently with the advent of neural network language models yielding state-of-the-art results on the semantic similarity task the interest to this field increased even more. Many authors tried to performed exhaustive comparisons of semantic similarity measures and developed a whole range of benchmarks and evaluations datasets.

Contribution

Unfortunately, most of the approaches to semantic similarity were implemented and evaluated only on a handful of European languages, mostly in English. While some Russian researchers sporadically tried to adopt several methods developed for English, these efforts were mostly done in a context of some specific applications without any proper evaluation. To the best of our knowledge, no systematic investigation of semantic similarity measures of Russian language was ever performed.

Expected Results

The goal of the RUSSE is to fill this gap and to conduct an evaluation campaign of key currently available methods. The RUSSE competition will perform a systematic comparison and evaluation of the baseline and the most recent approaches to semantic similarity in the context of Russian language. This will let us identify specific features of the semantic similarity phenomena in Russian language. The event will be organized in a form of a competition of systems that calculate similarity between words.

Contacts

Further details, including task rationale, schedule and datasets can be found on the RUSSE website: http://russe.nlpub.ru/. Participants will be invited to submit a paper to the Dialogue-2015 conference describing their system.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].