All Projects → inspur-hsslab → iMIX

inspur-hsslab / iMIX

Licence: Apache-2.0 license
A framework for Multimodal Intelligence research from Inspur HSSLAB.

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to iMIX

rosita
ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration
Stars: ✭ 36 (+71.43%)
Mutual labels:  vqa, vision-and-language
slp
Utils and modules for Speech Language and Multimodal processing using pytorch and pytorch lightning
Stars: ✭ 17 (-19.05%)
Mutual labels:  multimodal, multimodal-deep-learning
Mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
Stars: ✭ 4,713 (+22342.86%)
Mutual labels:  vqa, multimodal
just-ask
[TPAMI Special Issue on ICCV 2021 Best Papers, Oral] Just Ask: Learning to Answer Questions from Millions of Narrated Videos
Stars: ✭ 57 (+171.43%)
Mutual labels:  vqa, vision-and-language
VideoNavQA
An alternative EQA paradigm and informative benchmark + models (BMVC 2019, ViGIL 2019 spotlight)
Stars: ✭ 22 (+4.76%)
Mutual labels:  vqa, multimodal
pytorch violet
A PyTorch implementation of VIOLET
Stars: ✭ 119 (+466.67%)
Mutual labels:  vision-and-language
gakg
GAKG is a multimodal Geoscience Academic Knowledge Graph (GAKG) framework by fusing papers' illustrations, text, and bibliometric data.
Stars: ✭ 21 (+0%)
Mutual labels:  multimodal
MultiGraphGAN
MultiGraphGAN for predicting multiple target graphs from a source graph using geometric deep learning.
Stars: ✭ 16 (-23.81%)
Mutual labels:  multimodal-deep-learning
MinkLocMultimodal
MinkLoc++: Lidar and Monocular Image Fusion for Place Recognition
Stars: ✭ 65 (+209.52%)
Mutual labels:  multimodal
CBP
Official Tensorflow Implementation of the AAAI-2020 paper "Temporally Grounding Language Queries in Videos by Contextual Boundary-aware Prediction"
Stars: ✭ 52 (+147.62%)
Mutual labels:  vision-and-language
wikiHow paper list
A paper list of research conducted based on wikiHow
Stars: ✭ 25 (+19.05%)
Mutual labels:  vision-and-language
mix-stage
Official Repository for the paper Style Transfer for Co-Speech Gesture Animation: A Multi-Speaker Conditional-Mixture Approach published in ECCV 2020 (https://arxiv.org/abs/2007.12553)
Stars: ✭ 22 (+4.76%)
Mutual labels:  multimodal
Transformer-MM-Explainability
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (+2204.76%)
Mutual labels:  vqa
img2dataset
Easily turn large sets of image urls to an image dataset. Can download, resize and package 100M urls in 20h on one machine.
Stars: ✭ 1,173 (+5485.71%)
Mutual labels:  multimodal
mmgnn textvqa
A Pytorch implementation of CVPR 2020 paper: Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Text
Stars: ✭ 41 (+95.24%)
Mutual labels:  vqa
lipnet
LipNet with gluon
Stars: ✭ 16 (-23.81%)
Mutual labels:  multimodal
MSAF
Offical implementation of paper "MSAF: Multimodal Split Attention Fusion"
Stars: ✭ 47 (+123.81%)
Mutual labels:  multimodal-deep-learning
Social-IQ
[CVPR 2019 Oral] Social-IQ: A Question Answering Benchmark for Artificial Social Intelligence
Stars: ✭ 37 (+76.19%)
Mutual labels:  multimodal-deep-learning
attentive-modality-hopping-for-SER
TensorFlow implementation of "Attentive Modality Hopping for Speech Emotion Recognition," ICASSP-20
Stars: ✭ 25 (+19.05%)
Mutual labels:  multimodal-deep-learning
calvin
CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks
Stars: ✭ 105 (+400%)
Mutual labels:  vision-and-language

This project is not maintained,We have moved to a new product called MIX-Kalman

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].