probnmn-clevrCode for ICML 2019 paper "Probabilistic Neural-symbolic Models for Interpretable Visual Question Answering" [long-oral]
Stars: ✭ 63 (+53.66%)
Vqa TensorflowTensorflow Implementation of Deeper LSTM+ normalized CNN for Visual Question Answering
Stars: ✭ 98 (+139.02%)
MICCAI21 MMQMultiple Meta-model Quantifying for Medical Visual Question Answering
Stars: ✭ 16 (-60.98%)
AoA-pytorchA Pytorch implementation of Attention on Attention module (both self and guided variants), for Visual Question Answering
Stars: ✭ 33 (-19.51%)
OscarOscar and VinVL
Stars: ✭ 396 (+865.85%)
Pytorch VqaStrong baseline for visual question answering
Stars: ✭ 158 (+285.37%)
FigureQA-baselineTensorFlow implementation of the CNN-LSTM, Relation Network and text-only baselines for the paper "FigureQA: An Annotated Figure Dataset for Visual Reasoning"
Stars: ✭ 28 (-31.71%)
plurPLUR (Programming-Language Understanding and Repair) is a collection of source code datasets suitable for graph-based machine learning. We provide scripts for downloading, processing, and loading the datasets. This is done by offering a unified API and data structures for all datasets.
Stars: ✭ 67 (+63.41%)
Bottom Up AttentionBottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
Stars: ✭ 989 (+2312.2%)
hcrn-videoqaImplementation for the paper "Hierarchical Conditional Relation Networks for Video Question Answering" (Le et al., CVPR 2020, Oral)
Stars: ✭ 111 (+170.73%)
GraknTypeDB: a strongly-typed database
Stars: ✭ 2,947 (+7087.8%)
Mac NetworkImplementation for the paper "Compositional Attention Networks for Machine Reasoning" (Hudson and Manning, ICLR 2018)
Stars: ✭ 444 (+982.93%)
OpenvqaA lightweight, scalable, and general framework for visual question answering research
Stars: ✭ 198 (+382.93%)
Awesome Visual Question AnsweringA curated list of Visual Question Answering(VQA)(Image/Video Question Answering),Visual Question Generation ,Visual Dialog ,Visual Commonsense Reasoning and related area.
Stars: ✭ 295 (+619.51%)
L2λ² is a tool for synthesizing functional programs from input-output examples.
Stars: ✭ 59 (+43.9%)
rositaROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration
Stars: ✭ 36 (-12.2%)
Vqa regatResearch Code for ICCV 2019 paper "Relation-aware Graph Attention Network for Visual Question Answering"
Stars: ✭ 129 (+214.63%)
just-ask[TPAMI Special Issue on ICCV 2021 Best Papers, Oral] Just Ask: Learning to Answer Questions from Millions of Narrated Videos
Stars: ✭ 57 (+39.02%)
ArcThe Abstraction and Reasoning Corpus
Stars: ✭ 1,598 (+3797.56%)
Transformer-MM-Explainability[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-based network. Including examples for DETR, VQA.
Stars: ✭ 484 (+1080.49%)
VqaCloudCV Visual Question Answering Demo
Stars: ✭ 57 (+39.02%)
ZS-F-VQACode and Data for paper: Zero-shot Visual Question Answering using Knowledge Graph [ ISWC 2021 ]
Stars: ✭ 51 (+24.39%)
GASGenerative Art Synthesizer - a python program that generates python programs that generates generative art
Stars: ✭ 42 (+2.44%)
appMonorepo for the client, server, etc. of the Debate Map website.
Stars: ✭ 53 (+29.27%)
lsw2OWL and Semantic Web toolkit for Common Lisp, used for construction and reasoning over ontologies and ontology-structured data
Stars: ✭ 22 (-46.34%)
MmfA modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
Stars: ✭ 4,713 (+11395.12%)
deepcoderDeepcoder paper implementation
Stars: ✭ 96 (+134.15%)
Awesome VqaVisual Q&A reading list
Stars: ✭ 403 (+882.93%)
Tbd NetsPyTorch implementation of "Transparency by Design: Closing the Gap Between Performance and Interpretability in Visual Reasoning"
Stars: ✭ 345 (+741.46%)
Clipbert[CVPR 2021 Oral] Official PyTorch code for ClipBERT, an efficient framework for end-to-end learning for image-text and video-text tasks.
Stars: ✭ 168 (+309.76%)
Nscl Pytorch ReleasePyTorch implementation for the Neuro-Symbolic Concept Learner (NS-CL).
Stars: ✭ 276 (+573.17%)
self critical vqaCode for NeurIPS 2019 paper ``Self-Critical Reasoning for Robust Visual Question Answering''
Stars: ✭ 39 (-4.88%)
bottom-up-featuresBottom-up features extractor implemented in PyTorch.
Stars: ✭ 62 (+51.22%)
vqa-softAccompanying code for "A Simple Loss Function for Improving the Convergence and Accuracy of Visual Question Answering Models" CVPR 2017 VQA workshop paper.
Stars: ✭ 14 (-65.85%)
suslikSynthesis of Heap-Manipulating Programs from Separation Logic
Stars: ✭ 107 (+160.98%)
DVQA datasetDVQA Dataset: A Bar chart question answering dataset presented at CVPR 2018
Stars: ✭ 20 (-51.22%)
Papers读过的CV方向的一些论文,图像生成文字、弱监督分割等
Stars: ✭ 99 (+141.46%)
VT-UNet[MICCAI2022] This is an official PyTorch implementation for A Robust Volumetric Transformer for Accurate 3D Tumor Segmentation
Stars: ✭ 151 (+268.29%)
typedbTypeDB: a strongly-typed database
Stars: ✭ 3,152 (+7587.8%)
iMIXA framework for Multimodal Intelligence research from Inspur HSSLAB.
Stars: ✭ 21 (-48.78%)
MullowbivqaHadamard Product for Low-rank Bilinear Pooling
Stars: ✭ 57 (+39.02%)
mmgnn textvqaA Pytorch implementation of CVPR 2020 paper: Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Text
Stars: ✭ 41 (+0%)
PopperPopper is an inductive logic programming (ILP) system.
Stars: ✭ 95 (+131.71%)
cfvqa[CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias
Stars: ✭ 96 (+134.15%)
Conditional Batch NormPytorch implementation of NIPS 2017 paper "Modulating early visual processing by language"
Stars: ✭ 51 (+24.39%)
VideoNavQAAn alternative EQA paradigm and informative benchmark + models (BMVC 2019, ViGIL 2019 spotlight)
Stars: ✭ 22 (-46.34%)
autogoalA Python framework for program synthesis with a focus on Automated Machine Learning.
Stars: ✭ 153 (+273.17%)
typeqlTypeQL: the query language of TypeDB - a strongly-typed database
Stars: ✭ 157 (+282.93%)
Vizwiz Vqa PytorchPyTorch VQA implementation that achieved top performances in the (ECCV18) VizWiz Grand Challenge: Answering Visual Questions from Blind People
Stars: ✭ 33 (-19.51%)
iPerceiveApplying Common-Sense Reasoning to Multi-Modal Dense Video Captioning and Video Question Answering | Python3 | PyTorch | CNNs | Causality | Reasoning | LSTMs | Transformers | Multi-Head Self Attention | Published in IEEE Winter Conference on Applications of Computer Vision (WACV) 2021
Stars: ✭ 52 (+26.83%)
neural inverse knittingCode for Neural Inverse Knitting: From Images to Manufacturing Instructions
Stars: ✭ 30 (-26.83%)
Bottom Up Attention VqaAn efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
Stars: ✭ 667 (+1526.83%)
pytorch-convcnpA PyTorch Implementation of Convolutional Conditional Neural Process.
Stars: ✭ 41 (+0%)
appsAPPS: Automated Programming Progress Standard (NeurIPS 2021)
Stars: ✭ 174 (+324.39%)
Vqa.pytorchVisual Question Answering in Pytorch
Stars: ✭ 602 (+1368.29%)