All Projects → watsonyanghx → Image Text Papers

watsonyanghx / Image Text Papers

Licence: mit
Image Caption and Text to Image papers.

Projects that are alternatives of or similar to Image Text Papers

Contrastive Unpaired Translation
Contrastive unpaired image-to-image translation, faster and lighter training than cyclegan (ECCV 2020, in PyTorch)
Stars: ✭ 822 (+1057.75%)
Mutual labels:  image-generation
Rnn Vae
Variational Autoencoder with Recurrent Neural Network based on Google DeepMind's "DRAW: A Recurrent Neural Network For Image Generation"
Stars: ✭ 39 (-45.07%)
Mutual labels:  image-generation
Coco Cn
Enriching MS-COCO with Chinese sentences and tags for cross-lingual multimedia tasks
Stars: ✭ 57 (-19.72%)
Mutual labels:  image-captioning
Domain Transfer Network
TensorFlow Implementation of Unsupervised Cross-Domain Image Generation
Stars: ✭ 850 (+1097.18%)
Mutual labels:  image-generation
Punny captions
An implementation of the NAACL 2018 paper "Punny Captions: Witty Wordplay in Image Descriptions".
Stars: ✭ 31 (-56.34%)
Mutual labels:  image-captioning
Knpsnappybundle
Easily create PDF and images in Symfony by converting html using webkit
Stars: ✭ 1,038 (+1361.97%)
Mutual labels:  image-generation
Mixnmatch
Pytorch implementation of MixNMatch
Stars: ✭ 694 (+877.46%)
Mutual labels:  image-generation
Pix2pix
Image-to-image translation with conditional adversarial nets
Stars: ✭ 8,765 (+12245.07%)
Mutual labels:  image-generation
Bottom Up Attention
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
Stars: ✭ 989 (+1292.96%)
Mutual labels:  image-captioning
Image Captioning
Image Captioning: Implementing the Neural Image Caption Generator with python
Stars: ✭ 52 (-26.76%)
Mutual labels:  image-captioning
Show Attend And Tell
TensorFlow Implementation of "Show, Attend and Tell"
Stars: ✭ 869 (+1123.94%)
Mutual labels:  image-captioning
Im2p
Tensorflow implementation of paper: A Hierarchical Approach for Generating Descriptive Image Paragraphs
Stars: ✭ 15 (-78.87%)
Mutual labels:  image-captioning
Img2imggan
Implementation of the paper : "Toward Multimodal Image-to-Image Translation"
Stars: ✭ 49 (-30.99%)
Mutual labels:  image-generation
Multi Viewpoint Image Generation
Given an image and a target viewpoint, generate synthetic image in the target viewpoint
Stars: ✭ 23 (-67.61%)
Mutual labels:  image-generation
Matlab Gan
MATLAB implementations of Generative Adversarial Networks -- from GAN to Pixel2Pixel, CycleGAN
Stars: ✭ 63 (-11.27%)
Mutual labels:  image-generation
Self Critical.pytorch
Unofficial pytorch implementation for Self-critical Sequence Training for Image Captioning. and others.
Stars: ✭ 716 (+908.45%)
Mutual labels:  image-captioning
Congan
Continious Generative Adversarial Network
Stars: ✭ 41 (-42.25%)
Mutual labels:  image-generation
Dcgan Tensorflow
A Tensorflow implementation of Deep Convolutional Generative Adversarial Networks trained on Fashion-MNIST, CIFAR-10, etc.
Stars: ✭ 70 (-1.41%)
Mutual labels:  image-generation
Cameramanager
Simple Swift class to provide all the configurations you need to create custom camera view in your app
Stars: ✭ 1,130 (+1491.55%)
Mutual labels:  image-captioning
Image captioning
generate captions for images using a CNN-RNN model that is trained on the Microsoft Common Objects in COntext (MS COCO) dataset
Stars: ✭ 51 (-28.17%)
Mutual labels:  image-captioning

Image-Text-Papers

Image Caption and Image Generation related papers.

Still working on it ....


Image Caption (Image --> Text)

Survey

  • Bernardi, Raffaella, et al. Automatic Description Generation from Images: A Survey of Models, Datasets, and Evaluation Measures. J. Artif. Intell. Res.(JAIR) 55 (2016): 409-442. [pdf]

  • Karpathy, Andrej. CONNECTING IMAGES AND NATURAL LANGUAGE. Diss. STANFORD UNIVERSITY, 2016. [pdf]

Visual-semantic Embedding Based

  • Kiros R, Salakhutdinov R, Zemel R S. Unifying visual-semantic embeddings with multimodal neural language models. arXiv preprint arXiv:1411.2539, 2014. [pdf]

  • Karpathy A, Fei-Fei L. Deep visual-semantic alignments for generating image descriptions. CVPR, 2015: 3128-3137. [pdf]

Encoder-Decoder

  • Vinyals, Oriol, et al. Show and tell: A neural image caption generator. CVPR, 2015. [pdf]

  • Xu, Kelvin, et al. Show, attend and tell: Neural image caption generation with visual attention. ICML, 2015. [pdf]

  • Karpathy, Andrej, and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. CVPR, 2015. [pdf]

  • Anderson, Peter, et al. Bottom-up and top-down attention for image captioning and VQA. arXiv preprint arXiv:1707.07998 (2017). [pdf]

Reinforcement Learning

  • Rennie, Steven J., et al. Self-critical Sequence Training for Image Captioning. CVPR, 2017. [pdf]

  • Liu, Siqi, et al. Improved Image Captioning via Policy Gradient optimization of SPIDEr. ICCV, 2017. [pdf] [video]

  • Zhou Ren, Xiaoyu Wang, Ning Zhang, et al. Deep Reinforcement Learning-based Image Captioning with Embedding Reward. CVPR, 2017. [pdf] [video]

  • Chen T H, Liao Y H, Chuang C Y, et al. Show, Adapt and Tell: Adversarial Training of Cross-domain Image Captioner[C]. ICCV, 2017. [pdf] [Supplementary]

  • Dai B, Lin D, Urtasun R, et al. Towards diverse and natural image descriptions via a conditional gan. ICCV, 2017. [pdf] [video]


Image Generation (Text --> Image)

RNN

  • Oord, Aaron van den, Nal Kalchbrenner, and Koray Kavukcuoglu. Pixel recurrent neural networks. arXiv preprint arXiv:1601.06759 (2016). [pdf]

  • Zhang H, Xu T, Li H, et al. Gregor, Karol, et al. DRAW: A recurrent neural network for image generation. arXiv preprint arXiv:1502.04623 (2015). [pdf]

  • Mansimov, Elman, et al. Generating images from captions with attention. arXiv preprint arXiv:1511.02793 (2015). [pdf]

GAN

  • Gauthier, Jon. Conditional generative adversarial nets for convolutional face generation. Class Project for Stanford CS231N: Convolutional Neural Networks for Visual Recognition, Winter semester 2014.5 (2014): 2. [pdf]

  • Reed S, Akata Z, Yan X, et al. Generative adversarial text to image synthesis. ICML, 2016. [pdf] [Supplementary]

  • Reed, Scott E., et al. Learning what and where to draw. NIPS, 2016. [pdf]

  • Zhang H, Xu T, Li H, et al. StackGAN: Text to Photo-realistic Image Synthesis with Stacked Generative Adversarial Networks. ICCV, 2017. [pdf] [video]

  • Zhang H, Xu T, Li H, et al. StackGAN++: Realistic Image Synthesis with Stacked Generative Adversarial Networks. arXiv preprint arXiv:1710.10916, 2017. [pdf]

  • Xu, Tao, et al. AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial Networks. arXiv preprint arXiv:1711.10485 (2017). [pdf]

  • Hao Dong, Simiao Yu, Chao Wu, Yike Guo. Semantic Image Synthesis via Adversarial Learning. ICCV, 2017. [pdf] [Supplementary]

  • Ayushman, John, et al. TAC-GAN - Text Conditioned Auxiliary Classifier Generative Adversarial Network . arXiv preprint arXiv:1703.06412, 2017. [pdf]

  • Nguyen, Anh, et al. Plug & play generative networks: Conditional iterative generation of images in latent space. CVPR, 2017. [pdf] [Supplementary]

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].