All Projects → salu133445 → bmusegan

salu133445 / bmusegan

Licence: MIT license
Code for “Convolutional Generative Adversarial Networks with Binary Neurons for Polyphonic Music Generation”

Programming Languages

python
139335 projects - #7 most used programming language
shell
77523 projects

Projects that are alternatives of or similar to bmusegan

Music-generation-cRNN-GAN
cRNN-GAN to generate music by training on instrumental music (midi)
Stars: ✭ 38 (-34.48%)
Mutual labels:  generative-adversarial-network, music-generation
binarygan
Code for "Training Generative Adversarial Networks with Binary Neurons by End-to-end Backpropagation"
Stars: ✭ 25 (-56.9%)
Mutual labels:  generative-adversarial-network, binary-neuron
seqgan-music
Implementation of a paper "Polyphonic Music Generation with Sequence Generative Adversarial Networks" in TensorFlow
Stars: ✭ 21 (-63.79%)
Mutual labels:  generative-adversarial-network, music-generation
speech-enhancement-WGAN
speech enhancement GAN on waveform/log-power-spectrum data using Improved WGAN
Stars: ✭ 35 (-39.66%)
Mutual labels:  generative-adversarial-network
gans-2.0
Generative Adversarial Networks in TensorFlow 2.0
Stars: ✭ 76 (+31.03%)
Mutual labels:  generative-adversarial-network
text2imageNet
Generate image from text with Generative Adversarial Network
Stars: ✭ 26 (-55.17%)
Mutual labels:  generative-adversarial-network
Pytorch models
PyTorch study
Stars: ✭ 14 (-75.86%)
Mutual labels:  generative-adversarial-network
FAST-RIR
This is the official implementation of our neural-network-based fast diffuse room impulse response generator (FAST-RIR) for generating room impulse responses (RIRs) for a given acoustic environment.
Stars: ✭ 90 (+55.17%)
Mutual labels:  generative-adversarial-network
text2image-benchmark
Performance comparison of existing GAN based Text To Image algorithms. (GAN-CLS, StackGAN, TAC-GAN)
Stars: ✭ 25 (-56.9%)
Mutual labels:  generative-adversarial-network
DCGAN-CIFAR10
A implementation of DCGAN (Deep Convolutional Generative Adversarial Networks) for CIFAR10 image
Stars: ✭ 18 (-68.97%)
Mutual labels:  generative-adversarial-network
SMILE
SMILE: Semantically-guided Multi-attribute Image and Layout Editing, ICCV Workshops 2021.
Stars: ✭ 28 (-51.72%)
Mutual labels:  generative-adversarial-network
GAN-LTH
[ICLR 2021] "GANs Can Play Lottery Too" by Xuxi Chen, Zhenyu Zhang, Yongduo Sui, Tianlong Chen
Stars: ✭ 24 (-58.62%)
Mutual labels:  generative-adversarial-network
Deep-Learning-Pytorch
A repo containing code covering various aspects of deep learning on Pytorch. Great for beginners and intermediate in the field
Stars: ✭ 59 (+1.72%)
Mutual labels:  generative-adversarial-network
Paper-Notes
Paper notes in deep learning/machine learning and computer vision
Stars: ✭ 37 (-36.21%)
Mutual labels:  generative-adversarial-network
timegan-pytorch
This repository is a non-official implementation of TimeGAN (Yoon et al., NIPS2019) using PyTorch.
Stars: ✭ 46 (-20.69%)
Mutual labels:  generative-adversarial-network
HashGAN
HashGAN: Deep Learning to Hash with Pair Conditional Wasserstein GAN
Stars: ✭ 63 (+8.62%)
Mutual labels:  generative-adversarial-network
MMD-GAN
Improving MMD-GAN training with repulsive loss function
Stars: ✭ 82 (+41.38%)
Mutual labels:  generative-adversarial-network
Self-Supervised-GANs
Tensorflow Implementation for paper "self-supervised generative adversarial networks"
Stars: ✭ 34 (-41.38%)
Mutual labels:  generative-adversarial-network
progressive-growing-of-gans.pytorch
Unofficial PyTorch implementation of "Progressive Growing of GANs for Improved Quality, Stability, and Variation".
Stars: ✭ 51 (-12.07%)
Mutual labels:  generative-adversarial-network
Deep-Fakes
No description or website provided.
Stars: ✭ 88 (+51.72%)
Mutual labels:  generative-adversarial-network

BinaryMuseGAN

BinaryMuseGAN is a follow-up project of the MuseGAN project. In this project, we first investigate how the real-valued piano-rolls generated by the generator may lead to difficulties in training the discriminator for CNN-based models. To overcome the binarization issue, we propose to append to the generator an additional refiner network, which try to refine the real-valued predictions generated by the pretrained generator to binary-valued ones. The proposed model is able to directly generate binary-valued piano-rolls at test time.

We trained the network with training data collected from Lakh Pianoroll Dataset. We used the model to generate four-bar musical phrases consisting of eight tracks: Drums, Piano, Guitar, Bass, Ensemble, Reed, Synth Lead and Synth Pad. Audio samples are available here.

Run the code

Configuration

Modify config.py for configuration.

  • Quick setup

    Change the values in the dictionary SETUP for a quick setup. Documentation is provided right after each key.

  • More configuration options

    Four dictionaries EXP_CONFIG, DATA_CONFIG, MODEL_CONFIG and TRAIN_CONFIG define experiment-, data-, model- and training-related configuration variables, respectively.

    The automatically-determined experiment name is based only on the values defined in the dictionary SETUP, so remember to provide the experiment name manually (so that you won't overwrite a trained model).

Run

python main.py

Training data

  • Prepare your own data

    The array will be reshaped to (-1, num_bar, num_timestep, num_pitch, num_track). These variables are defined in config.py.

  • Download our training data with this script or download it manually here.

Citing

Please cite the following paper if you use the code provided in this repository.

Hao-Wen Dong and Yi-Hsuan Yang, "Convolutional Generative Adversarial Networks with Binary Neurons for Polyphonic Music Generation," Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), 2018.
[homepage] [video] [paper] [slides] [slides (long)] [poster] [arXiv] [code]

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].