All Projects → NU-CUCIS → Elemnet

NU-CUCIS / Elemnet

Licence: other
Deep Learning the Chemistry of Materials From Only Elemental Composition for Enhancing Materials Property Prediction

Projects that are alternatives of or similar to Elemnet

Deep Learning Time Series
List of papers, code and experiments using deep learning for time series forecasting
Stars: ✭ 796 (+1709.09%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Deepfake Detection
DeepFake Detection: Detect the video is fake or not using InceptionResNetV2.
Stars: ✭ 23 (-47.73%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Deep Embedded Memory Networks
https://arxiv.org/abs/1707.00836
Stars: ✭ 19 (-56.82%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Saliency
TensorFlow implementation for SmoothGrad, Grad-CAM, Guided backprop, Integrated Gradients and other saliency techniques
Stars: ✭ 648 (+1372.73%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Servenet
Service Classification based on Service Description
Stars: ✭ 21 (-52.27%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Pytorch Multi Style Transfer
Neural Style and MSG-Net
Stars: ✭ 687 (+1461.36%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Csc deeplearning
3-day dive into deep learning at csc
Stars: ✭ 22 (-50%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Easy Deep Learning With Keras
Keras tutorial for beginners (using TF backend)
Stars: ✭ 367 (+734.09%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Kalasalingam
IEEE "Invited Talk on Deep Learning" 03/02/2018
Stars: ✭ 13 (-70.45%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Tf Keras Surgeon
Pruning and other network surgery for trained TF.Keras models.
Stars: ✭ 25 (-43.18%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Speech Emotion Analyzer
The neural network model is capable of detecting five different male/female emotions from audio speeches. (Deep Learning, NLP, Python)
Stars: ✭ 633 (+1338.64%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Dl Colab Notebooks
Try out deep learning models online on Google Colab
Stars: ✭ 969 (+2102.27%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Stock Analysis Engine
Backtest 1000s of minute-by-minute trading algorithms for training AI with automated pricing data from: IEX, Tradier and FinViz. Datasets and trading performance automatically published to S3 for building AI training datasets for teaching DNNs how to trade. Runs on Kubernetes and docker-compose. >150 million trading history rows generated from +5000 algorithms. Heads up: Yahoo's Finance API was disabled on 2019-01-03 https://developer.yahoo.com/yql/
Stars: ✭ 605 (+1275%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Gans In Action
Companion repository to GANs in Action: Deep learning with Generative Adversarial Networks
Stars: ✭ 748 (+1600%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Docs
TensorFlow documentation
Stars: ✭ 4,999 (+11261.36%)
Mutual labels:  jupyter-notebook, deep-neural-networks
All Classifiers 2019
A collection of computer vision projects for Acute Lymphoblastic Leukemia classification/early detection.
Stars: ✭ 22 (-50%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Action Recognition Visual Attention
Action recognition using soft attention based deep recurrent neural networks
Stars: ✭ 350 (+695.45%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Magnet
Deep Learning Projects that Build Themselves
Stars: ✭ 351 (+697.73%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Concise Ipython Notebooks For Deep Learning
Ipython Notebooks for solving problems like classification, segmentation, generation using latest Deep learning algorithms on different publicly available text and image data-sets.
Stars: ✭ 23 (-47.73%)
Mutual labels:  jupyter-notebook, deep-neural-networks
Densedepth
High Quality Monocular Depth Estimation via Transfer Learning
Stars: ✭ 963 (+2088.64%)
Mutual labels:  jupyter-notebook, deep-neural-networks

ElemNet

ElemNet is a deep neural network model that takes only the elemental compositions as inputs and leverages artificial intelligence to automatically capture the essential chemistry to predict materials properties. ElemNet can automatically learn the chemical interactions and similarities between different elements which allows it to even predict the phase diagrams of chemical systems absent from the training dataset more accurately than the conventional machine learning models based on physical attributes levaraging domain knowledge.

This repository contains the code for performing data processing, model training and analysis along with the trained model. If you have a large dataset such as OQMD, the model should be trained from scratch. Otherwise for smaller DFT-computed or experimental datasets, it is best to train the model using trasfer learning from a pretrained model as shown below.

Installation Requirements

The basic requirement for re-using these environments are a Python 3.6.3 Jupyter environment with the packages listed in requirements.txt.

Some analyses required the use of Magpie, which requires Java JDK 1.7 or greater. See [the Magpie documentation for details].

Source Files

The code for training the ElemNet model along with the trained model produced in our work [1] are available in the elemnet folder. The other folders contain scripts associted with different analyses performed to characterize ElemNet. The analysis notebooks should be self-describing, in other case there is a README file that describes the folder content.

Please download and extract files for the following paths as required:

Here is a brief explanation about the folder content:

  • chemical-interpolation: code for creating training sets for chemical interpolation test to determine whether machine learning models are able to infer the interactions between elements that are not included in the training set by plotting their phase diagrams [1].

  • elemnet: code for data preprocessing and training the ElemNet model from scratch [1] and using transfer learning from a pretrained model [2], and also for making predictions using a trained model. It contains a README file that explains the details about training and making predictions using ElemNet model.

  • evaluate-predictions: code for evaluating the predictions made by the ElemNet model [1].

  • icsd-search: list of matched compositions between ICSD and ones from combinatorial screening using ElemNet [1].

  • layer-projection: analysis of the activations from ElemNet to interpret how the model is learning the underlying chemistry.

  • data: the datasets used for training ElemNet model in [1] and [2]. We use 'oqmd_all.data' in [1]. For [2], we used 'oqmd_all-22Mar18.csv' (OQMD), 'jv.csv' (JARVIS), 'mp.csv' (the Materials Project) and 'exp.csv' (experimental observations) as the starting datasets. We also provide the preprocessed experimental data- 'experimental.csv', which contains the elemental fractions and can be used for creating test, train and validation splits; the respective training and test sets for both 8:2 and 9:1 hold-out train:test splits are provided inside 'holdout-new/20' and 'holdout-new/10' folders respectively.

Running the code

The code to run the ElemNet model is privided inside the elemnet folder. Inside this folder, you can simply run the code by passing a sample config file to the dl_regressors.py as follows:

python dl_regressors.py --config_file sample/sample-run.config

The config file defines the loss_type, training_data_path, test_data_path, label, input_type [elements_tl for ElemNet] and other runtime parameters. For transfer learning, you need to set 'model_path' [e.g. sample/sample_model]. The output log from this sample run is provided in the sample/sample.log file.

Developer Team

The code was developed by Dipendra Jha from the CUCIS group at the Electrical and Computer Engineering Department at Northwestern University and Logan Ward from the Computation Institute at University of Chicago.

Publications

Please cite the following works if you are using ElemNet model:

  1. Dipendra Jha, Logan Ward, Arindam Paul, Wei-keng Liao, Alok Choudhary, Chris Wolverton, and Ankit Agrawal, “ElemNet: Deep Learning the Chemistry of Materials From Only Elemental Composition,” Scientific Reports, 8, Article number: 17593 (2018) [DOI:10.1038/s41598-018-35934-y] [PDF].

  2. Dipendra Jha, Kamal Choudhary, Francesca Tavazza, Wei-keng Liao, Alok Choudhary, Carelyn Campbell, Ankit Agrawal, "Enhancing materials property prediction by leveraging computational and experimental data using deep transfer learning," Nature Communications, 10, Article number: 5316 (2019) [DOI: https:10.1038/s41467-019-13297-w] [PDF].

Questions/Comments

email: [email protected], [email protected] or [email protected] Copyright (C) 2019, Northwestern University.
See COPYRIGHT notice in top-level directory.

Funding Support

This work was performed under the following financial assistance award 70NANB14H012 and 70NANB19H005 from U.S. Department of Commerce, National Institute of Standards and Technology as part of the Center for Hierarchical Materials Design (CHiMaD). Partial support is also acknowledged from DOE awards DE-SC0014330, DE-SC0019358.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].