All Projects → JesperDramsch → Seismic Transfer Learning

JesperDramsch / Seismic Transfer Learning

Licence: mit
Deep-learning seismic facies on state-of-the-art CNN architectures

Projects that are alternatives of or similar to Seismic Transfer Learning

Ner Bert
BERT-NER (nert-bert) with google bert https://github.com/google-research.
Stars: ✭ 339 (+959.38%)
Mutual labels:  jupyter-notebook, transfer-learning
Xlearn
Transfer Learning Library
Stars: ✭ 406 (+1168.75%)
Mutual labels:  jupyter-notebook, transfer-learning
Fast Pytorch
Pytorch Tutorial, Pytorch with Google Colab, Pytorch Implementations: CNN, RNN, DCGAN, Transfer Learning, Chatbot, Pytorch Sample Codes
Stars: ✭ 346 (+981.25%)
Mutual labels:  jupyter-notebook, transfer-learning
Deep-Learning
It contains the coursework and the practice I have done while learning Deep Learning.🚀 👨‍💻💥 🚩🌈
Stars: ✭ 21 (-34.37%)
Mutual labels:  transfer-learning, vgg16
Tensorflow 101
TensorFlow 101: Introduction to Deep Learning for Python Within TensorFlow
Stars: ✭ 642 (+1906.25%)
Mutual labels:  jupyter-notebook, transfer-learning
Grad Cam Tensorflow
tensorflow implementation of Grad-CAM (CNN visualization)
Stars: ✭ 261 (+715.63%)
Mutual labels:  jupyter-notebook, vgg16
Trainyourownyolo
Train a state-of-the-art yolov3 object detector from scratch!
Stars: ✭ 399 (+1146.88%)
Mutual labels:  jupyter-notebook, transfer-learning
Bert Sklearn
a sklearn wrapper for Google's BERT model
Stars: ✭ 182 (+468.75%)
Mutual labels:  jupyter-notebook, transfer-learning
Video Classification
Tutorial for video classification/ action recognition using 3D CNN/ CNN+RNN on UCF101
Stars: ✭ 543 (+1596.88%)
Mutual labels:  jupyter-notebook, transfer-learning
Food Recipe Cnn
food image to recipe with deep convolutional neural networks.
Stars: ✭ 448 (+1300%)
Mutual labels:  jupyter-notebook, vgg16
Skin Lesions Classification DCNNs
Transfer Learning with DCNNs (DenseNet, Inception V3, Inception-ResNet V2, VGG16) for skin lesions classification
Stars: ✭ 47 (+46.88%)
Mutual labels:  transfer-learning, vgg16
Skin Cancer Image Classification
Skin cancer classification using Inceptionv3
Stars: ✭ 16 (-50%)
Mutual labels:  jupyter-notebook, transfer-learning
Deeppicar
Deep Learning Autonomous Car based on Raspberry Pi, SunFounder PiCar-V Kit, TensorFlow, and Google's EdgeTPU Co-Processor
Stars: ✭ 242 (+656.25%)
Mutual labels:  jupyter-notebook, transfer-learning
Pytorch Nlp Notebooks
Learn how to use PyTorch to solve some common NLP problems with deep learning.
Stars: ✭ 293 (+815.63%)
Mutual labels:  jupyter-notebook, transfer-learning
Deep Learning With Python
Deep learning codes and projects using Python
Stars: ✭ 195 (+509.38%)
Mutual labels:  jupyter-notebook, vgg16
Amazon Forest Computer Vision
Amazon Forest Computer Vision: Satellite Image tagging code using PyTorch / Keras with lots of PyTorch tricks
Stars: ✭ 346 (+981.25%)
Mutual labels:  jupyter-notebook, transfer-learning
Cvpr18 Inaturalist Transfer
Large Scale Fine-Grained Categorization and Domain-Specific Transfer Learning. CVPR 2018
Stars: ✭ 164 (+412.5%)
Mutual labels:  jupyter-notebook, transfer-learning
Pytorch Retraining
Transfer Learning Shootout for PyTorch's model zoo (torchvision)
Stars: ✭ 167 (+421.88%)
Mutual labels:  jupyter-notebook, transfer-learning
Modelsgenesis
Official Keras & PyTorch Implementation and Pre-trained Models for Models Genesis - MICCAI 2019
Stars: ✭ 416 (+1200%)
Mutual labels:  jupyter-notebook, transfer-learning
Getting Things Done With Pytorch
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.
Stars: ✭ 738 (+2206.25%)
Mutual labels:  jupyter-notebook, transfer-learning

Deep learning seismic facies on state-of-the-art CNN architectures

Jesper S. Dramsch, Technical University of Denmark, and Mikael Lüthje, Technical University of Denmark

Abstract

We explore propagation of seismic interpretation by deep learning in stacked 2D sections. We show the application of state-of-the-art image classification algorithms on seismic data. These algorithms were trained on big labeled photograph databases. We use transfer learning to benefit from pre-trained networks and evaluate their performance on seismic data.

Presentation Date: Wednesday, October 17, 2018
Start Time: 8:30:00 AM
Location: 204B (Anaheim Convention Center)
Presentation Type: Oral

Citation

Paper

Jesper S. Dramsch and Mikael Lüthje (2018) Deep-learning seismic facies on state-of-the-art CNN architectures. SEG Technical Program Expanded Abstracts 2018: pp. 2036-2040.

Presentation

Dramsch, Jesper Soeren; Lüthje, Mikael (2018): Deep-learning seismic facies on state-of-the-art CNN architectures. figshare. Presentation. https://doi.org/10.6084/m9.figshare.7301645.v1

Code

Dramsch, Jesper Soeren; Lüthje, Mikael (2018): Deep-learning seismic facies on state-of-the-art CNN architectures. figshare. Code. https://doi.org/10.6084/m9.figshare.7227545

Usage

Interpretation of VGG

Interpretation of VGG

Loss of VGG

Loss of VGG

References

  • Abadi, M., A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G. S. Corrado, A. Davis, J. Dean, M. Devin, S. Ghemawat, I. Goodfellow, A. Harp, G. Irving, M. Isard, Y. Jia, R. Jozefowicz, L. Kaiser, M. Kudlur, J. Levenberg, D. Mane, R. Monga, S. Moore, D. Murray, C. Olah, M. Schuster, J. Shlens, B. Steiner, I. Sutskever, K. Talwar, P. Tucker, V. Vanhoucke, V. Vasudevan, F. Viegas, O. Vinyals, P. Warden, M. Wattenberg, M. Wicke, Y. Yu, and X. Zheng, 2015, TensorFlow: Large-scale machine learning on heterogeneous systems. (Software available from tensorflow.org).
  • Baxter, J., 1998, Theoretical models of learning to learn, in Learning to learn: Springer, 71–94.
  • Charles Rutherford Ildstad, P. B., 2017, MalenoV. Machine learning of Voxels.
  • Chollet, F., et al., 2015, Keras,
  • Dahl, G. E., T. N. Sainath, and G. E. Hinton, 2013, Improving deep neural networks for LVCSR using rectified linear units and dropout: Presented at the IEEE International Conference on Acoustics Speech and Signal Processing.
  • Deng, J., W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, 2009, ImageNet: A large-scale hierarchical image database: Presented at the CVPR09.
  • He, K., X. Zhang, S. Ren, and J. Sun, 2016, Deep residual learning for image recognition: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 770–778.
  • Krizhevsky, A., I. Sutskever, and G. E. Hinton, 2012, ImageNet classification with deep convolutional neural networks, in Advances in neural information processing systems: Curran Associates, Inc. 25, 1097–1105.
  • Lecun, Y., 1989, Generalization and network design strategies, in Connectionism in perspective: Elsevier.
  • Lin, T.-Y., P. Goyal, R. Girshick, K. He, and P. Dollar, 2017, Focal loss for dense object detection: arXiv preprint arXiv:1708.02002.
  • Long, J., E. Shelhamer, and T. Darrell, 2015, Fully convolutional networks for semantic segmentation: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 3431–3440.
  • Ruder, S., 2016, An overview of gradient descent optimization algorithms: arXiv preprint arXiv:1609.04747.
  • Rumelhart, D., G. Hinton, and R. Williams, 1988, Learning internal representations by error propagation, in Readings in cognitive science: Elsevier, 399–421.
  • Simonyan, K., and A. Zisserman, 2014, Very deep convolutional networks for large-scale image recognition: arXiv preprint arXiv:1409.1556.
  • Srivastava, N., G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, 2014, Dropout: A simple way to prevent neural networks from overfitting: Journal of Machine Learning Research, 15, 1929–1958.
  • Waldeland, A., and A. Solberg, 2016, 3D attributes and classification of salt bodies on unlabelled datasets: 78th Annual International Conference and Exhibition, EAGE, Extended Abstracts, https://doi.org/10.3997/2214-4609.201600880
  • Widrow, B., and M. Lehr, 1990, 30 years of adaptive neural networks: Perceptron Madaline, and backpropagation: Proceedings of the IEEE, 78, 1415–1442, https://doi.org/10.1109/5.58323
  • Yilmaz, Ö., 2001, Seismic data analysis: SEG.

Notes

We explore transfer training for automatic seismic interpretation without fine-tuning. See and cite the Powerpoint


Read More: https://library.seg.org/doi/abs/10.1190/segam2018-2996783.1 Or at: https://dramsch.net/#portfolio

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].