All Projects → deontaepharr → Retinal-Disease-Diagnosis-With-Residual-Attention-Networks

deontaepharr / Retinal-Disease-Diagnosis-With-Residual-Attention-Networks

Licence: other
Using Residual Attention Networks to diagnose retinal diseases in medical images

Programming Languages

python
139335 projects - #7 most used programming language

Projects that are alternatives of or similar to Retinal-Disease-Diagnosis-With-Residual-Attention-Networks

Medicalzoopytorch
A pytorch-based deep learning framework for multi-modal 2D/3D medical image segmentation
Stars: ✭ 546 (+3800%)
Mutual labels:  medical-imaging, resnet
Eeg Dl
A Deep Learning library for EEG Tasks (Signals) Classification, based on TensorFlow.
Stars: ✭ 165 (+1078.57%)
Mutual labels:  resnet, attention-mechanism
resnet-cifar10
ResNet for Cifar10
Stars: ✭ 21 (+50%)
Mutual labels:  resnet, residual-networks
Sca Cnn.cvpr17
Image Captions Generation with Spatial and Channel-wise Attention
Stars: ✭ 198 (+1314.29%)
Mutual labels:  resnet, attention-mechanism
caffe-wrn-generator
Caffe Wide-Residual-Network (WRN) Generator
Stars: ✭ 19 (+35.71%)
Mutual labels:  resnet, residual-networks
ReZero-ResNet
Unofficial pytorch implementation of ReZero in ResNet
Stars: ✭ 23 (+64.29%)
Mutual labels:  resnet, residual-networks
wideresnet-tensorlayer
Wide Residual Networks implemented in TensorLayer and TensorFlow.
Stars: ✭ 44 (+214.29%)
Mutual labels:  resnet, residual-networks
pyro-vision
Computer vision library for wildfire detection
Stars: ✭ 33 (+135.71%)
Mutual labels:  resnet
Attention
一些不同的Attention机制代码
Stars: ✭ 17 (+21.43%)
Mutual labels:  attention-mechanism
mri-deep-learning-tools
Resurces for MRI images processing and deep learning in 3D
Stars: ✭ 56 (+300%)
Mutual labels:  medical-imaging
MoChA-pytorch
PyTorch Implementation of "Monotonic Chunkwise Attention" (ICLR 2018)
Stars: ✭ 65 (+364.29%)
Mutual labels:  attention-mechanism
Keras-CIFAR10
practice on CIFAR10 with Keras
Stars: ✭ 25 (+78.57%)
Mutual labels:  resnet
RMNet
RM Operation can equivalently convert ResNet to VGG, which is better for pruning; and can help RepVGG perform better when the depth is large.
Stars: ✭ 129 (+821.43%)
Mutual labels:  resnet
coursera-ai-for-medicine-specialization
Programming assignments, labs and quizzes from all courses in the Coursera AI for Medicine Specialization offered by deeplearning.ai
Stars: ✭ 80 (+471.43%)
Mutual labels:  medical-imaging
DE resnet unet hyb
Depth estimation from RGB images using fully convolutional neural networks.
Stars: ✭ 40 (+185.71%)
Mutual labels:  resnet
pytorch2keras
PyTorch to Keras model convertor
Stars: ✭ 788 (+5528.57%)
Mutual labels:  resnet
neurdicom
RESTful PACS server with plugins
Stars: ✭ 97 (+592.86%)
Mutual labels:  medical-imaging
pynmt
a simple and complete pytorch implementation of neural machine translation system
Stars: ✭ 13 (-7.14%)
Mutual labels:  attention-mechanism
co-attention
Pytorch implementation of "Dynamic Coattention Networks For Question Answering"
Stars: ✭ 54 (+285.71%)
Mutual labels:  attention-mechanism
fw4spl
Main repository for fw4spl
Stars: ✭ 30 (+114.29%)
Mutual labels:  medical-imaging

Identifying Eye Diseases with Residual Attention Networks

Eye Diseases

  • The leading causes of blindness
    • Age-related macular degeneration: Almost 10mill individuals suffer in the US
    • Diabetic Retinopathy: Nearly 750K individuals aged 40+ suffer
  • Diseases likely to increase due to the aging population and global diabetes epidemic
  • Optical Coherence Tomography (OCT) Imaging special way to visualize individual retina layers so that specialists can analyze the state of the retina
  • With the OCT images, creating a diagnostic tool based on Residual Attention Networks (RAN) for the screening of patients with common treatable blinding retinal diseases may ultimately aid in expediting the diagnosis and referral of treatable conditions 

Optical Coherence Tomography (OCT) Imaging

  • OCT imaging is the standard retinal imaging tool for the diagnosis and treatment of some of the leading causes of blindness
    • approx. 30 million OCT scans are performed each year world-wide
  • Provides a clear visualization of individual retinal layers that would be impossible with clinical examination by the human eye
  • OCT uses light waves to capture distinctive layers of the retina in vivo

Residual Attention Networks

  • Paper Title: “Residual Attention Network for Image Classification”
  • Authors: Wang, Fei and Jiang, Mengqing and Qian, Chen and Yang, Shuo and Li, Cheng and Zhang, Honggang and Wang, Xiaogang and Tang, Xiaoou
  • Convolutional Neural Network with Attention Mechanism and Residual Units
  • Built by stacking Attention Modules, which generate attention-aware features.
  • These attention-aware features from different modules change adaptively as layers going deeper and brings more discriminative feature representation
  • Inside each Attention Module, bottom-up top-down feedforward structure is used to unfold the feedforward and feedback attention process into a single feedforward process.
  • Implements attention residual learning to train very deep Residual Attention Networks

Interesting Properties About Architecture

  • Increasing Attention Modules leads to consistent performance improvement
    • With multiple Attention Modules, different types of attention are able to be captured
  • Attention Residual Learning: naively stacking Attention Modules directly would lead to the obvious performance drop.
    • Proposes attention residual learning mechanism - optimizes very deep Residual Attention Network with hundreds of layers.
  • Bottom-Up Top-Down Feedforward Attention Structure as part of Attention - Module to add soft weights on features.
  • Mimics bottom-up fast feedforward process and top-down attention feedback in a single feedforward process
    • Allows an end-to-end trainable network

Attention Modules

  • Main component of network
  • Divided into two branches (components)
    • Mask branch
    • Trunk branch
  • Includes Residual Units within module and branches

Residual Units

  • Inspired by Residual Neural Network (ResNet)
  • Utilizes skip-connections to jump over 2–3 layers with nonlinearities (e.g. in ReLU CNNs) and batch normalizations
  • Motivation for skipping: to avoid the vanishing gradients and degradation problem
  • Vanishing Gradient: when the gradient becomes vanishingly small due to the deepness of the network
    • Thus preventing the weights from changing its value and stopping the network from further training
  • The degradation problem occurs as network depth increases. As additional layers are added to the network, its accuracy decreases despite being capable of gathering all of the intricacies of our data.

Branches

  • Trunk Branch performs feature processing with Residual Units
  • Mask Branch uses bottom-up top-down structure softly weight output features with the goal of improving trunk branch features
  • Bottom-up Step: collects global information of the whole image by downsampling (i.e. max pooling) the image
  • Top-down Step: combines global information with original feature maps by upsampling (i.e. bilinear interpolation) to keep the output size the same as the input feature map.
  • Two consecutive 1 × 1 convolution layers follows
  • Finally, a sigmoid layer normalizes the output range to [0, 1]

Attention Residual Learning

  • Naive stacking Attention Modules leads to a performance drop
    • Dot production with mask range from zero to one repeatedly will degrade the value of features in deep layers.
    • Soft mask can potentially break good property of trunk branch, for example, the identical mapping of Residual Unit.
  • Attention Residual Learning (ARL) eases the above problems.
  • Similar to ideas in residual learning, if soft mask unit can be constructed as identical mapping,
  • Modify output H of Attention Module as
    • Original Residual Learning Formula: H(x) = F(x) * x
    • ARL Formula: H(x) = F(x) * (1+M(x))
  • The key difference lies in the mask branches M(x). They work as feature selectors which enhance good features and suppress noises from trunk features.

Retina OCT Dataset

  • Before training, each image went through a tiered grading system of trained graders of increasing expertise for verification and correction of image labels.
    • First tier of graders: undergraduate and medical students who had taken and passed an OCT interpretation course review
    • Second tier of graders: four ophthalmologists who independently graded each image that had passed the first tier
    • Third tier of graders: two senior independent retinal specialists, each with over 20 years of clinical retina experience, verified the true labels for each image.
  • Resulted in 84,484 images
  • 4 Categories
    • Choroidal Neovascularization – 37,455
    • Diabetic Macular Edema – 11,598
    • Drusen - 8,866
    • Normal – 26,565

Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning

  • Paper that inspired this project
  • Authors : Daniel S. Kermany, Michael Goldbaum, Wenjia Cai, ..., M. Anthony Lewis, Huimin Xia, Kang Zhang
  • Utilized a pre-trained network which demonstrated competitive performance of OCT image analysis that was comparable to that of human experts with significant clinical experience with retinal diseases.
  • Obtained immaculate results
  • Read paper here: Paper Link

Implementation

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].