Skip to content

NYUMedML/plaquebox-paper

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

46 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Interpretable Classification of Alzheimer's Disease Pathologies with a Convolutional Neural Network Pipeline

bioRxiv 454793

Zenodo Data Available at: https://doi.org/10.5281/zenodo.1470797

This repository accompanies the publication above. Specifically, we include Jupyter notebooks to reproduce all image preprocessing and processing, training of convolutional neural networks, confidence visualizations, and saliency maps. The available code is provided as-is, and does not constitute a full-fledged software package for analysis.

Systems requirements

This code repository was developed on Linux CentOS 7 and Ubuntu 18 and has not been tested on on other systems (Windows, MacOS).

Code requires the following Python packages:

python                    3.6.5                hc3d631a_2  
ipython                   6.4.0                    py36_0  
jupyter                   1.0.0                    py36_4  
matplotlib                2.2.2            py36h0e671d2_1  
numpy                     1.14.3           py36hcd700cb_1  
pandas                    0.23.0           py36h637b7d7_0  
scikit-learn              0.19.1           py36h7aa7ec6_0     
scikit-image              0.13.1           py36h14c3975_1    
scipy                     1.1.0            py36hfc37229_0  
pytorch                   0.3.0            py35_cuda8.0.61_cudnn7.0.3hb362f6e_4    pytorch
torchvision               0.2.1                    py36_1    pytorch   
libopencv                 3.4.1                h1a3b859_1   
opencv                    3.4.1            py36h6fd60c2_2  
py-opencv                 3.4.1            py36h0676e08_1  
pyvips                    2.1.2                     <pip>
tqdm                      4.23.4                   py36_0

In addition, libvips (version 8.2.2-1) was used in this study for handling of whole slide images. The open-source python package pyvips is a wrapper for libvips ((https://jcupitt.github.io/libvips/), which can be installed in Linux with the following:

sudo apt-get install libvips

Hardware Requirements

Graphics Cards for Deep Learning - All deep learning models were trained using 4 X NVIDIA 1080 GPUs. As indicated above, PyTorch requires CUDA 8.0 and cuDNN 7.0 for compatibility.

Installation guide

We recommend creating a new Anaconda (https://www.anaconda.com/) environment with the dependencies above.

This repository can be cloned directly through:

git clone https://github.com/keiserlab/plaquebox-paper.git

Demo

Notebook 2.2) CNN Models - Test Cases is a demo that shows how to apply the trained CNN model on unseen dataset. Simply download the tiles from Zenodo repository and unzip it to the /data folder, then the notebook can be run through Jupyter.

Instructions for use

This repository contains 11 notebooks to reproduce the results from the linked paper. Each notebook includes details relevant to a portion of the described pipeline, with detailed descriptions at the top of each notebook. For results reproduction, these files are presented in sequential order and depend on the previous notebook.

Data Download

Before running the code, it is necessary to download the raw datafiles from the corresponding Zenodo repository above and unzip the files to the /data folder.

Modifying Filepaths

The filepaths must be specified as indicated in each notebook to specify the location of the downloaded data.

1. Preprocessing Steps

Notebooks 1.1-1.3 describe necessary preprocessing steps, including: color normalization, whole slide image tiling, blob detection, and dataset splitting.

2. Model Training and Development

Notebooks 2.1 and 2.2 detail model development, training, and testing.

3. Visualizing Predictions

Notebook 3 describes prediction confidence heatmaps.

4. Saliency Maps

Notebooks 4.1 and 4.2 describe feature interpretation studies, including feature occlusion and guided-grad cam studies.

5. CERAD-like Scoring on Whole Slide Images

Notebooks 5.1-5.3 describe whole slide scoring.

About

Repo for Tang et al, bioRxiv 454793 (2018)

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Jupyter Notebook 99.9%
  • Python 0.1%