Skip to content

I2Cvb/alsaih-2016-aug

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 

Repository files navigation

Classification of SD-OCT volumes with multi pyramids, LBP and HOG descriptors: application to DME detections

How to use the pipeline?

Pre-processing pipeline

The follwoing pre-processing routines were applied:

  • BM3D denoising,
  • Flattening,
  • Cropping.

Data variables

In the file pipeline/feature-preprocessing/pipeline_preprocessing.m, you need to set the following variables:

  • data_directory: this directory contains the orignal SD-OCT volume. The format used was .img.
  • store_directory: this directory corresponds to the place where the resulting data will be stored. The format used was .mat.

Algorithm variables

The variables which are not indicated in the inital publication and that can be changed are:

  • x_size, y_size, z_size: the original size of the SD-OCT volume. It is needed to open .img file.
  • sigma: the estimate of the standard deviation for the BM3D denoising.
  • h_over_rpe, h_under_rpe, width_crop: the different variables driving the cropping.

Run the pipeline

From the root directory, launch MATLAB and run:

>> run pipeline/feature-preprocessing/pipeline_preprocessing.m

Extraction pipeline

For this pipeline, the following features were extracted:

  • HOG,
  • LBP with radius set to 8, 16, and 24. Both rotation and non-rotation invariant features were extracted.

Data variables

In the file pipeline/feature-extraction/pipeline_extraction_***.m, you need to set the following variables:

  • data_directory: this directory contains the pre-processed SD-OCT volume. The format used was .mat.
  • store_directory: this directory corresponds to the place where the resulting data will be stored. The format used was .mat.

Run the pipeline

From the root directory, launch MATLAB and run:

>> run pipeline/feature-extraction/pipeline_extraction_hog.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_8_ri.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_16_ri.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_24_ri.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_8_nri.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_16_nri.m
>> run pipeline/feature-extraction/pipeline_extraction_lbp_24_nri.m

Classification pipeline

The method for classification used was:

  • Linear SVM.

Data variables

In the file pipeline/feature-preprocessing/pipeline_classifier_***.m, you need to set the following variables:

  • data_directory: this directory contains the feature extracted from the SD-OCT volumes. The format used was .mat.
  • store_directory: this directory corresponds to the place where the resulting data will be stored. The format used was .mat.
  • gt_file: this is the file containing the label for each volume. You will have to make your own strategy.

Run the pipeline

From the root directory, launch MATLAB and run:

>> run pipeline/feature-classification/pipeline_classifier_pca_hog.m
>> run pipeline/feature-classification/pipeline_classifier_pca_hog_lbp_8_16_24_ri.m
>> run pipeline/feature-classification/pipeline_classifier_pca_hog_lbp_8_16_24_nri.m

Validation pipeline

Data variables

In the file pipeline/feature-validation/pipeline_validation_***.m, you need to set the following variables:

  • data_directory: this directory contains the classification results. The format used was .mat.
  • gt_file: this is the file containing the label for each volume. You will have to make your own strategy.

Run the pipeline

From the root directory, launch MATLAB and run:

>> run pipeline/feature-validation/pipeline_validation_pca_hog.m
>> run pipeline/feature-validation/pipeline_validation_pca_hog_lbp_8_16_14_ri.m
>> run pipeline/feature-validation/pipeline_validation_pca_hog_lbp_8_16_14_nri.m