You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Provide an Implementation of the AutoAugment policies for ImageNet, CIFA10 and SVHN on TorchVision.
Motivation
AutoAugment is a common Data Augmentation technique that when applied, typically increases the accuracy of the models on specific datasets. On the original paper, the authors used RL to search for better data augmentation policies that increase the accuracy on 3 datasets: ImageNet, CIFA10 and SVHN. Though the found data augmentation policies were directly linked to their trained dataset, the paper shows that ImageNet policies provide significant improvements when applied to other datasets such as Oxford Flowers, Caltech-101, Oxford-IIT Pets, FGVC Aircraft and Stanford Cars. Similarly the CIFAR10 policies can be used for boosting the performance of CIFAR100.
Hence providing an out-of-the-box implementation for the aforementioned policies can be beneficial for the users of TorchVision.
Pitch
Provide the implementation of 3 the following policies:
ImageNet AutoAugment Policy
CIFAR10 AutoAugment Policy
SVHN AutoAugment Policy
Currently there are already community implementations for the above, nevertheless they all use PIL as backend. Therefor to add this in TorchVision we need to add support for the following missing Tensor Functional transforms:
posterize
solarize
equalize
sharpness
autocontrast
invert
Note that the original paper also experimented with the following augmentation techniques:
Cutout: Very similar to RandomErasing but with does not apply thresholds on rations/scales, can produce multiple rectangles, uses the value of the average pixel per channel of the dataset. TorchVision already provides an implementation for RandomErasing which closely approximates Cutout, so it's probably not worth reimplementing it. Moreover though cutout was investigated by AutoAugment's RL as a policy, it was not selected for any of the datasets, though it was applied always at the end on the CIFAR10 experiments.
Sample Pairing: Similar to the above, though Sample Pairing was investigated during the RL phase, it was not selected by any of the dataset policies and thus we don't need to implement it here.
Lastly it's important to note that it's worth supporting the aforementioned transforms even if we decide not to implement AutoAugment. This is because it will enable the easier addition of newer Data Augmentation techniques such as FixMatch (official implementation, @vfdev-5's implementation), ReMixMatch, RandAugment etc.
🚀 Feature
Provide an Implementation of the AutoAugment policies for ImageNet, CIFA10 and SVHN on TorchVision.
Motivation
AutoAugment is a common Data Augmentation technique that when applied, typically increases the accuracy of the models on specific datasets. On the original paper, the authors used RL to search for better data augmentation policies that increase the accuracy on 3 datasets: ImageNet, CIFA10 and SVHN. Though the found data augmentation policies were directly linked to their trained dataset, the paper shows that ImageNet policies provide significant improvements when applied to other datasets such as Oxford Flowers, Caltech-101, Oxford-IIT Pets, FGVC Aircraft and Stanford Cars. Similarly the CIFAR10 policies can be used for boosting the performance of CIFAR100.
Hence providing an out-of-the-box implementation for the aforementioned policies can be beneficial for the users of TorchVision.
Pitch
Provide the implementation of 3 the following policies:
Currently there are already community implementations for the above, nevertheless they all use PIL as backend. Therefor to add this in TorchVision we need to add support for the following missing Tensor Functional transforms:
Note that the original paper also experimented with the following augmentation techniques:
Lastly it's important to note that it's worth supporting the aforementioned transforms even if we decide not to implement AutoAugment. This is because it will enable the easier addition of newer Data Augmentation techniques such as FixMatch (official implementation, @vfdev-5's implementation), ReMixMatch, RandAugment etc.
cc @vfdev-5 @fmassa
The text was updated successfully, but these errors were encountered: