You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
🐛 Dataset Statistics do not work for sourcefolder dataset
The get_statistics function was designed to iterate over the complete audio data in an deterministic manner, therefore loading the full audio samples. This doesn't work together with the sourcefolder dataset as it allows to have different length of files as it get short chunks of fixed lengths from each item.
Expected behavior
sourcefolder dataset should work with get_statistics
Proposed solutions
Solution 1
replace dataset_scaler.seq_duration = None with dataset_scaler.seq_duration = args.seq_dur. That would solve the issue but then would only train the dataset statistics on the first n seconds from each sample.
Solution 2
use stochastic sampling and use a dataloader instead of a dataset: e.g.:
the second option would get better distributed samples and users can maybe specify an argument that selects the number of samples randomly drawn to train the dataset statistics
The text was updated successfully, but these errors were encountered:
Then we extract random chunks (random start positions) and it would still be deterministic because we use the dataset in valid mode.
that would work, but then we will use the validation set instead of the training set. Maybe we introduce a deterministic flag for all dataloaders instead of making this depend on valid vs. train?
🐛 Dataset Statistics do not work for
sourcefolder
datasetThe
get_statistics
function was designed to iterate over the complete audio data in an deterministic manner, therefore loading the full audio samples. This doesn't work together with the sourcefolder dataset as it allows to have different length of files as it get short chunks of fixed lengths from each item.Expected behavior
sourcefolder
dataset should work with get_statisticsProposed solutions
Solution 1
replace
dataset_scaler.seq_duration = None
withdataset_scaler.seq_duration = args.seq_dur
. That would solve the issue but then would only train the dataset statistics on the first n seconds from each sample.Solution 2
use stochastic sampling and use a dataloader instead of a dataset: e.g.:
the second option would get better distributed samples and users can maybe specify an argument that selects the number of samples randomly drawn to train the dataset statistics
The text was updated successfully, but these errors were encountered: