Background, dataloader slows down over time, especially when using a large number of slides; data that is persistent in memory loads quickly (case for very small number sslides), but not when training from large number of slides; issues with having .compute() within getitem(), yet needing to take into account data augmentations (albumentations) for the mask of the image for semantic segmentation task when loading data, which can make the dataloading operation if more daskified a bit more complex:
Issue is with the getitem, when the data is loaded, it passes quickly through the DL model.
Potentially nice ideas:
@lvaickus , can you comment more here?
@sumanthratna
Background, dataloader slows down over time, especially when using a large number of slides; data that is persistent in memory loads quickly (case for very small number sslides), but not when training from large number of slides; issues with having .compute() within getitem(), yet needing to take into account data augmentations (albumentations) for the mask of the image for semantic segmentation task when loading data, which can make the dataloading operation if more daskified a bit more complex:
Issue is with the getitem, when the data is loaded, it passes quickly through the DL model.
Potentially nice ideas:
@lvaickus , can you comment more here?
@sumanthratna