Preprocessing is currently done on CPU which is quite annoying as we need to transfer data between host and device all the time during training of our neural networks. This issue could be solved by at least providing means for tri-linear interpolation on GPU. We might make use of PyTorch's interpolation code: torch.lerp