@APJansen, @goord and myself conducted recently hyperoptimization experiments at multi-replica level using our local Snellius cluster.
With a limited max wall time (120h in gpu partition), it became clear that the increase in the number of replicas (and consequently increase in the computation time per trial) would result in a relatively poor sample of the hyperparameter space per calculation.
It would be very nice if we could restart the hyperoptimization process using info from a previous tries.json file and input seeds.
Not sure whether this option might be easily adapted or considerable changes must be made in ModelTrainer.hyperparametrizable.
I could assign myself for the task but would appreciate any suggestion and thoughts about this issue.
@APJansen, @goord and myself conducted recently hyperoptimization experiments at multi-replica level using our local Snellius cluster.
With a limited max wall time (120h in gpu partition), it became clear that the increase in the number of replicas (and consequently increase in the computation time per trial) would result in a relatively poor sample of the hyperparameter space per calculation.
It would be very nice if we could restart the hyperoptimization process using info from a previous
tries.jsonfile and input seeds.Not sure whether this option might be easily adapted or considerable changes must be made in
ModelTrainer.hyperparametrizable.I could assign myself for the task but would appreciate any suggestion and thoughts about this issue.