I had an idea while trying the new experiment feature. For the use case of grid-search or hyper-parameter optimization, could we imagine an option for dvc experiments checkout to checkout the best experiment regarding one of the metrics? Something like dvc experiments checkout --best metrics.json:accuracy. With an option like this, you can run (manually or automatically) a lot of experiments and ask dvc to return the best experiment, which can be then commited to git. Or instead of --best, it could be --lowest or --highest to take into account that some metrics should be minimized and some others maximized.
I had an idea while trying the new experiment feature. For the use case of grid-search or hyper-parameter optimization, could we imagine an option for
dvc experiments checkoutto checkout the best experiment regarding one of the metrics? Something likedvc experiments checkout --best metrics.json:accuracy. With an option like this, you can run (manually or automatically) a lot of experiments and ask dvc to return the best experiment, which can be then commited to git. Or instead of--best, it could be--lowestor--highestto take into account that some metrics should be minimized and some others maximized.