Skip to content

reproducing results with corrected dataset #5

@mrudorfer

Description

@mrudorfer

Hello,

Following up on my previous question regarding the offset in the dataset, we unfortunately still have problems reproducing the results in the paper, and we were wondering whether you have any further advice.
We compared the available pretrained model with models we trained using the code in this repository and the provided dataset. For the training dataset we used two conditions:

  1. "offset" which is the dataset as we downloaded it, and
  2. "fixed", where we removed the offsets from grasp centers and contact points.

Similarly, for the testing we used the same two conditions:

  1. "offset", where we compared with the test annotations as downloaded and performed simulations without any adjustments of the predicted grasps, and
  2. "fixed", where we used the corrected dataset for rule-based evaluation and we introduced the offset to the predicted grasps for simulation.

Our expectation was that using the code from the GPNet repository with fixed training dataset and fixed evaluation should be the correct way to reproduce the results, however, we also checked other combinations and here are all results:

gpnet_results_reproduce

As you can see, the version of GPNet that we trained with the fixed dataset performs very poorly in both evaluation conditions, which is surprising to us. We have very thoroughly checked that we fixed the dataset in the correct way and we are certain that grasp centers and contact points are correct now.

The pre-trained model and the one we trained in "offset" condition work ok and produce very similar results.
Although the rule-based success rates are more or less matched, neither of those models is able to reproduce the high simulation success rates reported in the paper. We think that the simulation-based score is the more important one, as it produces a label for the actually predicted grasps instead of simply looking up nearby grasps.

So to summarise, our questions are:

  1. Can you confirm that fixed-fixed condition should be the correct one? Is this also the one you used for producing the paper results and the pre-trained model?
  2. Is there any possible explanation why the model we trained with the fixed dataset performs so poorly? We just used the repository code with corrected dataset and are a bit clueless on this one...
  3. Should we expect to be able to reliably reproduce the high simulation success rates reported in the paper? Are we doing something wrong, as we do not reproduce these scores even with the pre-trained model? Or would that score maybe be rather some sort of upper bound on the achievable simulation success rate (as we experience that variance is relatively high)?

We would very much appreciate your thoughts on this. Thank you in advance!

Best wishes,
Martin

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions