The repository at hand is designed into two modular DNNs such that they can ultimately be use used as in the following figure
Model Part 1 is the name given to the model used to determine the likelihood maps asssociated with the probability of the absolute position of a grasp occuring. This prediction is given based off of an input 2-D image which was created via a hemispherical transform on a 3-D object mesh.
Visual results of such predictions using Model Part I are as depicted:
Model Part 2 is the meta-learned (using First-Order MAML) model that was tasks with updating the likelihood maps produced in Model Part 1 according to demonstrated human grasps. Model Part 2 also outputs the maximum liklihood angles (azimuth, zenith and our rotational angle gamma). As input Model Part 2 takes both the spherically transformed mesh image and the priors produced in Model part 1.
When training such a model with FOMAML , we found that the best form of task-augmentation was additive discrete noise towards the angular data (see below). In such cases, the base-learner was still capable of adapting its predictions. When attempting task-augmentations on the labelled likelihood maps, the FOMAML trained model was not flexible enough.
a) the Labelled data for b) the predicted dataFor data Generation used for these models, see my other repository: https://github.com/KryptixOne/Spherical-Data-Generation-For-3D-Meshes