A details list of all metrics currently supported by Panoptica.
- DICE (DSC)
- Intersection over Union (IoU)
- Centerline Dice (clDSC)
- Average Symmetric Surface Distance (ASSD)
- Relative Volume Difference (RVD)
For instance-wise metrics:
- True Positives (tp)
- False Positives (fp)
- False Negatives (fn)
And most importantly, the panoptic metrics:
- Recognition Quality (rq)
- Segmentation Quality (sq)
- Panoptic Quality (pq)
These three original come from arXiv: Panoptic Segmentation.
Missing a metric? Write us an github: issue
Be sure to use the "metric" label.
Formulas to calculate the different metrics
$$ \text{DSC}(X,Y) = \frac{2 | X \cap Y | }{ |X| + |Y|} \in [0,1] $$ Represents an overlap score.
$$ \text{IoU}(X,Y) = \frac{| X \cap Y | }{ |X \cup Y|} \in [0,1] $$ Represents an overlap score. Is related to DSC, so giving both metrics doesn't really make sense (you can compute one with the other).
Originally from arXiv: clDice
Represents an topology-preserving overlap score. Can be used as loss. Uses skeletonize and calculates the Dice coefficient on the skeletonized version of prediction and reference.
$$
\text{ASSD}(X,Y) = \frac{\text{asd}(A,B) + \text{asd}(B,A)}{ |X| + |Y|} \in [0,\infty]
$$
with
ASSD is a typically good metric to report, as it shows whether errors are local or if the prediction has widespread noise voxels not even close to the reference.
The relative volume difference is the predicted volume of an instance in relation to the reference volume. For a journal, this might not be the most important metric. However, when the RVD is consistently positive, the predictions are oversegmenting. If negative, it is underpredicting (overall).
It is the F1-score basically. Represents how well your instances match the references well (well = determined by threshold).
The segmentation quality is the average of all true positive metrics. So for sq_dsc, this is the average of all dice scores among the true positives in this prediction/reference pair.
As this metric is linked to another metric, we support all combinations of it. You can calculate the segmentation quality with IoU, DSC, clDSC, ASSD, and even RVD.
Combines the F1-score of instances with the Segmentation Quality.