Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Improve your understanding of qualitative instance segmentation results to gain contextual insights from your output data.

With the help of your qualitative results output, you can compare the predictions of your algorithm application with your manual (ground truth) annotations. This can help you better assess the predictive power of your algorithmtrained model.

What to expect from this page?

Table of Contents
minLevel1
maxLevel7
excludeWhat to expect from this page?

📑 Location of qualitative output visualizations

Qualitative results for all validation images and ROI(s) are provided in a ZIP-folder in the form of visualizations in separate image files in addition to the PDF report.

📑 Types of visualizations included

ROI(s) overview

If you have included ROI(s) in your validation images, an overview with a visualization of the ROI(s)' position(s) within the image will be provided.

You can find it in the folder rois_visualizations: [your-image-name]_rois_visualization.jpg.

instance-segmentation-view-rois-visualizations.mp4
Info

Please note:

  • Each ROI has an automatically generated [roi-id]according to the creation date of the ROI.

  • ROI(s) completely outside of the image are not shown in this file.

Image and ROI Performance

You will also find validation visualizations showing the performance of your trained algorithm model for a specific label [your-label-name] within the image or the ROI. They are located in the validation_vis folder. This folder also includes confidence visualizations for each instance.

instance-segmentation-validation-visualizations.mp4

The image and ROI files are named accordingly:

  • [your-label-name][your-image-name].jpg - visualization name of an image that contains no ROI(s)

  • [your-label-name][your-image-name]_[roi-id].jpg - visualization name of a ROI

📑 Understanding performance visualizations

Each of the validation visualizations is divided into four sections:

Info

Please note: that all visualizations are downscaled to 25 megapixels (MP), if the visualization in the original image size is larger than 25 MP.

Input image

(Left upper image)

Shows the input image.

Instance correctness, prediction [label-name]

(Right upper image)

Shows an overlay of predicted instances (instance contours) as predicted by the algorithmapplication, marked with a different color (e.g. as true-positives or as false-positives) depending on the presence of an overlap.

True-positive areas (overlap ≥ 50%) are shown in

Status
colourGreen
titleGREEN
.

Predicted instances of the algorithm Instances predicted by the application that match annotated instances with an overlap of ≥ 50%. In other words, these instances have been predicted correctly.

False-positive areas (overlap < 50%) are shown in

Status
colourRed
titleRED
.

Instances that have not been annotated, but have been predicted by your algorithmapplication.

Instance correctness, annotation [label-name]

(Left lower image)

Shows an overlay of ground-truth instances as drawn by the user, marked with a different color (e.g. as true-positives or as false-negatives) depending on the presence of an overlap.

True-positive areas (overlap ≥ 50%) are shown in

Status
colourGreen
titleGREEN
.

Annotated instances that match instances predicted by the algorithm application with an overlap of ≥ 50%. In other words, the algorithm app detects and labels these objects in the images correctly.

False-negative areas (overlap < 50%) are shown in

Status
colourBlue
titleBLUE
.

Instances that have been annotated, but not predicted by your algorithmapplication.

Pixel correctness [label-name]

(Right lower image)

Visualizes correctness cases resulting from the automated prediction (compared to manual annotations) shown on a pixel level.

True-positive areas (overlap ≥ 50%) are shown in

Status
colourGreen
titleGREEN
.

Areas where the prediction of the algorithm application matches annotations, represent the ground truth for the algorithmapplication. In other words, it labels areas of the image correctly.

False-positive areas (overlap < 50%) are shown in

Status
colourRed
titleRED
.

Areas that have not been annotated, however, were predicted by your algorithmapp. In other words, it labels parts of the image that should not be labeled.

False-negative areas (overlap < 50%) are shown in

Status
colourBlue
titleBLUE
.

Areas that have been annotated, however, were not predicted by your algorithmapplication.

Now that you have mastered the handling of qualitative results, you are perfectly equipped to gain solid insights from your analysis outputs.


If you still have questions regarding your algorithm application training, feel free to send us an email at support@ikosa.ai. Copy-paste your training ID in the subject line of your email.

📚 Related articles

Filter by label (Content by label)
showLabelsfalse
showSpacefalse
cqllabel = "instance-segmentation"