4.4.1 Semantic Segmentation App
Application Name | IKOSA AI Semantic Segmentation App |
Version | 4.4.1 |
Documentation Version | 28.03.2023 - 1 |
Input Image(s) | 2D (standard and/or WSI) / Time Series / z-Stack / Multichannel Images; RGB or Grayscale (8 bit) |
Input Parameter(s) | Regions of interest (optional) |
Keywords | - |
Short Description | Segmentation of regions/objects that resemble the appearance of regions/objects annotated and labeled in the training data for this application. |
References / Literature |
|
Table of contents
IKOSA AI Semantic Segmentation App
Your IKOSA AI application was created using our training for semantic segmentation. All information presented in this document is applicable to your trained application.
Application description
This application automatically segments the regions/objects of the images that resemble the appearance of regions/objects that were annotated and labeled in the training data of this application. The application supports multiple classes and also performs an instance separation in post-processing. Areas of the different classes are measured, and the number and morphological parameters of the objects belonging to the different classes are calculated. This analysis can also be performed on time-lapse recordings (Time Series), z-stacks, or multichannel images, uploaded as 8-bit multipage TIFF files.
In the following, the prerequisites for an accurate analysis are outlined and the output of the application is described.
Input data requirements
Input image(s)
Input for this application is the following image data:
Image type | Color channels | Color depth (per channel) | Size (px) | Resolution (μm/px) |
---|---|---|---|---|
2D (standard and/or WSI), Time Series, Multichannel, Z-stack*
Check image format | 3 (RGB) 1 (Grayscale) | 8 Bit
| WSI formats: arbitrary Standard images: max. 25,000 x 25,000 | arbitrary |
Image content Arbitrary Additional requirements None |
*Please note: Z-stack images cannot be uploaded into IKOSA but still can be analyzed via IKOSA Prisma API.
Important:
For all images, the following requirements apply:
The illumination must be constant throughout the image(s).
The sample must be in focus, i.e. no blurry regions in image(s).
Input parameter(s)
No additional input parameters are required for this application.
As an optional parameter, a single or multiple regions of interest (ROIs) can be defined in which the analysis should be performed (‘inclusion ROIs’).
Please note: Parameters that were set during training may affect also prediction with the deployed application. More information can be found under How to set custom training parameter values?.
Description of output files and their content
Files
File format | Description | |
---|---|---|
1 | csv | results.csv A csv file containing the overall analysis results for the input image or all inclusion ROIs. |
2 | csv | results_<xx>_<class-name>.csv A csv file containing the analysis results for all detected objects of class number <xx> with label name <class-name> (in training data) in the input image or inclusion ROIs. |
3 | jpg | results_vis/<xx>_<class-name>_vis.jpg (2D image, no ROI), or results_vis/<xx>_<class-name>_t<time-step>_z<z-layer>_c<channel>.jpg (for time series, z-stack, or multichannel image, no ROI), or results_vis/<xx>_<class-name>_<roi-id>.jpg (2D image, ROI <roi-id>), or results_vis/<xx>_<class-name>_t<time-step>_z<z-layer>_c<channel>_<roi-id>.jpg for time series, z-stack, or multichannel image, ROI <roi-id>): A visualization of the analysis result for a specific time step (of a time series), z-layer (of a z-stack), or channel (of a multichannel image) for either the whole image (if no inclusion ROIs selected for analysis) or each individual inclusion ROI, for each class number <xx> with label name <class-name> (in training data). Each visualization includes two parts:
|
4 | json | annotation_results.json: JSON file containing segmented regions. The position is measured from the left upper corner (1,1) of the image. |
5 | json | roiMeta.json A json file containing all information regarding the ROIs defined for the analysis job to ensure reproducibility. The file is empty if no ROIs were defined for analysis. |
6 | jpg | rois_visualization.jpg, or t<time-step>_z<z-layer>_c<channel>_rois_visualization.jpg: An overview visualization to show locations of all analyzed ROIs for the 2D image or time step <time-step> of a time series, z-layer <z-layer> of a z-stack, or channel <channel> of a multichannel image. |
7 | json | jobResultBundleMeta.json: A json file containing all information regarding the analysis job (application name and version, project, etc.) to ensure reproducibility. |
Content
results.csv
Single csv-file
If one or more time steps (of a Time Series), z-layers (of a z-Stack), or channels (of a multichannel image) were specified, the results in a specific row refer to the time step/z-layer/channel specified in the corresponding column.
If one or more ROIs were specified, the results in a specific row refer to the ROI specified in the corresponding columns, otherwise (empty ROI columns) the results refer to the whole image.
Column NO. | Column name | Examples | Value range | Description |
---|---|---|---|---|
1 | t | 3 | 1 - | Time step, i.e. the position of the image in the time series. |
2 | z | 5 | 1 - | z-layer, i.e. the position of the layer in the z-stack. |
3 | c | 2 | 1 - | Channel, i.e. the position of the channel in the multichannel image. |
4 | roi_id | ROI-03 | ROI-01 - | <roi-id> starting from “ROI1”. Empty, if no inclusion ROI is specified and the whole image was analyzed. |
5 | roi_name | “central” | text | Custom text to identify the ROI. Empty, if no inclusion ROI is specified and the whole image was analyzed. |
6 | roi_size [Px^2] | 1212212 | 1 - | Size of the ROI that was analyzed in pixels^2. The size of the whole image is given if no inclusion ROI is specified and the whole image was analyzed. |
7 | <class-name>_total_area [Px^2] | 122438 | 0 - no. of image px | Total area covered by detected objects of class <class-name> in Pixels^2. |
8 | <class-name>_total_area [%] | 3.66 | 0 - 100 | Total area covered by detected objects of class <class-name> as percentage of overall image area or ROI area inside the image. |
9 | <class-name>_total_num_of_objects | 3796 | 0 - | Total number of detected objects of class <class-name> in ROI or image. |
... |
|
|
| Similar to columns 6-8 with further classes. |
results_<xx>_<class-name>.csv
Single or multiple csv-file(s)
If one or more time steps (of a Time Series), z-layers (of a z-Stack), or channels (of a multichannel image) were specified, the results in a specific row refer to the time step/z-layer/channel specified in the corresponding column.
If one or more ROIs were specified, the results in a specific row refer to the ROI specified in the corresponding columns, otherwise (empty ROI columns) the results refer to the whole image.
Column NO. | Column name | Examples | Value range | Description |
---|---|---|---|---|
1 | t | 3 | 1 - | Time step, i.e. the position of the image in the time series. |
2 | z | 5 | 1 - | z-layer, i.e. the position of the layer in the z-stack. |
3 | c | 2 | 1 - | Channel, i.e. the position of the channel in the multichannel image. |
4 | roi_id | ROI-03 | ROI-01 - | <roi-id> starting from “ROI1”. Empty, if no inclusion ROI is specified and the whole image was analyzed. |
5 | roi_name | “central” | text | Custom text to identify the ROI. Empty if no inclusion ROI is specified and the whole image was analyzed. |
6 | roi_size [Px^2] | 1212212 | 1 - | Size of the ROI that was analyzed in pixels^2. The size of the whole image is given if no inclusion ROI is specified and the whole image was analyzed. |
7 | object_id | 5 | 1 - | ID of object corresponding to id in visualization of ROI or image |
8 | area [Px^2] | 132 | 1 - | Area of detected object in Pixels^2. |
9 | bbox_area [Px^2] | 175 | 1 - | Area of bounding box of detected object in Pixels^2. |
10 | area_ratio [%] | 0.4 | 0 - 100 | Area of detected object as percentage of overall image area or ROI area inside the image. |
11* | perimeter [Px] | 12.3 | 0 - | Perimeter of detected object in Pixels. |
12* | circularity | 0.91 | 0 - | Circularity factor of detected object; circularity = 4*pi*area/(perimeter^2). The circularity of a circle is 1. |
13* | solidity | 0.98 | 0 - 1 | Ratio of pixels in the region to pixels of the convex hull image. |
14* | eccentricity | 0.96 | 0 - 1 | Eccentricity of the ellipse that has the same second-moments as the region. The eccentricity is the ratio of the focal distance (distance between focal points) over the major axis length. When it is 0, the ellipse becomes a circle. |
15* | equivalent_diameter [Px] | 10.3 | 0 - | Diameter of a circle having the same area as the detected object. |
16* | extent | 0.73 | 0 - 1 | Ratio of pixels in the region to pixels in the total bounding box |
17* | minor_axis_length [Px] | 6 | 1 - | The length of the minor axis of the ellipse that has the same normalized second central moments as the region. |
18* | major_axis_length [Px] | 12 | 1 - | The length of the major axis of the ellipse that has the same normalized second central moments as the region. |
Error information
More information about errors can be found in the Application Error Documentation.
Contact
If you have any questions about IKOSA AI and the applications you can create with it, please refer to Application Training with IKOSA AI section in the Knowledge Base.
You can also always contact our team at support@ikosa.ai for any further clarifications.
Feel free to book a 30-minute meeting to speak with us about IKOSA and the apps!
Copyright 2024 KOLAIDO GmbH. IKOSA® is a registered EU trademark.