Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 6 Current »

Application Name

IKOSA AI Semantic Segmentation App

Version

4.1.0

Documentation Version

02.03.2022 - 1

Input Image(s)

2D (standard and/or WSI) / Time Series / z-Stack / Multichannel Images;

RGB or Grayscale (8 bit)

Input Parameter(s)

Regions of interest (optional)

Keywords

-

Short Description

Segmentation of regions/objects that resemble the appearance of regions/objects annotated and labeled in the training data for this application.

References / Literature

Table of contents

IKOSA AI Semantic Segmentation App

Your IKOSA AI application was created using our training for semantic segmentation models. All information presented in this document is applicable to your trained application.

Application description

This application automatically segments the regions/objects of the images that resemble the appearance of regions/objects that were annotated and labeled in the training data of this application. The application supports multiple classes and also performs an instance separation in post-processing. Areas of the different classes are measured, and the number and morphological parameters of the objects belonging to the different classes are calculated. This analysis can also be performed on timelapse recordings (Time Series), z-stacks, or multichannel images, uploaded as 8-bit multipage TIFF files.

In the following, the requirements for an accurate analysis are given and the output of the application is described.

Input data requirements

Input image(s)

Input for this application is the following image data:

Image type

Color channels

Color depth (per channel)

Size (px)

Resolution (μm/px)

2D (standard and/or WSI),

Time Series,

Multichannel,

Z-stack*

Check image format

File formats

3 (RGB)

1 (Grayscale)

8 Bit

WSI formats: arbitrary

Standard images: max. 25,000 x 25,000

arbitrary

Image content

Arbitrary

Additional requirements

None

*Please note: Z-stack images cannot be uploaded into IKOSA but still can be analyzed via IKOSA Prisma API.

Important:

For all images, the following requirements apply:

  • The illumination must be constant throughout the image(s).

  • The sample must be in focus, i.e. no blurry regions in image(s).

Input parameter(s)

No additional input parameters are required for this application.

As an optional parameter, a single or multiple regions of interest (ROIs) can be defined in which the analysis should be performed (‘inclusion ROIs’).

Description of output files and their content

Files

File format

Description

1

csv

results.csv

A csv file containing the overall analysis results for the input image or all inclusion ROIs.

2

csv

results_<xx>_<class-name>.csv

A csv file containing the analysis results for all detected objects of class number <xx> with label name <class-name> (in training data) in the input image or inclusion ROIs.

3

jpg

results_vis/<xx>_<class-name>_vis.jpg (2D image, no ROI), or

results_vis/<xx>_<class-name>_t<time-step>_z<z-layer>_c<channel>.jpg (for time series, z-stack, or multichannel image, no ROI), or

results_vis/<xx>_<class-name>_<roi-id>.jpg (2D image, ROI <roi-id>), or

results_vis/<xx>_<class-name>_t<time-step>_z<z-layer>_c<channel>_<roi-id>.jpg for time series, z-stack, or multichannel image, ROI <roi-id>):

A visualization of the analysis result for a specific time step (of a time series), z-layer (of a z-stack), or channel (of a multichannel image) for either the whole image (if no inclusion ROIs selected for analysis) or each individual inclusion ROI, for each class number <xx> with label name <class-name> (in training data).

Each visualization includes two parts:

  • segmentation

    • predicted regions/objects areas are shown as an overlay in green color.

  • objects

    • object instances are shown as an overlay in random color.

    • object instance index number corresponds to the object_id as listed in the results_<xx>_<class-name>.csv file.

4

json

roiMeta.json

A json file containing all information regarding the ROIs defined for the analysis job to ensure reproducibility. The file is empty if no ROIs were defined for analysis.

5

jpg

rois_visualization.jpg, or

t<time-step>_z<z-layer>_c<channel>_rois_visualization.jpg:

An overview visualization to show locations of all analyzed ROIs for the 2D image or time step <time-step> of a time series, z-layer <z-layer> of a z-stack, or channel <channel> of a multichannel image.

Please note: This file is only created if inclusion ROIs were defined for analysis.

6

json

jobResultBundleMeta.json:

A json file containing all information regarding the analysis job (application name and version, project, etc.) to ensure reproducibility.

Please note: This file is only included if bundled or merged analysis jobs are downloaded.

Please note:

  • In the case of  inclusion ROIs that are partially outside of the image, the ROIs are cropped to the areas that lie inside the image.

  • In the case of inclusion ROIs that are completely outside of the image, no analysis is performed. However, they are still listed in corresponding results files.

  • A <roi-id> is generated automatically by the application corresponding to the creation date of a ROI. The location of a ROI within an image with its specific <roi-id> can be seen in the file “rois_visualization.jpg.” ROIs that are completely outside of the image are not shown in this file.

  • All visualizations are downscaled to 25 megapixels (MP) if the original image or inclusion ROI is larger than 25 MP.

Content

results.csv

Single csv-file

If one or more time steps (of a Time Series), z-layers (of a z-Stack), or channels (of a multichannel image) were specified, the results in a specific row refer to the time step/z-layer/channel specified in the corresponding column.

If one or more ROIs were specified, the results in a specific row refer to the ROI specified in the corresponding columns, otherwise (empty ROI columns) the results refer to the whole image.

Column NO.

Column name

Examples

Value range

Description

1

t

3

1 - 

Time step, i.e. the position of the image in the time series.

2

z

5

1 - 

z-layer, i.e. the position of the layer in the z-stack.

3

c

2

1 - 

Channel, i.e. the position of the channel in the multichannel image.

4

roi_id

ROI-03

ROI-01 - 

<roi-id> starting from “ROI1”. Empty, if no inclusion ROI is specified and the whole image was analyzed.

5

roi_name

“central”

text

Custom text to identify the ROI. Empty, if no inclusion ROI is specified and the whole image was analyzed.

6

roi_size [Px^2]

1212212

1 -

Size of the ROI that was analyzed in pixels^2. The size of the whole image is given if no inclusion ROI is specified and the whole image was analyzed.

7

<class-name>_total_area [Px^2]

122438

0 - no. of image px

Total area covered by detected objects of class <class-name> in Pixels^2.

8

<class-name>_total_area [%]

3.66

0 - 100

Total area covered by detected objects of class <class-name> as percentage of overall  image area or ROI area inside the image.

9

<class-name>_total_num_of_objects

3796

0 - 

Total number of detected objects of class <class-name> in ROI or image.

...

Similar to columns 6-8 with further classes.

results_<xx>_<class-name>.csv

Single or multiple csv-file(s)

If one or more time steps (of a Time Series), z-layers (of a z-Stack), or channels (of a multichannel image) were specified, the results in a specific row refer to the time step/z-layer/channel specified in the corresponding column.

If one or more ROIs were specified, the results in a specific row refer to the ROI specified in the corresponding columns, otherwise (empty ROI columns) the results refer to the whole image.

Column NO.

Column name

Examples

Value range

Description

1

t

3

1 - 

Time step, i.e. the position of the image in the time series.

2

z

5

1 - 

z-layer, i.e. the position of the layer in the z-stack.

3

c

2

1 - 

Channel, i.e. the position of the channel in the multichannel image.

4

roi_id

ROI-03

ROI-01 - 

<roi-id> starting from “ROI1”. Empty, if no inclusion ROI is specified and the whole image was analyzed.

5

roi_name

“central”

text

Custom text to identify the ROI. Empty if no inclusion ROI is specified and the whole image was analyzed.

6

roi_size [Px^2]

1212212

1 -

Size of the ROI that was analyzed in pixels^2. The size of the whole image is given if no inclusion ROI is specified and the whole image was analyzed.

7

object_id

5

1 - 

ID of object corresponding to id in visualization of ROI or image

8

area [Px^2]

132

1 -

Area of detected object in Pixels^2.

9

bbox_area [Px^2]

175

1 -

Area of bounding box of detected object in Pixels^2.

10

area_ratio [%]

0.4

0 - 100

Area of detected object as percentage of overall  image area or ROI area inside the image.

11

perimeter [Px]

12.3

0 -

Perimeter of detected object in Pixels.

12

circularity

0.91

0 -

Circularity factor of detected object; circularity = 4*pi*area/(perimeter^2). The circularity of a circle is 1.

13*

solidity

0.98

0 - 1

Ratio of pixels in the region to pixels of the convex hull image.

14*

eccentricity

0.96

0 - 1

Eccentricity of the ellipse that has the same second-moments as the region. The eccentricity is the ratio of the focal distance (distance between focal points) over the major axis length. When it is 0, the ellipse becomes a circle.

15*

equivalent_diameter [Px]

10.3

0 -

Diameter of a circle having the same area as the detected object.

16*

extent

0.73

0 - 1

Ratio of pixels in the region to pixels in the total bounding box

17*

minor_axis_length [Px]

6

1 -

The length of the minor axis of the ellipse that has the same normalized second central moments as the region.

18*

major_axis_length [Px]

12

1 -

The length of the major axis of the ellipse that has the same normalized second central moments as the region.

Please note: The parameters marked with an asterisk (*) are calculated using https://scikit-image.org/.

Error information

More information about errors can be found in the Application Error Documentation.

Contact

If you have any questions about IKOSA AI and the applications you can create with it, please refer to Application Training with IKOSA AI section in the Knowledge Base.

You can also always contact our team at support@kmlvision.com for any further clarifications.

Feel free to book a 30-minute meeting to speak with us about IKOSA and the apps!

https://calendly.com/kmlvision/book-the-ikosa-platform-demo

  • No labels