AI 2D Fct

 

In this mode, you can use an AI to predict numerical values from 2D images.

 

The mode work with 2D images and 1D CSV (Comma Separated Values) files at the same time.  The numerical values are in the CSV files.

 

 

Note:

 

This mode work with the Python language.  So, you need to install Python on your system.  For help on that subject, please consult the page: www.TomoVision.com/support/sup_sliceO_python.html

 

 

From the Graphic Interface

 

 

 

 

The main "AI with Python" interface has 3 tabs: Train, Test and Predict.  Each of theses has a "Config" button that will open the configuration menu, and one (or more) "Compute" button that will open the appropriate page.

 

 

The Configuration Menu

 

Each of the 3 main pages of the "AI Python" interface has a "Config" button.  This will open the configuration menu.  Depending on the main pages you are in, the configuration menu will have between 3 and 9 sub-pages.  For training, you need access to all the pages, for "Predict", you only need 3.

 

 

The config "File 2D" page

 

 

This page is used to specify the images that will be used either for training or testing your AI. 

 

 

 

Progress bar

Show the progress of the program when parsing directories

 

Add slices

Files can be added to this list in 3 different ways:

 

Simply drag&drop a directory on the interface.  The program will recursively parse this directory and all its sub-directories.  Each directories that contain segmented images will be add as a separate line in the interface or use the "From Path" or "From Browser" options.

 

From Path

You can achieve the same results by specifying the directory in the input line and clicking the "Add from path" button.

 

From Browser

You can add files with the "Medi Browser" interface.

 

Clear List

Clear the list.

 

Dir List

For each directory specified that contain segmented files, you will see a line of information specifying the number of studies in that directory, the number of series, the modality of the images in the directory, the resolution of these images, the total number of images and the total number of segmented images, the number of TAGs present in these segmented images and a list of these TAGs.

 

Note:

Please note that the TAG information is not needed in this module.  TAG files are still read but that's just because I use the same function in this module as in the AI 2D TAG module.

 

You can also select and de-select any of the directories by clicking on the associated button in front of each line.

 

A double click on a line in this list will open the associated files in sliceOmatic.

 

Files preview

 

A preview of all the selected files is available.  Clicking an image will select it and highlight the dir it came from in the directory list.  A double click on an image will open that image in sliceOmatic.

 

Files status

Show the number of images found in the different selected directories.  Also, the different TAGs found in these directories (along with their prevalence) will be displayed.

 

Note:

Please note that the TAG information is not needed in this module.  TAG files are still read but that's just because I use the same function in this module as in the AI 2D TAG module.

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "File 1D" page

 

 

This page is used to specify the datasets that will be used either for training or testing your AI. 

 

The data must come from CSV (Comma Separated Values) files.  These files contain tables of data with a 1 to 3 lines header.  SliceOmatic expect these files to have either the ".csv" or ".txt" extensions.  The files can be in ASCII or one of 3 UTF encoding (8, 16 or 32 bits).  If the files start with the text "sep=", then the next character is used as cell separator, otherwise the program will guess the separator by analyzing the files. (accepted values are ",", "tab" or "space").

 

If the files contain 3 lines of header (files created by sliceO do), they will be used to create the header and its hierarchy.  If only one line is used, each column header will be split in its components and analyzed to create the hierarchy.

 

The header and data will be presented using colors:

 

 

Normal data is presented on a grey background.

 

Empty cells are presented in yellow.

 

Cells that contain an invalid (Not a Number) value have a red background.

 

Cells that contain multiple values have a red background.

 

Cells that contain text have a brown background.

 

Index columns have a cyan background.

 

Data columns (in AI modes) have a green background.

 

Class columns (in AI modes) have a blue background.

 

Also, if all the data associated with a column are either empty, "NaN", "Multiple" or have all the same value, then that column can not be used by the AI and the background of the associated 3rd line in the header box will be red.

 

 

 

Progress bar

Show the progress of the program when analyzing directories

 

Add cells

Files can be added to this list in 3 different ways:

 

Simply drag&drop a CSV file on the interface or use the "From Path" or "From Browser" options.

 

From Path

You can achieve the same results as the drag&drop by specifying the file path in the input line and clicking the "Add from path" button.  You can use the folder button on the path line to open a Windows file browser.  The selected file's path will then be copies to the input line.

 

From Browser

You can add files with the "CSV Browser" interface.

 

Clear List

Clear the list.

 

Header display

The 3 lines of column headers will be displayed in this box.  You can use the horizontal slider to scroll the headers.  You can also use the mouse wheel if the cursor is over the header box.

 

You can "mark" these columns to be either "Index" or "Class".

 

The "Index" can be used to help merge files in the database.  For example is one file contain the study ID and the Radiomics values derived from these studies, while another file contain the study ID and the matching patient information, specifying the "study ID" as index will ensure that all the information from both files are present in the same rows for each study.

 

Also, the "Index" will be used to macth the rows to the loaded images.  More on that in the "PreProc" page...

 

The "Class" column is the one the AI will be trained to predict.

 

Note:

If you use a "Categorical" AI model, you can only specify one class column.  For "Float" models, you can have multiple classes at the same time.

 

Data display

 

You can see all the data cells in this box.  You can use the vertical slider to scroll the rows.  You can also use the mouse wheel if the cursor is over the data box.

 

Status

The number of columns of the different types along with the number of valid rows available to the AI are displayed here.  If some columns or rows are invalid they will also be listed. 

 

Columns can be rejected if the associated values contains a text cell, or if all the associated values are empty, NaN, Multiple or all identical.

 

Rows are rejected if one of the associated values are text, empty or invalid (NaN or Multiple).

 

Note:

Please note that the in this module the Data information come form the 2D images associeated with the rows.  So the number of Data columns will always be "0".

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "ROI" page

 

 

This page is used to limit the surface of each slices used in training or testing.

 

 

 

 

 

Slice selection

You can select the slice you want to clip from this list. 

- You can also use the mouse scroll wheel when the cursor is not over one of the clip sliders.

- You can also click on the desired slice in the "Clip status" region.

 

Clip planes

You can move the 4 clip planes used to define the ROI for each slice.

 

Clip status

The status for each slice is composed of 3 lines:

The slice number (directory - slice) along with the ratio of the ROI over the complete slice.

The clipped FoV.  Or the dimensions of the ROI defined by the clip planes.  The highest FoV value will be displayed in red.

The original FoV of the slice.

 

Exit

Exit the configuration interface.

 

Help

Display this web page.

 

 

 

The config "Preproc" page

 

 

This page is used to control any pre-processing that need to be applied to the data.

 

 

 

 

CSV to Image index pairing

 

Since we have both 2D and 1D datasets loaded, we need a way to match the 2D images with specific rows in the 1D CSV files. This is done with the "CSV to Image index pairing".  There is one row of pairing for each index column selected in the "Files 1D" page.

 

On the left hand side, we have the headers of the index column we selected in the "Files 1D" page, we need to find some matching information in the header of the loaded images.

 

By pressing the  button, we open a list of available header information.  We can either select one of these or specify a DICOM tag in the text field.

 

DICOM tags are specified using the syntax: "(xxxx,yyyy)" where "xxxx" and "yyyy" are the DICOM tags in hexadecimal.

 

You can see the results for each of the pairing row by pressing on the  button.

 

 

CSV table & Images Files

 

Here, we see the results of the pairing specified in the previous section.

 

The left hand side, the "CSV table" side, we see how many rows our CSV data contain, we see how many of these rows are matched to a single image how many are unmatched and how many are matched to more than one image.

 

On the right hand side, we have the corresponding values for the loaded images: how many are loaded, how many are matched to a single CSV row, how many are unmatched and how many are matched to more than one row.

 

Pixel resizing

Before training, testing or predicting on the slices, they will all be re-sized to the desired pixel size.  So the pixel size you chose will affect the precision of your prediction.  You want  a smaller pixel for more precise results.  On the other hand, a smaller pixel size will result in bigger images (higher resolution) so the training will take longer. 

 

There is no reason to have a pixel size smaller than the smallest original pixel size (see "Resizing status").

 

Once you select a pixel size, the program will use it to divide the biggest FoV (in both X or Y) from all the loaded images to find the resolution that will be used in the training.  This resolution must be a multiple of 16 for the U-Net model to work. The FoV of the training dataset will then be the pixel size multiplied by new resolution.

 

If you want to increase the training speed, you may want to reduce the FoV of the biggest image using the ROI interface.  The biggest FoV value will be displayed in red in the Resizing status region.

 

Resizing status

For each slice we will have 3 lines of information:

- The slice number (dir - slice)

- The original pixel size

- The clipped FoV

- The original FoV

 

Intensity normalization

The AI expect floating point pixels values between -1 and 1. While medical images use integer values.  So, we need to "normalize" the pixel values.

 

You have 3 choices: "Min/Max", "Z-Score" or "GLI range".

 

The "Min/Max" normalization will normalize the GLI values so that the minimum value found in an image will be -1 and the maximum will be 1.

 

The "Z-Score" normalization will normalize the GLI values so that the mean value of an image will be at 0 and and the standard deviation is 1.  The idea is that all the values within 1 standard deviation of the mean will be between -1 and 1.

 

The "GLI range" is mainly used for CT images.  All HU values under the minimum will be assigned a value of -1, and all HU greater than the maximum will be assigned a value of +1 in the normalized images.  All HU values in-between will be linearly mapped between -1 and 1.

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "Augment" page

 

 

This page enable you to select and control the different augmentations that are applied to the images before they are used for training. The augmentations will be applied to all the "training" images before each epoch.  Only the "Training" images are augmented, the images reserved for validation and testing are never augmented.

 

 

 

Original Image

Rotated

Scaled

Translated

 

 

 

Distorted

Flipped

With Noise

With "Pseudo MR"

 

 

 

 

 

Augmentations

You can enable/disable the different augmentations, and specify the intensity of each of these from this interface.

You have access to 9 augmentations.  Each of these is controlled by a selected value "val":

Rotation:

A rotation of the image, centered on the center of the image, of a random angle going from -val to +val.

Scaling:

A random scaling of the image, ranging from -val % to +val %.

Translate Horizontal:

An horizontal translation of the image ranging from -val % to +val % of the width of the image.

Translate Vertical:

A vertical translation of the image ranging from -val % to +val % of the height of the image.

Distort:

A grid of 8x8 bi-cubic splines are superimposed on the image.  Each of the control points of these splines are randomly jiggled by a factor scaled by the value selected in the interface.  The resulting splines are used to distort the image.

Flip Horizontal:

If a random value is over a threshold (fixed by "val"), the image is flipped horizontally.

Flip Vertical:

If a random value is over a threshold (fixed by "val"), the image is flipped vertically.

Perlin (Noise):

We use a Perlin noise function of an high frequency to simulate noise in the image.

Perlin (Pseudo MR):

We use a Perlin noise function of a low frequency to simulate the intensity variations present in MR images.

 

Demo Original

The slider under the image is used to change the demo frame.  All the images used for training are available.

 

Demo Augmented

The slider under the image is used to show the range of augmentation. 

 

In normal training images, each time an image is augmented, a random value between -1 to +1 is assigned to each augmentation of each image. 

 

In this demo however, we use the slider to show the range of possible augmentation. The slider simulate the random variable, with a range of -1 (complete left) to +1 (complete right).

 

Exit

Exit the configuration interface.

 

Help

Display this web page.

 

 

The config "Metrics" page

 

 

When you train the AI, 2 sets of "metrics" are used: The Loss Metrics is the one used to compute the AI weights.  The other group of metrics are just computed to give you an idea of what is going on.  They are only there to provide you some feedback on the accuracy of the AI, they are optional and can be all off if you want.

 

At this time, you can only have 1 "Loss" metric when training your AI.  You can however have as many "Feedback" metrics as you want.

 

These metrics will be reported back to you when you train your AI.  They can also be used when you will select the weights you want to save for your AI.  During training you will be presented with 2 graphs:  One for the training and validation loss, and one for the training and validation values of the feedback metrics.

 

Depending on the type of data you want to predict, you will select a model that is either "Categorical" or "float" (more on that subject in the "Model" section of this page).  The loss and feedback metrics you select should also depend on your model.  For a "categorical" model, you should use the "Categorical_Cross_Entropy" loss and the "Categorical_Accuray" metric.  For float models, I would suggest the "Mean_Squared_Error" loss and the "Absolute_Error" metrics.

 

For more info on the different loss and metrics functions I suggest reading the web page: "How to Choose Loss Functions When Training Deep Learning Neural Networks"

 

 

 

 

Loss Metrics

Select one of the metrics as a "Loss" function to train the AI.

 

Feedback Metrics

Select any metrics you want from this list to get additional information on the AI training progress.

 

Text Window

This window reports on the metrics selections and the description of each metrics.

 

Description

This will cause the description  associated with each of the selected metrics to be displayed in the text window.

 

Edit Loss

The different metrics are all Python files that you can edit if you want.  You can also add your own metrics.  Just follow the prescribed syntax and add the python file of your new metric in the Python directory.  Then click the "re-scan dir" to cause sliceO to re-analyze the directory and add your metrics.  The syntax will be described later at the end of this section.

Edit Feedback

Re-scan dir

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "Split" page

 

 

 

 

 

Testing

A portion of the data images you selected in the config "File" page can be reserved for testing.

 

Testing Status

Report on the number of available data images.

 

Validation

You have access to 3 validation methods:

Training/Validation Split.  Probably the simplest, a part of the available images (those left after the images reserved for testing are removed) is reserved for validating the AI at each epoch.  The same rows are used for validation at each epoch.

K-Fold Cross-Validation.  The available images are split in "n" groups using the ratio defined in the K-Fold interface.  For example if the K-Fold is 1/5, then n=5.  We do a complete training of all the epochs reserving the images of one of these groups validation.  We then redo the complete training using another group for validation.  We do this n times.  After that, we assume that we have a good idea of the validation since by now all the images have been used for that, we do a last training without any validation.  In the "Train Save" page, the graphs will show the mean value of the n+1 pass for training metrics and the n pass with validation for validation metrics.

Leave one out Cross-Validation.  This is the extreme case of K-Fold where n is equal to the number of images.

 

Validation Status

Report the number of imagesthat will be used for training and for validation.

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "Model" page

 

 

In this page, you will select the model you want to use for the AI.  The models can be split in 2 groups: The "categorical" models and the "float" models.

 

The "categorical" models are used to predict either a binary value or are used to class images in categories.  For binaries values, the class column will have values that are either 1 or 0, for categories, the values will be an integer value representing the different categories (0 to n).

 

If you want to predict a floating point value, then you should use one of the "float" models.

 

 

 

 

Model List

Select a model from the list.

 

Model Feedback

Information on the selected model will be displayed here.

 

Model Summary

Call the Keras function "Summary" for the selected model.

 

Edit Model

The different models are all Python files that you can edit if you want.  You can also add your own models.  Just follow the prescribed syntax and add the python file of your new model in the Python directory.  Then click the "re-scan dir" to cause sliceO to re-analyze the directory and add your model.  The syntax will be described later at the end of this section.

 

Re-scan Python dir

 

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The config "Weights" page

 

 

 

 

 

Weight List

This list will display all the available weights for the selected model.  Select one of these from the list.

 

Weight Feedback

Information on the selected weights will be displayed here.

 

 

Weight Description

Cause the description of the weights to be displayed in the text window.

 

Edit Description

The weight description is a text file associated with the weights.  It has the same name as the weights but with the ".txt" extension. It can be edited with a simple text editor.  If you click "Edit", this file will be opened in Notepad.

 

Re-Scan weights

 

Re-scan the directory if you have saved some new weights for this model.

Exit

Exit the configuration interface.

 

Help

Display this web page

 

 

The "Train" page

 

 

 

 

Text Feedback

 

 

Config

Open the Configuration window

 

Train

 

Open the "Train" Window.

Save

Open the "Save" Window.

 

 

 

The Train "Train" page

 

 

Once you used the config pages to select the files you want to train on, and you selected the preprocessing steps, the augmentations, the validation split and the model you want to use, you are ready to start training your model to obtain the AI weights.

 

This is the page that will enable you to train the AI.

 

The only things that remain for you to select are the learning rate, the number of epochs and the batch size.  After that, you click on "Train" and wait (and wait... and wait...) for the results!

 

Once started, the training will:

Pre-load and normalize all the images used for training and validation in memory.

Then, for each step it will:

Reset or pre-load the weights (if you selected to re-train existing weights).

Then, for each epoch it will:

Reset and augment all images used for training.

Make a scrambled list of the images and associated data rows used for training.

Call the Keras function "train_on_batch" for the batches of training images (using the scrambled list).

Call the Keras function "test_on_batch" for the batches of validation images.

Report the results of this epoch and possibly save the current weights, the current metrics and a script associated with these to the "AI Temp" folder. We only save a total of 100 weights, so if you have more than 100 eopchs, we will only save weights each (nb_epoch/100) epochs.

 

The number of "steps" is dependant on the validation technique. A simple Train/Validation split us done in a single step. K-Fold and Leave-one-out cross validation require multiple steps.

 

 

Note:

 

The first call to "train_on_batch" is always longer than the others.  Do not be allarmed if the red dot in the progress bar stop for a few seconds.

 

Note:

 

The "ETA" information is computed at the end of each epoch.  So you need to wait for at least one epoch to finish before having an idea of how long the complete training will take.

 

Note:

 

If you stop the training, or if the computer (or sliceOmatic) crash.  You will have the option to re-start the process from where it stopped the next time you press "Start"

 

 

 

 

 

Progress Bar

Report on the training progression.

 

ETA

Estimated time of completion (computed after each epoch).

 

Step Feedback

Current step number.

 

Epoch Feedback

Current Epoch number.

 

Batch Feedback

Currently processed batch.

 

Text Feedback

textual information about the training progression.

 

Loss Graphic

A graphic showing the training and validation loss values as a function of the epochs.

 

Validation Graphic

A graphic showing the training and validation values of all the selected feedback metrics

 

Learning Rate

The learning rate used in training the AI.  the default value is 0.001

 

Max Epoch

The number of epoch used for the training.

 

Batch Size

The number of images used in each training and testing batch.  If the AI failed because of a lack of GPU memory, you can decrease this value.

 

Normalization

The AI need the values of the training dataset to be close to the range "-1" to "1".  But a lot of the model's computations will cause the values to drift from this range.  You can select to "Normalize" the data at each step of the model.  You have the choice of 3 normalization modes:

Batch

Group

Instance

 

The "Normalization" value is placed in the system variable: "$AI_NORMALIZATION" and is accessible to the model.  This is explained further down.

 

Dropout

A technique to help the training to converge is to "drop" a number of the computed weights at each step of the computation.  A "dropout" of 0.3 for example will reset 30% of the weights to zero.  This help prevent the training from converging toward a false local minima.

 

The "Dropout" value is placed in the system variable: "$AI_DROPOUT" and is accessible to the model.  This is explained further down.

 

Start

Start the training.

 

Pause/Resume

You can pause and resume the training.

 

Stop

Stop the training.

 

Exit

Exit the "Training" interface.

 

Help

Display this web page

 

 

The Train "Save" page

 

 

After having trained your AI, you want to save the weights you just computed.

 

This page will show you the weights of all the epochs of the training.  You need to select one of these and save it.  It will then be transfered from the AI_Temp folder to the sub-folder of the selected model with the name you specified.  A "description" file will also be saved along the weights in a file with the same name but with the ".txt" extension.

 

To help you select the optimal epoch for the weights, you have access to plots of the loss and feedback values along with their numerical values.

 

 

 

 

Epoch Selection

Select the desired epoch from the list.

 

Name

Name use for the saved weights.  A default value is proposed, but you can change it as you which.

 

Description

The description associated with the saved weights.  Again, a default value is proposed, but you can edit this test to your liking.

 

Epoch Feedback

Numerical values of the different metrics at the end of the selected epoch.

 

Loss Graphic

Plot of the training and validation loss as a function of the epochs.  The yellow vertical bar represent the selected epoch.  You can also select the epoch to save by clicking at the desired location in the graphic window

 

Validation Graphic

Plot of the training and validation feedback metrics as a function of the epochs.  The yellow vertical bar represent the selected epoch.  You can also select the epoch to save by clicking at the desired location in the graphic window

 

Save

Save the selected weights to the model's sub-folder.

 

Exit

Exit the "Save" interface.

 

Help

Display this web page

 

 

The "Test" page

 

This page is used to test your AI against a reserved portion of the selected rows.

 

The program will go through all the "test" rows, and call the Keras "test_on_batch" function for each of them.  It will report the different metrics for each rows and once all the rows are tested, compute the mean and standard deviation of each of the metrics. 

 

These results will also be saved in a Excel compatible file in the AI_Temp directory.

 

 

 

 

 

Text Feedback

 

 

Config

Open the Configuartion window

 

Test

Open the "Test Progress" Window.

 

 

 

The Test "Test" page

 

 

The results for each row will be presented in a table.  You can sort this table according to any of its column (just click on the column header).  You can also highlight a specific row of values by clicking on it.

 

 

 

 

 

Progress Bar

Show the progression of the testing

 

Text window

Report on the metrics for each image and the mean values.

 

Results table

 

Once the testing is done, this box will display the numerical results of the different metrics for each rows.  The results can be sorted acoording to any of the columns.  Just click on the desired column to change the sort order.

 

You can highlight a row by clicking on it.

 

Start

Start the computation of the tests.

 

Pause/Resume

Pause/resume the computation.

 

Stop

Stop the computation.

 

Exit

Exit this page.

 

Help

Display this web page

 

 

The Predict page

 

This is the page where the magic happens!

 

When clicking "Save" 4 files will be transferd from the "AI Temp" folder to a sub-directory of the "Python Code" folder.  This sub-folder will have the name of the model used for the training (ex: "U-Net", "Att Res U-Net"...). That sub-folder will contain 4 files, they will all have the name you selected but with different extensions:

The Keras generated weights (with the extension ".data-00000-of-00001").  This file will be big (more than 100Mb).

The Keras generated index (with the extension ".index").  This file has between 30Kb to 40Kb.

A sliceOmatic script file containing the training parameters (with the extension ".scp").

An ASCII "description" file (with the extension ".txt").

 

Note:

 

If you want to use weights provided to you by an external source, you need to create a sub-directory in the "Python Code" folder using the name of the model used to create the weights and copy the 4 weight files to that sub-directory.

 

Vice-versa, if you want to provide your weights to a colleague, you may want to "zip" the sub-folder containing the 4 files.  That way the sub-folder itself is preserved.  The recipient just have to unzip the file in the "Python Code" directory.

 

All the valid rows of the 1D CSV database loaded in sliceOmatic will then be used to "predict" the desired column when you press the "Compute" button.  If your loaded databse does not have a column matching the "class" column of the AI trained model, that column will be created and filled by the program.  If the column already exit, its cells will be overwritten by the program.

 

 

 

 

AI Selection

Select the desired AI from the list.  This will automatically set the Model and the associated Weights.

 

Rename

Rename the files associated with the selected AI (script, text and AI weights).

 

Delete

Delete the files associated with the selected AI.

 

Config

Open the Configuration window

 

Compute

Open the "Predict Progress" Window.

 

 

 

The Predict "Compute" page

 

 

 

 

 

Progress Bar

Show the progression of the predicting

 

Text Feedback

Report some info on the progression of the prediction.

 

Start

Start the computation of the prediction.

 

Pause/Resume

Pause/resume the computation.

 

Stop

Stop the computation.

 

Exit

Exit this page.

                    

 

The Python files

 

You can add your own metrics and models to this mode.  You just have to follow the syntax described here:

 

The Python directory.

 

All the Python files used by sliceOmatic's AI module are stored in the Python directory.  By default this is in "Documents\SliceO_Python".  The location of this directory can be changed through the sliceOmatic's "Config" interface.

 

The Metrics Python files.

 

For my metrics files, I use names that start with "metrics ", but the name is not really important.  What is important is that the file contain either a "metrics_2D_Fct" function, a "loss_2D_Fct" function or both.

 

You can have 3 types of metrics functions

 

The "Loss" metrics:

- The Loss functions compute a loss metrics used to train the AI.

The value of the loss function must decrease when we are approach perfect results.

The name of the loss function must be: "loss_2D_Fct" and have 2 arguments: "y_true" and "y_pred", 2 tensors, as described in Keras doc.

You must define a flag for the function: loss_2D_Fct.flag =... (The choices are between: "mean" "weighted" and "class" )

You must define the name of the function: loss_2D_Fct.__name__ =...

You must define a version for the function: loss_2D_Fct.__version__ =... (The version syntax is: "major.minor".  ex: "2.3")

 You must define the description of the function: loss_2D_Fct.description =...

 

 

import keras as kr

 

def loss_2D_Fct( y_true, y_pred ) :

    # --- We simply call the Keras function ---

    return( kr.losses.mean_squared_error( y_true, y_pred ) )

 

# ----------------------------------------------------

#   Give a name for this metrics

# ----------------------------------------------------

loss_2D_Fct.__name__ = "Mean_Squared_Error"

 

# ----------------------------------------------------

#   Give a version for this metrics

# ----------------------------------------------------

loss_2D_Fct.__version__ = "2.0"

 

# ----------------------------------------------------

#   Give a name for this metrics

# ----------------------------------------------------

loss_2D_Fct.description = "Call the Keras \"mean_squared_error\" function:\n" \

           "Computes the mean of squares of errors between labels and predictions."

 

You can have 2 types of metrics: "global" or "Per class".

 

The "global" metrics:

 

- The Metrics functions compute a metrics used as feedback when training or testing.

The name of the metrics function must be: "metrics_2D_Fct" and have 2 arguments: "y_true" and "y_pred", 2 tensors, (as described in Keras doc.)

You must define a flag for the function: metrics_2D_Fct.flag =... (The choices are between: "mean" "weighted" and "class" )

You must define the name of the function: metrics_2D_Fct.__name__ =...

You must define a version for the function: metrics_2D_Fct.__version__ =... (The version syntax is: "major.minor".  ex: "2.3")

You must define the description of the function: metrics_2D_Fct.description =...

 

 

from keras import backend as K

 

def metrics_2D_Fct( y_true, y_pred ) :

 

    # --- We simply call the Keras function ---

    return( kr.metrics.mean_absolute_error( y_true, y_pred ) )

 

# ----------------------------------------------------

#   Give a name for this metrics

# ----------------------------------------------------

metrics_2D_Fct.__name__ = "Absolute_Error"

 

# ----------------------------------------------------

#   Give a version for this metrics

# ----------------------------------------------------

metrics_2D_Fct.__version__ = "2.0"

 

# ----------------------------------------------------

#   This metrics is computed for all classes

# ----------------------------------------------------

metrics_2D_Fct.flag = "mean"

 

# ----------------------------------------------------

#   Give a name for this metrics

# ----------------------------------------------------

metrics_2D_Fct.description = "Call the Keras \"mean_absolute_error\" function:\n" \

           "Computes the mean of absolute errors between labels and predictions."

 

 

The "per class" metric:

 

The name of the metrics function must be: "metrics_2D_Fct" and have 2 arguments: "class_idx" and "name", the index and name of the target class, for "Per class" metrics.

The "per class" metrics return a "global" metric function for the desired class.

You must define the name of the function: metrics_2D_Fct.__name__ =...

You must define a version for the function: metrics_2D_Fct.__version__ =... (The version syntax is: "major.minor".  ex: "2.3")

You must define the description of the function: metrics_2D_Fct.description =...

You must also define a "flag" that must contain the string "class". (metrics_2D_Fct.flag = "class")

 

 

from keras import backend as K

 

def metrics_2D_Fct( class_idx: int, name: str ) :

 

    def test( y_true, y_pred ) :

 

#        print( "class[", class_idx, "] = ", name )

 

        # --- Compute this metrics on a single class ---

        y_true_single_class = y_true[..., class_idx]

        y_pred_single_class = y_pred[..., class_idx]

 

        # --- We simply call the Keras function ---

        return( kr.metrics.mean_absolute_error( y_true_single_class, y_pred_single_class ) )

 

    test.__name__ = f"test {name}"

 

    return test

 

# ----------------------------------------------------

#   Give a name for this metrics

# ----------------------------------------------------

metrics_2D_Fct.__name__ = "Absolute_Error"

 

# ----------------------------------------------------

#   Give a version for this metrics

# ----------------------------------------------------

metrics_2D_Fct.__version__ = "2.0"

 

# ----------------------------------------------------

#   This metrics is computed "per class"

# ----------------------------------------------------

metrics_2D_Fct.flag = "class"

 

# ----------------------------------------------------

#   Give a decription for this metrics

# ----------------------------------------------------

metrics_2D_Fct.description = "Compute the mean value of the abolute error for each class:\n" \

           "Calculates how often predictions match one-hot labels"

 

 

 

The Model Python files.

 

Again, I use the prefix "model " for my models files but the name is no important.  The important is that the file contain a "model_2D_Fct" function.

That function receive 4 arguments: dim_x, dim_y, dim_z and num_classes.  dim_x is the number of column used to train this model with and num_classes is the number of classes.  The "dim_z" argument is not used in 2D models.  The function must define the layers of your model and return the "model" variable returned by the Keras "Model" function.

You must define the name of the function: model_2D_Fct.__name__ =...

You must define a version for the function: model_2D_Fct.__version__ =... (The version syntax is: "major.minor".  ex: "2.3")

You must define the description of the function: model_2D_Fct.description =...

 

 

 

from tensorflow import keras

from keras.models import Model

from tensorflow.keras import layers

 

# =============================================================================

# =============================================================================

# ==================== Model 2D Fct Function ==================================

# =============================================================================

# =============================================================================

 

def model_2D_Fct( dim_x, dim_y, dim_z, num_classes ) :

 

#    print("Regression: dim_x = ", dim_x )

#    print("Regression: dim_y = ", dim_y )

#    print("Regression: dim_z = ", dim_z )

#    print("Regression: class = ", num_classes )

 

    inputs = keras.Input( shape=(dim_x,) )

 

    layer1 = layers.Dense(64, activation='relu')(inputs)

    layer1 = layers.Dropout(0.25)(layer1)

    layer2 = layers.Dense(64, activation='relu')(layer1)

    layer2 = layers.Dropout(0.25)(layer2)

    outputs = layers.Dense(num_classes)(layer2)

 

    # Define the model

    model = Model( inputs, outputs )

    return model

 

# ----------------------------------------------------

#   Give a name for this model

# ----------------------------------------------------

model_2D_Fct.__name__ = "Regression 2-layers"

 

# ----------------------------------------------------

#   Give a name for this model

# ----------------------------------------------------

model_2D_Fct.flag = "float"

 

# ----------------------------------------------------

#   Give a version for this model

# ----------------------------------------------------

model_2D_Fct.__version__ = "2.1"

 

# ----------------------------------------------------

#   Give a name for this model

# ----------------------------------------------------

model_2D_Fct.description = "Simple regression with 2 layers for predicting floating point values"

 

 

The Test Python files.

 

By default, I test for the presence of Python, Numpy, Tensorflow and Keras.  If you use another module in your Python code, you can add a new test file.  SliceO will use this to test that the desired module is indeed loaded when it start the AI module.

 

The file must contain a "version" function that return the version number for the target module.

It must also contain a "__name__" definition.

 

 

import numpy as np

 

# ----------------------------------------------------

#   The function must be called "version"

# ----------------------------------------------------

def version():

    print( "numpy vers: ", np.__version__ )

    return np.__version__

 

# ----------------------------------------------------

#   Give a name for this test

# ----------------------------------------------------

version.__name__ = "NumPy vers: "

 

 

Accessing sliceOmatic variables from Python.

 

You can access any of the sliceOmatic variables from your python code.

 

The functions for this are:

 

sliceO_Var_Get_Int( Variable_Name )

sliceO_Var_Get_Float( Variable_Name )

sliceO_Var_Get_Str( Variable_Name )

 

return one (or an array of) value(s), depending on the variable.

 

sliceO_Var_Set_Int( Variable_Name, Value )

sliceO_Var_Set_Float( Variable_Name, Value )

sliceO_Var_Set_Str( Variable_Name, Value )

 

Assign the value "Value" to the variable "Var_Name".  At this time only variable with one value can be "Set" (no arrays).

 

                    

From the Display Area

                    

There is no display area interaction specific to this mode.

 

 

From the Keyboard

 

There is no keyboard interface specific to this mode.

 

 

From the Command Line

 

System Variables defined in this library:

 

 

$AI_RES_X

(I16)

 

 

$AI_RES_Y

(I16)

 

 

$AI_RES_Z

(I16)

 

 

$AI_CLASS_NB

(I16)

 

 

$AI_CLASS_LABEL

A(W)

* $AI_CLASS_NB

 

$AI_CLASS_RATIO

A(F32)

* $AI_CLASS_NB

 

$AI_CLASS_WEIGHTS

A(F32)

* $AI_CLASS_NB

          

                    Importance computation:

 

$AI_IMP_BATCH

(I32)

Number of rows in each batch computation (def=1)

 

$AI_IMP_REPEAT

(I32)

Number of computation done to compute the mean value (def=4)

          

                    Training Parameters:          

 

$AI_TRAIN_STEP_MAX

(I32)

 

 

$AI_TRAIN_STEP_CUR

(I32)

 

 

$AI_TRAIN_EPOCH_MAX

(I32)

 

 

$AI_TRAIN_EPOCH_CUR

(I32)

 

 

$AI_TRAIN_BATCH_MAX

(I32)

 

 

$AI_TRAIN_BATCH_CUR

(I32)

 

 

$AI_TRAIN_LR

(F32)

 

 

$AI_TRAIN_EARLY_FLAG

(I32)

 

 

$AI_TRAIN_EARLY_NB

(I32)

 

          

                    Python Variables:          

 

$PYTHON_FLAG

(U32)

 

 

$PYTHON_MODULE_PATH

(W)

 

 

$PYTHON_INTERPRETER_PATH

(W)

 

 

$PYTHON_SLICEO

(P)

 

 

$PYTHON_TEMP

(W)

 

          

 

Commands recognized in this mode:

 

Python: path "path"

Python: temp "path"

 

A2F: file read "path"

A2F: file "path" (on|off)

 

A2F: class nb value

A2F: class label id "label"

 

A2F: data nb value

A2F: data label id "label"

A2F: data norm id a b id

 

A2F: index nb value

A2F: index label id "label"

 

A2F: metrics loss "name"

A2F: metrics feedback clear

A2F: metrics feedback "name"

 

A2F: split test value

A2F: split mode value

A2F: split split value

A2F: split kfold value

 

A2F: model cur "name"

A2F: model modality value

 

A2F: weights cur "name"

A2F: weights load