@@ -391,8 +391,9 @@ The following workflows have common variables with the above illustrated workflo
The following workflows contain a search space containing a set of possible neural networks architectures that can be used by `Distributed_Auto_ML` to automatically find the best combinations of neural architectures within the search space.
*Handwritten_Digit_Classification:* trains a simple deep CNN on the MNIST dataset using the PyTorch library. This example allows to search for two types of neural architectures defined in the Handwritten_Digit_Classification_Search_Space.json file.
*Single_Handwritten_Digit_Classification:* trains a simple deep CNN on the MNIST dataset using the PyTorch library. This example allows to search for two types of neural architectures defined in the Handwritten_Digit_Classification_Search_Space.json file.
*Multiple_Objective_Handwritten_Digit_Classification:* trains a simple deep CNN on the MNIST dataset using the PyTorch library. This example allows optimizing multiple losses, such as accuracy, number of parameters, and memory access cost (MAC) measure.
=== Distributed Training
...
...
@@ -1559,7 +1560,7 @@ AutoFeat currently supports the following encoding methods:
- Label: converts each value in a categorical feature into an integer value between 0 and n-1, where n is the number of distinct categories of the variable.
- Binary: stores categories as binary bitstrings.
- OneHot: creates a new feature for each category in the Categorical Variable and replaces it with either 1 (presence of the feature) or 0 (absence of the feature). The number of the new features depends on the number of categories in the Categorical Variable.
- OneHot: creates a new feature for each category in the categorical variable and replaces it with either 1 (presence of the feature) or 0 (absence of the feature). The number of the new features depends on the number of categories in the categorical variable.
- Dummy: transforms the categorical variable into a set of binary variables (also known as dummy variables). The dummy encoding is a small improvement over the one-hot-encoding, such it uses n-1 features to represent n categories.
- BaseN: encodes the categories into arrays of their base-n representation. A base of 1 is equivalent to one-hot encoding and a base of 2 is equivalent to binary encoding.
- Target: replaces a categorical value with the mean of the target variable.
...
...
@@ -1572,29 +1573,31 @@ There are numerous research papers and studies dedicated to the analysis of the
To access the AutoFeat page, please follow the steps below:
Open the link:https://try.activeeon.com/studio[Studio Portal].
. Open the link:https://try.activeeon.com/studio[Studio Portal].
Create a new workflow.
. Create a new workflow.
Drag and drop the `Import_Data_And_Automate_Feature_Engineering` task from the *machine-learning* bucket in the ProActive Machine Learning.
. Drag and drop the `Import_Data_And_Automate_Feature_Engineering` task from the *machine-learning* bucket in the ProActive Machine Learning.
Click on the task and click `General Parameters` in the left to change the default parameters of this task.
. Click on the task and click `General Parameters` in the left to change the default parameters of this task.
To get more information about the parameters of the service, please check the section <<Import_Data_And_Automate_Feature_Engineering>>.
Open the link:https://try.activeeon.com/automation-dashboard/#/portal/workflow-execution[Workflow Execution Portal].
[start=8]
. Open the link:https://try.activeeon.com/automation-dashboard/#/portal/workflow-execution[Workflow Execution Portal].
You can now access the AutoFeat Page by clicking on the endpoint `AutoFeat` as shown in the image below.
. You can now access the AutoFeat Page by clicking on the endpoint `AutoFeat` as shown in the image below.
.AutoFeat endpoint
image::AutoFeat_endpoint.png[align=center]
...
...
@@ -1658,6 +1661,15 @@ The user can also download the results as a csv file by clicking on the *Downloa
.Encoded data
image::AutoFeat_encoded_data.png[align=center]
=== ML Pipeline Example
You can connect different tasks in a single workflow to get the full pipeline from data preprocessing to model training and deployment. Each task will propagate the acquired variables to its children tasks.
The following workflow example `Vehicle_Type_Using_Model_Explainability` uses the `Import_Data_And_Automate_Feature_Engineering` task to prepare the data. It is available on the `machine_learning_workflows` bucket.
This workflow predicts vehicle type based on silhouette measurements, and apply ELI5 and Kernel Explainer to understand the model’s global behavior or specific predictions.
== ProActive Analytics
The *ProActive Analytics* is a dashboard that provides an overview of executed workflows