SmartPredict
  • Documentation
  • OVERVIEW
    • Presentation
      • Key features
      • Who may benefit from its use
      • The SmartPredict Graphical User Interface (GUI)
        • The SmartPredict Modules
        • Focus on the Notebooks
        • Practical shortcuts
    • Prerequisites
  • Getting started
    • Getting started (Part1) : Iris classification project
      • Project description
      • Step 1. Create the project
      • Step 2. Upload the dataset
      • Step 3. Preprocess the dataset
      • Step 4. Build the flowchart
        • Set up the flowchart
        • Configure the modules
      • Step 5. Run the build
      • Step 6. Deploy the project
      • Step 7. Make inferences with our model
      • Conclusion
  • Getting started (Part 2): Predicting the passengers' survival in the Titanic shipwreck
    • Project description
    • Step 1. Create the project
    • Step 2. Upload the dataset
    • Step 3. Preprocess the dataset
    • Step 4. Build the flowchart
    • Step 5. Run the build
    • Step 6. Deploy the pipeline
    • Step 7. Make inferences with our pipeline
  • MODULE REFERENCE
    • CORE MODULES
      • Introduction
      • Basic Operations
        • Item Saver
      • Web Services
        • Web Service IN and OUT
      • Data retrieval
        • Data fetcher
        • Data frame loader/converter
        • Image data loader
      • Data preprocessing
        • Introduction
        • Array Reshaper
        • Generic Data Preprocessor
        • Missing data handler
        • Normalizer
        • One Hot Encoder
        • Ordinal Encoder
      • Data selection
        • Features selector
        • Generic data splitter
        • Labeled data splitter
      • Training and Prediction
        • Predictor DL models
        • Predictor ML models
        • Predictor ML models (Probabilistic models)
        • Trainer ML models
        • Trainer/Evaluator DL models
      • Evaluation and fine-tuning
        • Cross Validator for ML
        • Evaluator for ML models
      • Machine Learning algorithms
        • ML modules in SmartPredict
        • Decision Tree Regressor
        • KNeighbors Classifier
        • KNeighbors Regressors
        • Linear Regressor
        • Logistic Regressor
        • MLP Regressor
        • Naive Bayes Classifier
        • Random Forest Classifier
        • Random Forest Regressor
        • Support Vector Classifier
        • Support Vector Regressor
        • XGBoost Classifier
        • XGBoost Regressor
      • Deep learning algorithms
        • Dense Neural Network
        • Recurrent Neural Networks
      • Computer Vision
        • Convolutional Recurrent Networks
        • Fully Convolutional Neural Networks
        • Face detector
        • Image IO
        • Image matcher
        • Yolo
      • Natural Language Processing
        • Introduction
        • Text cleaner
        • Text vectorizer
      • Times Series processing
        • TS features selector
      • TensorFlow API
        • LSTM Layer
        • Dense Layer
      • Helpers
        • Data/Object Logger
        • Object Selector (5 ports)
      • Conclusion
  • CUSTOM MODULES
    • Function
    • Class
    • Use cases
Powered by GitBook
On this page
  • The flowchart's elements
  • 1st Option : Workflow structure with a Dataframe loader
  • 2nd Option : Workflow structure without a Dataframe loader
  • Common configurations
  • The Features selector
  • The additional module : The Ordinal Encoder

Was this helpful?

  1. Getting started (Part 2): Predicting the passengers' survival in the Titanic shipwreck

Step 4. Build the flowchart

This section is about how to build or flowchart our Titanic project workflow .

PreviousStep 3. Preprocess the datasetNextStep 5. Run the build

Last updated 5 years ago

Was this helpful?

Building a flowchart with SmartPredict is just as easy as transporting modules into the workspace in drag and drop mode.

As our flowchart is borrowed from a classification template, all that is left to do is configure the modules' parameters to meet our specific needs.

It is deemed useful to remind that the flowchart we are going to build represents the ML pipeline from the processing steps to the model for initiating the prediction.

The flowchart's elements

The default build flowchart is composed of the following elements:

  • the Dataframe Loader

  • the Features Selector

  • the ML trainer

  • the Item saver

  • the ML evaluator

  • the Labeled Data splitter

  • the Data Object logger

  • and the Support Vector Classifier

We furthermore need to add :

  • an Ordinal Encoder in order to correctly handle the integer type of data.

  • [a processing pipeline + the original dirty dataset ] OR [a dataframe loader and a clean dataset]

Within the fields 'Columns to keep' and 'Columns to drop' enter the corresponding information .

The second flowchart shown above is another option for structuring our workflow , this time with a processing pipeline. It is what we are going to use for all the next steps.

To obtain the required configuration starting from the default flowchart, we furthermore need to add :

  1. the processing pipeline

  2. the unprocessed dataset

  3. an ordinal encoder

For our project , as we already have a processing pipeline, we are allowed to dispose of the Dataframe loader . To do so, select the Dataframe loader module in the flowchart and then delete it by clicking on the dot menu , then on delete. Click on OK.

Otherwise, by choosing to keep it , we shall need to save the processed dataset as a new dataset and place this latter within the flowchart right on top of the Dataframe loader.

Then, configure the settings and load parameters such as dataframes and features from the new processed dataset into the Dataframe loader.

In either ways , notice that both flowchart configurations need an Ordinal Encoder, else the model will signal an error.

Common configurations

The Features selector is part of the Core modules. It is located under the sub-tab of Data Selection modules. To configure it , select the features from the dataset inside the drag and drop area and select 'Survived' as a label.

The Ordinal Encoder module 's configuration is shown below:

1st Option : Workflow structure with a Dataframe loader

The dataframe loader is useful for loading the dataframe provided by a . This latter comes from the train dataset we initially had after some cleansing with the data processor .

2nd Option : Workflow structure without a Dataframe loader

The Features selector

The additional module : The Ordinal Encoder

Ordinal encoding deals with categorical data just like what we have here. You might already be familiar with the Ordinal encoding function. However, if you feel the need for more information , check its

🔩
👉
👉
✔️
🖇️
clean dataset
official documentation.
A Build flowchart with Dataframe loader included.
If we keep the DF loader, we need to insert the dataframe parameters.
Parameterizing the DF loader.
A Build flowchart without a DF loader but with a processing pipeline instead.
Insert 'Sex' in the columns to ordinal encode.