Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Tuning hyperparameters is a key concept in machine learning. You can perform hyperparameter tuning automatically using techniques such as grid search, random search, and gradient based optimization etc. This guide demonstrates an example on process of how to tune hyperparameters manually for a selected dataset and an algorithm, by performing a few tests using WSO2 Machine Learner.  You can use the same approach with different datasets and algorithms.

Table of Contents
maxLevel3

...

The approach on how to achieve the above goals on tuning the hyperparameters is described below.

Finding the optimal Learning Rate and the Number of Iterations which improves AUC

Follow the steps below to understand a fair number for the iterations to find the optimal learning rate.

  1. Upload your dataset (e.g. Pima Indians Diabetes dataset) to WSO2 ML. For instructions on uploading the dataset to the ML, see Exploring Data.
  2. Create a project, and then generate a model by creating an analysis. For instructions, see Generating Models.

    Tip

    Keep the Learning Rate at a fixed value (0.1), and vary the Number of Iterations in the Step 4 Parameters section of the model generating wizard in the WSO2 ML UI as shown below. 


    define hyper parameter values

  3.  Record the AUC value you obtain against each iterations number as shown in the example below. 

    table on varying number of iterations

    Tip

    You can get the AUC values from the Model Summary in the WSO2 ML UI as shown below.  

    view model summary

  4. Plot a graph using the results you obtained as shown in the example below.
    graph on varying number of iterations
    According to the above graph, AUC increases with the number of iterations. Hence, I picked you can pick 10000 as a fair number of iterations to find the optimal learning rate (of course I could have picked . (You may pick any number > 5000 (, where learning rate started to climb over 0.5)). Increasing number of iterations . However, increasing the Number of Iterations extensively would lead to an overfitted model).

  5. Now, find the optimal Learning Rate based on the ‘fair’ number for iterations you picked above. For that, create analyses by keeping the Number of Iterations at a fixed value (10000) and vary the Learning Rate in the Step 4 Parameters section of the model generating wizard in the WSO2 ML UI, as you did in step 2 above. 

  6. Record the AUC value you obtain against each Learning Rate as shown in the example below.

    table with varying learning ratesImage Added

  7. Plot a graph using the results you obtained as shown in the example below.

    graph with varying learning ratesImage Added

    According to the above graph, the AUC has a global maxima at 0.01 Learning Rate (precisely it is between 0.005 and 0.01). Hence, AUC maximizes when Learning Rate approaches 0.01. Therefore, 0.01 is the optimal Learning Rate for this particular dataset and algorithm.

  8. Now, keep the Learning Rate at 0.01, and vary the Number of Iterations again in the Step 4 Parameters section of the model generating wizard in the WSO2 ML UI, as you did in step 2 above. 
  9. Record the AUC value you obtain against each iterations number as shown in the example below.
    second table with varying number of iterationsImage Added
  10. Plot a graph using the results you obtained as shown in the example below.
    second graph with varying number of iterationsImage AddedAbove graph depicts that the AUC only increases slightly when you increase the Number of Iterations. 

Obtaining optimal values for the hyper parameters

Even though you further improve the Number of Iterations, AUC will probably not considerably improve. Therefore, decide on the optimal Number of Iterations based on the above observations, and also considering how much computing power you have and what level of AUC you expect. 

Also, use another binary classification algorithm (Support Vector Machine), or else carry out feature engineering on the dataset so that it reduces the noise of the training data to increase the AUC.