Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

This section describes some recommended performance tuning configurations to optimize WSo2 ML. It assumes that you have set up the BAM server on Unix/Linux, which is recommended for a production deployment. 

...

PropertyDescriptionRecommendation
maxActive

The maximum number of active connections that can be allocated from the connection pool at the same time. The default value is  100.

This value should match the maximum number of requests that can be expected at a time in your production environment. This is to ensure that, whenever there is a sudden increase in the number of requests to the server, all of them can be connected successfully without causing any delays. Note that this value should not exceed the maximum number of requests allowed for your database.
minIdleThe minimum number of connections that can remain idle in the pool, without extra ones being created. The connection pool can shrink below this number if validation queries fail. Default value is 0.This value should be similar or near to the average number of requests that will be received by the server at the same time. With this setting, you can avoid having to open and close new connections every time a request is received by the server.
testOnBorrow

The indication of whether connection objects will be validated before they are borrowed from the pool. If the object validation fails, it will be dropped from the pool, and we will attempt to borrow another connection.

Setting this property to 'true' is recommended as it will avoid connection requests from failing. The validationQuery property should be used if testOnBorrow is set to true. To increase the efficiency of connection validation and to improve performance, validationInterval property should also be used.

validationInterval

To avoid excess validation, run validation at most at this frequency (time in milliseconds). If a connection is due for validation, but has been validated previously within this interval, it will not be validated again. The default value is  30000  (30 seconds).

This time out can be as high as the time it takes for your DBMS to declare a connection as stale. For example, MySQL will keep a connection open for as long as 8 hours, which requires the validation interval to be within that range. However, note that having a low value for validation interval will not incur a big performance penalty, specially when database requests have a high throughput. For example, a single extra validation query run every 30 seconds is usually negligible.

validationQueryThe SQL query used to validate connections from this pool before returning them to the caller. If specified, this query does not have to return any data, it just can't throw an SQLException. The default value is null. Example values are SELECT 1(mysql), select 1 from dual(oracle), SELECT 1(MS Sql Server).Specify an SQL query, which will validate the availability of a connection in the pool. This query is necessary when testOnBorrow property is true.

...

Improvement areaPerformance recommendation
ML JVM options

Change the following values in the <ML_HOME>/bin/wso2server.sh file: -Xms256m -Xmx1024m Xmx2048m -XX:MaxPermSize=256m

Info

These options predominate the performance of the ML jobs especially when the ML is run in standalone mode since Spark contexts are created within the JVM.


Concurrency in ML

Change the number of threads in the thread pool and the length of queue of threads in the <ML_HOME>/repository/conf/machine-learner.xml file as follows:

  • <Property name="ml.thread.pool.size" value="100" />

  • <Property name="ml.thread.pool.queue.size" value="1000" />
 Data Compression and Serialization (Spark)

Change the following I/O compression and Serialization properties in the <ML_HOME>/repository/conf/etc/spark-config.xml file:

  • <property name="spark.io.compression.codec">lzf</property>

  • <property name="spark.serializer">org.apache.spark.serializer.KryoSerializer</property>

  • <property name="spark.kryoserializer.buffer">256k</property>

  • <property name="spark.kryoserializer.buffer.max">256m</property>

Executor memory (Spark)Change the properties that define the amount of memory allocated for executors in a worker node for the application in the <ML_HOME>/repository/conf/etc/spark-config.xml file as follows: <property name="spark.executor.memory">1g</property>

 
 Executor cores (Spark) Change the number of cores allocated for executors in a worker node in the <ML_HOME>/repository/conf/etc/spark-config.xml file as follows: <property name="spark.executor.cores">2</property>
Info

For more information on tuning Spark, go to Tuning Spark and Spark Configuration.