This section describes some recommended performance tuning configurations to optimize BAM. It assumes that you have set up the BAM server on Unix/Linux, which is recommended for a production deployment. If you have high volumes of data with high concurrency, it is recommended to use a distributed BAM setup. For instructions, see product deployment and clustering guide.
Info |
---|
|
OS-Level Settings
...
To optimize network and OS performance, configure the following settings in /etc/sysctl.conf
file of Linux. These settings specify a larger port range, a more effective TCP connection timeout value, and a number of other important parameters at the OS-level.
Code Block |
---|
net.ipv4.tcp_fin_timeout = 30
fs.file-max = 2097152
net.ipv4.tcp_tw_recycle = 1
net.ipv4.tcp_tw_reuse = 1
net.core.rmem_default = 524288
net.core.wmem_default = 524288
net.core.rmem_max = 67108864
net.core.wmem_max = 67108864
net.ipv4.tcp_rmem = 4096 87380 16777216
net.ipv4.tcp_wmem = 4096 65536 16777216
net.ipv4.ip_local_port_range = 1024 65535 |
Info |
---|
When we have the localhost port range configuration lower bound to 1024, there is a possibility that some processes may pick the ports which are already used by WSO2 servers. Therefore, it's good to increase the lower bound as sufficient for production, e.g., 10,000. |
To alter the number of allowed open files for system users, configure the following settings in /etc/security/limits.conf file of Linux.
Code Block |
---|
* soft nofile 4096
* hard nofile 65535 |
...
.
...
To alter the maximum number of processes your user is allowed to run at a given time, configure the following settings in /etc/security/limits.conf
file of Linux (be sure to include the leading * character). Each carbon server instance you run would require upto 1024 threads (with default thread pool configuration). Therefore, you need to increase the nproc value by 1024 per each carbon server (both hard and soft).
Code Block |
---|
* soft nproc 20000
* hard nproc 20000 |
JDBC Pool Configurations
Within the WSO2 platform, we use Tomcat JDBC pooling as the default pooling framework due to its production ready stability and high performance. The table below indicates some recommendations on how to configure the JDBC pool using the <AS_HOME>/repository/conf/datasources/master-datasources.xml file
. For more details about recommended JDBC configurations, see The Tomcat JDBC Connection Pool.
...
The maximum number of active connections that can be allocated from the connection pool at the same time. The default value is 100.
...
The indication of whether connection objects will be validated before they are borrowed from the pool. If the object validation fails, it will be dropped from the pool, and we will attempt to borrow another connection.
...
Setting this property to 'true' is recommended as it will avoid connection requests from failing. The validationQuery
property should be used if testOnBorrow is set to true. To increase the efficiency of connection validation and to improve performance, validationInterval
property should also be used.
...
To avoid excess validation, run validation at most at this frequency (time in milliseconds). If a connection is due for validation, but has been validated previously within this interval, it will not be validated again. The default value is 30000
(30 seconds).
...
Note |
---|
...
The |
...
values |
...
WSO2 BAM specific settings
Info |
---|
The values discussed below are general recommendations. They might not be optimal for the specific hardware configurations in your environment. We recommend you to carry out load tests on your environment to tune BAM accordingly. |
Improvement Area | Performance Recommendations | ||||||||
---|---|---|---|---|---|---|---|---|---|
Data receiver nodes |
| ||||||||
Analyzer nodes | Xms1024m -Xmx1024m -XX:MaxPermSize=512m | ||||||||
Dashboard nodes | Xms1024m -Xmx1024m -XX:MaxPermSize=512m | ||||||||
Hadoop nodes |
| ||||||||
Cassandra nodes |
For more information see Apache Cassandra 1.0 Documentation. | ||||||||
Memory size of Hive | In order to increase the memory size of Hive try the following configurations:
|
WSO2 Carbon platform-level settings
In multitenant mode, the WSO2 Carbon runtime limits the thread execution time. That is, if a thread is stuck or taking a long time to process, Carbon detects such threads, interrupts and stops them. Note that Carbon prints the current stack trace before interrupting the thread. This mechanism is implemented as an Apache Tomcat valve. Therefore, it should be configured in the <PRODUCT_HOME>/repository/conf/tomcat/catalina-server.xml
file as shown below.
Code Block | ||
---|---|---|
| ||
<Valve className="org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve" threshold="600"/> |
- The
className
is the Java class used for the implementation. Set it toorg.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve
. - The
threshold
gives the minimum duration in seconds after which a thread is considered stuck. The default value is 600 seconds.
JVM settings
When an XML element has a large number of sub-elements and the system tries to process all the sub-elements, the system can become unstable due to a memory overhead. This is a security risk.
To avoid this issue, you can define a maximum level of entity substitutions that the XML parser allows in the system. You do this using the entity expansion limit
attribute that is in the <DAS_HOME>/bin/wso2server.bat
file (for Windows) or the <DAS_HOME>/bin/wso2server.sh
file (for Linux/Solaris). The default entity expansion limit is 64000.
Code Block | ||
---|---|---|
| ||
-DentityExpansionLimit=100000 |
In a clustered environment, the entity expansion limit has no dependency on the number of worker nodes