|

I’m trying to redirect all url’s to https in a web application.
I have an ec2 instance behind an elastic load balancer. SSL is terminated on the load balancer.
Any attempts at redirection end up giving me the familiar ‘this page is requesting in a way which will never complete’.
The load balancer forwards 443 and 80 to port 80 on the instance.
This is what I have in my .htaccess.
RewriteCond %{X-FORWARDED-PROTO} !=on RewriteRule ^ https://%{HTTP_HOST}%{REQUEST_URI} [L,R=301]
Does anyone have any idea how to accomplish this? Kind thanks,

untested.. but I try this
RewriteCond %{HTTP:X-Forwarded-Proto} !https RewriteRule ^ https://%{SERVER_NAME}%{REQUEST_URI} [L,R]
ATG: Tuning Apache + Tomcat
I did some tuning on Apache which vastly decreased response times.
The tests show increase vast decrease in response. (see table below) which is due to the following improvements made effective by the tuning (Increased maxclients, KeepAlive on. KeepAliveTimeout etc):
1) Reusable connections to handle multiple requests from the same client, which in return causes less connections from client to the web-server and decreases response times.
2) Because of Point 1, apache now needs less Httpd clients to server more load, hence decrease in resource usage (Memory).
4) MaxClients has been increased to 1000, hence if required apache can now expand to serve more load without the queue buildup which causes high latencies.
Request |
100 |
100 (Tuned) |
Perf gain (%) |
400 |
400(Tuned) |
Perf gain (%) |
1-homePage |
850.21 |
253.52 |
70.18 |
4812.72 |
533.55 |
88.91 |
2-login |
1132.27 |
593.72 |
47.56 |
5411.01 |
1186.64 |
78.07 |
3-1-search |
1698.79 |
1616.14 |
4.87 |
4303.77 |
4366.99 |
-1.47 |
3-2-searchPagination |
1822.23 |
1005.70 |
44.81 |
8810.29 |
3429.22 |
61.08 |
4-productDetails |
1043.62 |
398.56 |
61.81 |
5454.44 |
686.17 |
87.42 |
5-cartOverviewDeleteAllProducts |
328.95 |
320.88 |
2.45 |
712.25 |
514.25 |
27.80 |
5-cartOverviewViaMiniCart |
1026.01 |
421.22 |
58.95 |
5248.49 |
743.63 |
85.83 |
6-cartOverviewAddProduct |
351.52 |
336.90 |
4.16 |
873.58 |
824.36 |
5.63 |
7-cartPreview |
1352.69 |
665.76 |
50.78 |
6218.38 |
1162.07 |
81.31 |
8-cartConfirm |
908.65 |
848.04 |
6.67 |
1662.67 |
1251.99 |
24.70 |
9-logout |
840.03 |
214.11 |
74.51 |
5095.36 |
523.79 |
89.72 |
The changes will require the following changes in the HTTP.cong (/etc/httpd/conf/httpd.conf)
—- KeepAlive on MaxKeepAliveRequests 100 KeepAliveTimeout 2
<IfModule prefork.c> StartServers 10 MinSpareServers 10 MaxSpareServers 20 ServerLimit 1000 MaxClients 1000 MaxRequestsPerChild 4000 </IfModule>
<IfModule worker.c> StartServers 10 MaxClients 1000 MinSpareThreads 25 MaxSpareThreads 75 ThreadsPerChild 25 MaxRequestsPerChild 1000 </IfModule> —
Reference: http://blog.springsource.org/2008/10/14/optimising-and-tuning-apache-tomcat-part-2/
Note: the above improves performance but exausts mod_jk thread pools, as the response are shorter now, hence more load.
Read the following on MPMs to fine tune it further:
Choosing an MPM
Apache 2.x supports pluggable concurrency models, called Multi-Processing Modules (MPMs). When building Apache, you must choose an MPM to use. There are platform-specific MPMs for some platforms: beos , mpm_netware , mpmt_os2 , and mpm_winnt . For general Unix-type systems, there are several MPMs from which to choose. The choice of MPM can affect the speed and scalability of the httpd:
- The
worker MPM uses multiple child processes with many threads each. Each thread handles one connection at a time. Worker generally is a good choice for high-traffic servers because it has a smaller memory footprint than the prefork MPM.
- The
prefork MPM uses multiple child processes with one thread each. Each process handles one connection at a time. On many systems, prefork is comparable in speed to worker, but it uses more memory. Prefork’s threadless design has advantages over worker in some situations: it can be used with non-thread-safe third-party modules, and it is easier to debug on platforms with poor thread debugging support.
For more information on these and other MPMs, please see the MPM documentation.
http://httpd.apache.org/docs/2.2/misc/perf-tuning.html
http://tomcat.apache.org/connectors-doc/reference/workers.html
https://community.jboss.org/wiki/OptimalModjk12Configuration
The first thing that we need to do as a part of this exercise is to identify the parameters that we have to work upon. Below is a list of different things that we can look upon before actually going into the tuning part.
JVM Settings a) JAVA HEAP SIZE b) PERM GEN SPACE c) Parallel GC? (Good for multi CPU systems)
DB settings a) Max Pool and Min Pool sizes for all Oracle and MySQL Pools. b) idle-timeout-minutes c) blocking-timeout-millis
HTTP connector Settings (Thread Pool Settings) a) Max Threads — Concurrency b) Accept Count — Queue size of each thread c) Compression — True/False
JSP Compilation Settings a) We don’t need to check for compiled JSP’s every now and then unlike in dev. b) No debugging for JSP’s.
Security Settings a) Remove universal access to JMS console and WEB consoles. b) We can even try to restrict the access to this from the local machine.
Log4J Settings a) Make sure we log only the critical errors. http://tomcat.apache.org/tomcat-5.5-doc/config/http.html
Load Testing the application a) This is one critical part that needs lot of application context. First of all we have to be really clear on what we are going to load test and the most probable pain points to look for. Then we can think of the data to test with.
Hibernate Settings a) Cache Size
OS Settings a) Large Pages (2.6 kernel supports it seems)
Apache for serving Static Content a) Traditionally App servers are not great at serving static content. Apache is generally good when it comes to a serving static content. Some reasons for this are web servers having their own disk caching etc. b) Over time, app servers have improved a lot in serving static content. But going for a web server for serving static content gives us the flexibility of creating a new domain and the content download can happen much faster.
One of my clients is a large e-commerce company (>$500MMin annual revenuse through their website). They use Akamai for caching some of their image data, but some of the other images are served off of their server. They are trying to minimize the load on their JBoss instances and wanted to have Apache serve the static content when possible while JBoss handled the dynamic content. Both Apache and JBoss were running on the same Linux server.
This is a problem that I have faced a couple of times and every time I spend a few hours futzing with Apache configurations trying to get this to work. So I figure I should post the basic solution and never have to look hard for it again. Also, if someone has better suggestions, let me know.
Step 1: Activate the Requisite Modules
I use four modules for getting this to work
Make sure these are activated in your configuration
Step 2: Setting Up The Static Content
First step is to set up a directory location that will ahve the content. In my example, I am dealing with an exploded EAR in the Jboss home directory. This results in a configuration like
#
# Configure the static content directory and alias. This is pointing at the cache
# file directory
#
Alias /static-content/ "/usr/local/jboss/cache/app"
#
# Do not allow any overriding from within the cache directory. #
<directory /usr/local/jboss/cache/app" >
Options None
AllowOverride None
Order allow,deny
Allow from all
</directory>
Step 3: Set up the rewrite rules
#
# Rewrite all static content to be served from the
# static content directory
#
<IfModule mod_rewrite.c<
RewriteEngine on
#
# Setup the static content rule. Any file that comes in that# we have on the filesystem, we should serve from there.
#
# We want this to be the last rule processed if we meet the conditions
# and we need to pass through so that the static-content alias works.
#
RewriteCond /usr/local/jboss/cache/app/%{REQUEST_FILENAME} -f
RewriteRule ^/?(.*)$ /static-content/$1 [L,PT]
</IfModule>
<IfModule mod_proxy.c>
# No need to forward proxy
ProxyRequests Off
# Make sure we clean up the headers
# you may also wish to use the mod_proxy_ajp to proxy
# over the ajp protocol
ProxyPassReverse / http://localhost:8080/
# Needed to maintain proper host information
ProxyPreserveHost On
</IfModule>
First need to install apache, installation steps, see its official website, and then install apache jk module
1.1. Wget http://mirror.bjtu.edu.cn/apache//tomcat/tomcat-connectors/jk/source/jk-1.2.31/tomcat-connectors-1.2.31-src.tar.gz 1.2. Tar-xzvf tomcat-connectors-1.2.31-src.tar.gz 1.3. Cd tomcat-connectors-1.2.31-src/native 1.4 .. / configure – with-apxs = / usr/local/apache2/bin/apxs # / usr/local/apache2 / your apache installation directory 1.5. Make The 1.6. Cp apache-2.0/mod_jk.so / usr/local/apache2/modules/mod_jk.so, and then restart apache 1.7 into the apache installation directory, cd / usr/local/apache2 /
2. Configuration mod_jk.conf 2.1. Cd conf, vi mod_jk.conf 2.2 to edit the following:
# Global config # Just configure log and load module # Load module The LoadModule jk_module modules / mod_jk.so # said the import jk module
JkLogFile logs / mod_jk.log # logs to save JkLogStampFormat “[% a% b% d% H:% M:% S% Y]” # log timestamp format JkRequestLogFormat “% w% V% U% T” # log records the contents of the JkLogLevel warn JkWorkersFile conf / workers.properties # worker configuration file is loaded
# All vhost should inherit this config The virtual host JkMountCopy All # copy the configuration file
2.3 JkRequestLogFormat parameter description:
% B Bytes sent, excluding HTTP headers (CLF format) % B Bytes sent, excluding HTTP headers % H The request protocol % M The request method % P The canonical Port of the server serving the request % Q The query string (prepended with a? If a query string exists, otherwise an empty string) % R First line of request % S Request HTTP status code % T Request duration, elapsed time to handle request in seconds ‘.’ Micro seconds % U The URL path requested, not including any query string. % V The canonical ServerName of the server serving the request % V The server name according to the UseCanonicalName setting % W Tomcat worker name % R Session route name (available with 1.2.19 and up)
3. Configuration workers.properties
vi workers.properties
# Can access the work machine, status used to monitor the status of the jk a working machine can be understood as a tomcat instance worker.list = balancer, master, status
# Type represents the type of work the machine, generally type = ajp13 and lb lb load balancing worker.gmaster.port = 8009 worker.master.host = 10.3.1.22 worker.master.type = ajp13 worker.master.lbfactor = 1 # load balancing factor, if you want this machine carrying the load, can set value # Define web app slave host worker.slave.port = 9009 worker.slave.host = localhost worker.slave.type = ajp13 worker.slave.lbfactor = 1
# Define web app balancer worker.balancer.type = lb # load balancing worker.balancer.balance_workers = master, lave # load balance work machine, this list of workers can not be placed worker.list worker.balancer.sticky_session = True # session the request refers to the original tomcat
# Define status worker.status.type = status # the specified monitoring work machine, you can not set
4. Configuration uriworkermap.properties, and it used to match the tomcat which url is forwarded to
4.1. Vi uriworkermap.properties
# Configure status pattern / Jkstatus = status
# Configure balancer patterns # Just forword *. Jsp, *. Do and page serverlet / Appname / *. Jsp = balancer / Appname / *. Do = balance / Appname / page = balance / Appname / * / page = balance / Appname / = balance / Appname = balance
# Configure admin functionality, the backstage management program can only be matched to a working machine, because the Admin user less / Appname / admin / *. Jsp = master / Appname / admin / *. Do = master / Appname / admin / *. Html = master / Appname / admin / page = master / Appname / admin / * / page = master / Appname / admin / = master
4.2. Matching priority instructions: The most restrictive URI pattern is applied first. More precisely the URI patterns are sorted by the number of ‘/’ characters in the pattern (highest number first), and rules with equal numbers are sorted by their string length (longest first).
If both distinctions still do not suffice, then the defining source of the rule is considered. Rules defined in uriworkermap.properties come first, before rules defined by JkMount (Apache) and inside workers.properties using the mount attribute.
All disabled rules are ignored. Exclusion rules are applied after all normal rules have been applied.
There is no defined behaviour, for the following configuration conflict: using literally the same URI pattern in the same defining source but with different worker targets.
5. Configuration httpd.conf
vi httpd.conf
Include conf / mod_jk.conf loaded in the last
6. Configuration httpd-vhosts.conf
vi extra / httpd-vhosts.conf
<VirtualHost *:80> ServerName yourweb.com ServerAdmin your email@mail.com DocumentRoot “/ usr/local/apache2/your doc root” # Placed Alias block RewriteEngine on RewriteRule ^ / logging.php * yourapp / admin / [PT]
The load configuration JkMountFile conf / uriworkermap.properties # rules, only the virtual host effect </ VirtualHost>
PT passthrough, sent directly apache internal R denotes redirect Faster than the the jk match, RewriteRule processing the Rewrite Rules can see apache official website, it can modify many acts like a hook,
7. Use apache Alias apache static file forwarding, static file requests tomcat, services provided directly by apache This period on <VirtualHost *:80> </ VirtualHost> DocumentRoot general on the next line to
Alias / yourapp / js / var / yourapp_static / js / <Directory “/var/yourapp_static/js”> Options Indexes MultiViews AllowOverride None Order allow, deny Allow from all </ Directory>
You can also place multiple aliases block
8. Configure tomcat 8.1 Configuring the the jvm memory of vi bin / catalina.sh add on top: JAVA_OPTS = ‘-Xms1000m-Xmx1000m’ 8.2 configure ajp connector, configured in server.xml <Connector protocol = “AJP/1.3” port = “0” channelNioSocket.port = “8009” channelNioSocket.redirectPort = “8443” channelNioSocket.maxThreads = “256” channelNioSocket.maxSpareThreads = “97” channelNioSocket.minSpareThreads = “50” channelNioSocket.URIEncoding = “UTF-8” URIEncoding = “UTF-8” channelNioSocket.connectionTimeout = “20000” channelNioSocket.keepAliveTimeout = “1000” channelNioSocket.bufferSize = “16384” />
Using nio 8.3 production environment, you can turn off the http connector
9. Configure the number of operating system file descriptor
vi / etc / profile
Add to ulimit-SHn 51200 After the save source / etc / profile
The # general system file descriptor is 1024, you can use ulimit-a view, when a large number of tcp connections, file descriptors might not be enough Parameter Description:
-H set hard resource limits, once set can not be increased. -S set the soft resource limit set to increase, but not more than the hard-resource settings. -A shows all current limit information. -C core file size units to blocks. The maximum data size of the segment of the process of-d, in Kbytes. -F process can create a maximum file blocks. -L maximum to lock memory size in Kbytes units. -M maximum memory size in Kbytes for. -N can open the maximum number of file descriptors. -P pipe buffer size in Kbytes units. -S thread stack size in Kbytes units. -T maximum CPU usage time, in seconds. -U user maximum available number of processes. -V process virtual memory available to Kbytes units.
In a continuing effort to make my hosting cloud more dynamic, I have had to become relatively creative with my Apache configurations so that my farm can remain relatively flexible with little or no reconfiguration on newly introduced servers. This presents a big problem when working with a multi-site Apache configuration, and an even bigger problem when SSL certificates are involved. As many of you probably already know, you need to have one network interface (virtual or physical) dedicated to a particular domain SSL certificate. It’s easy enough to grab a wildcard SSL certificate from GoDaddy and bind a single network interface to all subdomains, but when you have a farm configuration where you’re hosting multiple domains from the same web servers, then you’ll need to have a dedicated NIC for each wildcard domain SSL certificate. That’s ok, we can handle this by creating virtual NICs (eth0:1) and giving them their own IP address, and Apache is none the wiser. But, this is where our Apache configuration starts to become a problem. With Apache, we know that we can listen by IP address, but not by interface, so in order for us to keep a common configuration between all of the servers in our farm, we’ll need to figure out a way for it to be able to dynamically handle binding when the IP address is not statically defined.
If you’re not fully understanding the issue right now, that’s ok. Let me outline the environment consideration a little bit better with a pseudo-diagram:
Load Balancer -
VIP1 = 10.1.1.2 (www.example1.com)
- www01.example1.com
- www02.example1.com
- www03.example1.com
VIP2 = 10.1.1.3 (www.example2.com)
- www01.example2.com
- www02.example2.com
- www03.example2.com
SERVER1:
www01.example1.com: www01.example2.com:
\ /
eth0 = 10.1.1.10 eth0:1 = 10.1.1.11
SERVER2:
www02.example1.com: www02.example2.com:
\ /
eth0 = 10.1.1.20 eth0:1 = 10.1.1.21
SERVER3:
www03.example1.com: www03.example2.com:
\ /
eth0 = 10.1.1.30 eth0:1 = 10.1.1.31
This is a relatively common and efficient setup, and in fact may be extrapolated by many more sites and domains hosted from the same server farms. Our Apache configuration is relatively simple to setup by telling the VirtualHost sections what IP to listen on, as such (www01 example):
Listen 10.1.1.10:80
<VirtualHost 10.1.1.10:80>
DocumentRoot /var/www/example1_root
CustomLog logs/www.example1.com-access_log combined
ErrorLog logs/www.example1.com-error_log
</VirtualHost>
Listen 10.1.1.11:80
<VirtualHost 10.1.1.11:80>
DocumentRoot /var/www/example2_root
CustomLog logs/www.example2.com-access_log combined
ErrorLog logs/www.example2.com-error_log
</VirtualHost>
The problem with this setup is that as we deploy a new box into the farm, we’ll need to reconfigure Apache to listen on the new IP addresses that it gets for it’s eth0 and eth0:1 … That’s ok, and really can be done with relative ease, but what if we want to have it so that we can deploy a new box (say, from VMware template?) with a preloaded configuration, and have it just work (mmm, Linked Clones anyone? )? For this case, we’ll need to completely redesign the way that our Apache configuration is structured so that we don’t need to do any configuration after a box is deployed, and so that we can have a common configuration between www01,www02,www03, etc… This is nice for management too because we’ll always know that each farm server has the same configuration, and we’ll never have to worry about typos or anything like that. Sounds nice, so I’ll stop the suspense.
Now, Apache gives us the nice and native ability to import environment variables into our configuration, which means that we can figure out everything that we need for our configuration prior to it being loaded, and pass those values on. This means, through our advanced (eh?) bash scripting skills, that we can figure out what the IP addresses of our interfaces are, populate an environment variable, and pass that knowledge on to Apache. We’ll start with /etc/sysconfig/httpd. From the httpd startup script, /etc/sysconfig/httpd gets inherited prior to loading the Apache configuration, so we can declare in here any variables that we will need and they will be retained within Apache’s execution scope.
The bare metal…
/etc/sysconfig/httpd:
# Configuration file for the httpd service.
#
# The default processing model (MPM) is the process-based
# 'prefork' model. A thread-based model, 'worker', is also
# available, but does not work with some modules (such as PHP).
# The service must be stopped before changing this variable.
#
#HTTPD=/usr/sbin/httpd.worker
#
# To pass additional options (for instance, -D definitions) to the
# httpd binary at startup, set OPTIONS here.
#
#OPTIONS=
#
# By default, the httpd process is started in the C locale; to
# change the locale in which the server runs, the HTTPD_LANG
# variable can be set.
#
#HTTPD_LANG=C
IPADDR_ETH0=$(ifconfig eth0 | grep inet | grep -v inet6 | awk '{print $2}' | sed -e s/addr://g)
IPADDR_ETH01=$(ifconfig eth0:1 | grep inet | grep -v inet6 | awk '{print $2}' | sed -e s/addr://g)
export IPADDR_ETH0 IPADDR_ETH01
/etc/httpd/conf.d/vhosts.conf:
# Pass in our environment variables where we've defined the IP addresses
# for each of our NICs
PassEnv IPADDR_ETH0
PassEnv IPADDR_ETH01
# Listen on 80 and 443 on our www0x.example1.com NIC
NameVirtualHost ${IPADDR_ETH0}:80
NameVirtualHost ${IPADDR_ETH0}:443
# Listen on 80 and 443 on our www0x.example2.com NIC
NameVirtualHost ${IPADDR_ETH01}:80
NameVirtualHost ${IPADDR_ETH01}:443
# Redirect all of our hosts to HTTPs -- Regular HTTP is soooo lame

<VirtualHost ${IPADDR_ETH0}:80>
RewriteEngine on
RewriteRule ^/(.*)$ https://%{HTTP_HOST}/$1 [L,R]
</VirtualHost>
<VirtualHost ${IPADDR_ETH01}:80>
RewriteEngine on
RewriteRule ^/(.*)$ https://%{HTTP_HOST}/$1 [L,R]
</VirtualHost>
# www.example1.com SSL example
<VirtualHost ${IPADDR_ETH0}:443>
DocumentRoot /var/www/example1
CustomLog logs/${HTTP_HOST}-access_log combined
ErrorLog logs/${HTTP_HOST}-error_log
TransferLog logs/${HTTP_HOST}-access_log
SSLEngine on
SSLCertificateFile /etc/httpd/conf.d/certs/_.example1.com.crt
SSLCertificateKeyFile /etc/httpd/conf.d/certs/example1.com.key
SSLCertificateChainFile /etc/httpd/conf.d/certs/gd_bundle.crt
</VirtualHost>
# www.example2.com SSL example
<VirtualHost ${IPADDR_ETH01}:443>
DocumentRoot /var/www/example2
CustomLog logs/${HTTP_HOST}-access_log combined
ErrorLog logs/${HTTP_HOST}-error_log
TransferLog logs/${HTTP_HOST}-access_log
SSLEngine on
SSLCertificateFile /etc/httpd/conf.d/certs/_.example2.com.crt
SSLCertificateKeyFile /etc/httpd/conf.d/certs/example2.com.key
SSLCertificateChainFile /etc/httpd/conf.d/certs/gd_bundle.crt
</VirtualHost>
And in doing it this way, we can dynamically create Apache configurations for each new box that we put into the farm. Using this method, we’ll be able to expand or reduce the size of the farm with little impact or effort.
As always, feel free to contact me with any questions, comments, or corrections
I spent a little bit of time getting a JBoss cluster running on my development machine and then configuring a Apache 2.2.17 server to load balance between the 2 node cluster. Here is how I did it.
I had to enable the following modules on Apache:
LoadModule proxy_module modules/mod_proxy.so
LoadModule proxy_ajp_module modules/mod_proxy_ajp.so
LoadModule proxy_balancer_module modules/mod_proxy_balancer.so
LoadModule rewrite_module modules/mod_rewrite.so
LoadModule ssl_module modules/mod_ssl.so
LoadModule status_module modules/mod_status.so
Then, I had to setup the rules for the balancing Apache proxy at the end of the conf/httpd.conf file:
ProxyReceiveBufferSize 4096
<proxy balancer://ajpcluster>
BalancerMember ajp://test.local.host:8109
BalancerMember ajp://test.local.host:8209
</proxy>
<location /jconsole>
ProxyPass balancer://ajpcluster/jconsole stickysession=JSESSIONID|jsessionid nofailover=On
ProxyPassReverse balancer://ajpcluster/jconsole
</location>
RewriteEngine On
RewriteCond %{HTTPS} off
RewriteRule (.*) https://%{HTTP_HOST}%{REQUEST_URI}
Redirect / /jconsole
RedirectMatch ^/(?i)admin /jconsole
RedirectMatch ^/(?i)status /jconsole
Then, I needed to enable the conf/mod-jk.conf file, update the http.conf to reflect this, and add this VirtualHost:
<virtualHost localhost>
ServerName localhost
JkMount /jconsole loadbalancer
JkMount /jconsole/* loadbalancer
</virtualHost>
Then, I needed a conf/uriworkermap.properties file:
# Simple worker configuration file
# Mount the Servlet context to the ajp13 worker
# Example
# Mapping the URI /myapp1 and everything under /myapp1/:
# /myapp1|/*=myworker-a
/jconsole|/*=loadbalancer
/jkstatus|/*=jkstat
Then, I needed a conf/workers.properties for mod-jk:
# Define list of workers that will be used
# The configuration directives are valid
# for the mod_jk version 1.2.18 and later
worker.list=loadbalancer,jkstat
# Define Node1
worker.ports1.port=8109
worker.ports1.host=127.0.0.1
worker.ports1.type=ajp13
worker.ports1.lbfactor=1
worker.ports1.prepost_timeout=10000
worker.ports1.connect_timeout=10000
worker.ports1.ping_mode=A
worker.ports1.connection_pool_size=10
# Define Node2
worker.ports2.port=8209
worker.ports2.host=127.0.0.1
worker.ports2.type=ajp13
worker.ports2.lbfactor=1
worker.ports2.prepost_timeout=10000
worker.ports2.connect_timeout=10000
worker.ports2.ping_mode=A
worker.ports1.connection_pool_size=10
# Load-balancing behaviour
worker.loadbalancer.type=lb
worker.loadbalancer.balance_workers=ports1,ports2
# Status worker for managing load balancer
worker.jkstat.type=status
And, that is basically it. You can download my portable , self extracting zip file, of my Tomcat and JBoss cluster running on “localhost”. Just unzip and read the readme file enclosed. It works as-is , by navigating to the app called “http://test.local.host” after you have it running.
First configure apache conf directory establish mod_jk.conf file, which reads as follows:
LoadModule jk_module modules / mod_jk.so The Xml code Favorites code
JkWorkersFile / usr/local/apache2/conf/workers.properties JkLogFile logs / mod_jk.log # Set the jk log level [debug / error / info] JkLogLevel info # Select the log format JkLogStampFormat “[% a% b% d% H:% M:% S% Y]” # JkOptions indicate to send SSL KEY SIZE, JkOptions + ForwardKeySize + ForwardURICompat-ForwardDirectories # JkRequestLogFormat set the request format JkRequestLogFormat “% w% V% T” JkMount / servlet / * worker1 JkMount / *. Jsp worker1 # JkMount / * loadbalancer # Apache will serve the static picture. # This means that all images will be parsed by APACHE # JkUnMount / *. Jpg loadbalancer # JkUnMount / *. Gif loadbalancer # JkUnMount / *. Swf loadbalancer # JkUnMount / *. Bmp loadbalancer # JkUnMount / *. Png loadbalancer
Establish worker.properties content is as follows: The Xml code Favorites code
# Worker.list = loadbalancer, server105, server106 worker.list = worker1 # Set properties for worker1 worker.worker1.type = ajp13 worker.worker1.host = localhost worker.worker1.port = 8009 worker.worker1.lbfactor = 50 worker.worker1.cachesize = 10 worker.worker1.cache_timeout = 600 worker.worker1.socket_keepalive = 1 worker.worker1.socket_timeout = 300 # Define the first node … # Worker.server105.port = 9090 # Worker.server105.host = 127.0.0.1 # Worker.server105.type = ajp13 # Worker.server105.lbfactor = 1 # # Worker.server105.local_worker = 1 # Worker.server105.cachesize = 100 # Define the 2nd node … # Worker.server106.port = 9090 # Worker.server106.host = 127.0.0.1 # Worker.server106.type = ajp13 # Worker.server106.lbfactor = 1 # Worker.server106.local_worker = 1 # Worker.server106.cachesize = 100 # Now we define the load-balancing behaviour # Worker.loadbalancer.type = lb # Worker.loadbalancer.balanced_workers = server105, server106 # Worker.loadbalancer.sticky_session = 1
Finally, you need to join in httpd.conf:
The Xml code Favorites code
Include conf/mod_jk2.conf
At this time start will be reported to not find mod_jk.so the error, you need to manually generate, go to http://archive.apache.org/dist/tomcat/tomcat-connectors/jk/source/ address to download a jakarta-tomcat- connectors-jk-1.2.6-src.tar.tar Can then be compiled to generate: The Xml code Favorites code
# Tar xzvf jakarta-tomcat-connectors-jk-1.2.6-src.tar.tar # Cd jakarta-tomcat-connectors-jk-1.2.6-src/jk/native #. / Configure – with-apxs = / usr/local/apache2/bin/apxs # Make # Cp ./apache-2.0/mod_jk.so / usr/local/apache2/modules /
Configured to apache part next jboss, jboss do not have to configure, modify ports, do not tell apache port conflicts can, the default is 8080, change 9090 like configuration file in jboss-4.0. 5.GA/server/default/deploy/jbossweb-tomcat55.sar/server.xml.
To done here, to engage in a war package, which is simple to get hold of jsp files, source code such as: The Html code Favorites code
Hello ~ ~ ~ ~ Image: <img src=”10.jpg”/>
War on jboss default / deploy directory, start jboss and apache jsp files through apache port access war package to try it! ~ ~
Then the question came, my html files on apache at, war package named Test, that I visited the project file is http://www.test.com:8080/Test/index.html, but such access does not to mod rules selvet and jsp Jump to jboss html in apache, but the apache not established Test directory, how to make him jump to apache static files of the same name in the specified directory at it?
Corresponding VirtualHost httpd.con plus RewriteRule ^ / the WebTest / (*) html / $ 1.html [R, NE], such access will jump up.
The same used in the jsp jpg file, you need to jump to the apache default directory, you need to add RewriteRule ^ / the WebTest / (*). Jpg / $ 1.jpg [R, NE];
Complete httpd.conf VirtualHost configuration: The Xml code Favorites code
<VirtualHost *> ServerAdmin sefer@xiaofeng.com DocumentRoot / usr/local/apache2/htdocs/sefer ServerName www.test.com ErrorLog logs / test.sefer.com-error_log TransferLog logs / test.sefer.com-access_log RewriteEngine On RewriteRule / test / (. *) / Index.html [R, NE] RewriteRule / out / (. *) Http://www.163.com [R, NE] RewriteRule ^ / WebTest / (. *). Html / $ 1.html [R, NE] RewriteRule ^ / WebTest / (. *). Jpg / $ 1.jpg [R, NE] </ VirtualHost>
So that the picture can come out, the the access http://www.test.com:8080/WebTest/index.html jump directly to http://www.test.com:8080/index.jsp.
he apache main job of the partial pressure, static page as apache on processing, servlets, jsp page ajp13 protocol gave tomcat, tomcat to handle.
http://archive.apache.org/dist/ http://archive.apache.org/dist/tomcat/tomcat-5/v5.5.16/ jakarta-tomcat-connectors-jk-1.2.6-src.tar.tar
# Tar xzvf jakarta-tomcat-connectors-jk-1.2.6-src.tar.tar
# Cd jakarta-tomcat-connectors-jk-1.2.6-src/jk/native
#. / Configure – with-apxs = / usr/local/apache2/bin/apxs
# Make
# Cp ./apache-2.0/mod_jk.so / usr/local/apache2/modules /
the workers.properties will generate jk / conf
To copy it to the apache / conf / with httpd.conf in the same directory,
To the CD mod_jk.so in httpd.conf and Canada
LoadModule jk_module modules / mod_jk.so
If workers.properties. Named should join. JkWorkersFile “/ project / apache / conf / extra / workers.properties”
My machine plus mod_jk2.so
LoadModule jk_module modules / mod_jk.so
Install apache
sudo. / configure – with-apxs = / project / apache / bin / apxs
$ Tar xjvf httpd-2.2.3.tar.bz2 $ Cd httpd-2.2.3 / . / Configure – prefix = / project / apache – enable-rewrite – enable-so
Domain. / Configure – prefix = / usr / local / apache – enable-so – enable-mods-shared = all $ Make $ Sudo make install
jk operation http://localhost/jkstatus
Domain is not configured
<Directory “/home/splashnote/splashnote/webapp”> Options Indexes FollowSymLinks AllowOverride All Allow from all </ Directory>
<VirtualHost *:80> DocumentRoot / home / splashnote / splashnote / webapp ServerName dns.websketch.com ServerAlias *. Websketch.com DirectoryIndex home.html
<IfModule Mod_rewrite.c> RewriteEngine on RewriteCond% {HTTP_HOST} ^ [^.] + /. Websketch /. (Com) $ RewriteRule ^ (. +)% {HTTP_HOST} $ 1 [C] RewriteRule ^ ([^.] +) /. Websketch /. (Com) (. *) $ / Home / splashnote / splashnote / webapp / sylvan $ 3? $ 1 /% {QUERY_STRING} [L] </ IfModule>
</ VirtualHost>
jk operation http://localhost/jkstatus
apache operation
sudo. / bin / apachectl.-t
sudo. / bin / apachectl start
ubuntu 8.04 installed apache2.2.10 experiences tomcat 6.0
Download apache under
zlib: . / Configure – prefix = / root / dev_ent / zlib apr: . / Configure – prefix = / root / dev_ent / apr
apr-util: . / Configure – with-apr = / root / dev_ent / apr / – with-berkeley-db = / usr / local / bdb (Note: apr-util Do – prefix = / / .. automatic install ¬ / usr/local/apr/bin/apu-1-config
apache: . / Configure – prefix = / root / dev_ent / apache – enable-so – enable-dav – with-apr = / root/dev_ent/apr/bin/apr-1-config – with-apr-util = / usr/local/apr/bin/apu-1-config (Enable-so – enable-dav with mod_jk.so to the apache load so)
svn: . / Configure – prefix = / root / dev_ent / subversion – with-berkeley-db = / usr / local / bdb – with-apxs = / root / dev_ent / apache / bin / apxs – with-apr = / root/dev_ent/apr/bin/apr-1-config – with-apr-util = / usr/local/apr/bin/apu-1-config – with-ssl – with-zlib = / root / dev_ent / zlib – enable-maintainer-mode Newspaper database bberkeleyDB error # Tar zxvf db-4.7.25.tar.gz # Cd db-4.7.25/build_unix / #.. / Dist / configure – prefix = / usr / local / bdb # Make # Make install
(Note berkeleyDB will occupy 8080)
workers.properties, mod_jk.conf
workers.properties: # The following can also do load balancing # Worker.list = controller, tomcat1, tomcat2
workers.tomcat_home = / root/dev_ent/apache-tomcat-6.0.14 workers.java_home = / root/dev_ent/jdk1.6.0_05 worker.list = tomcat1 # ======== Tomcat1 ========
worker.tomcat1.port = 8009
worker.tomcat1.host = localhost
worker.tomcat1.type = ajp13
worker.tomcat1.lbfactor = 1
# ======== Tomcat2 ========
# Worker.tomcat2.port = 9009
# Worker.tomcat2.host = localhost
# Worker.tomcat2.type = ajp13
# Worker.tomcat2.lbfactor = 1
# ======== Controller ========
# Worker.controller.type = lb
# Worker.controller.balanced_workers = tomcat1, tomcat2
# Worker.controller.balanced_workers = tomcat1 # Worker.controller.sticky_session = 1
mod_jk.conf file
LoadModule jk_module modules / mod_jk.so
JkWorkersFile conf / workers.properties
JkMount / *. Jsp controller
JkMount / *. Do controller
JkMount / *. Img controller
conf / httpd.conf
Include conf / mod_jk.conf Include conf / extra / website.conf
website.conf
JkMount / *. Jsp tomcat1
<Directory “/root/project/website/webapp”> Options Indexes FollowSymLinks AllowOverride All Allow from all </ Directory>
<VirtualHost *:80> DocumentRoot / root / project / website / webapp ServerName www.website.com JkMount / * tomcat1 JkMount / *. Jsp tomcat1 DirectoryIndex student.jsp </ VirtualHost>
tomcat conf / server.xml configuration
<Context displayName = “Welcome to Tomcat” docBase = “/ home / splashnote / splashnote / webapp” path = “/ splashnote08” reloadable = “true” workDir = “work/Catalina/localhost/splashnote08”> </ Context>
(Note: there is no path = “/”) / Etc / hosts ip weburl
JBoss AS 4.0.5
Copyright © 2006> JBoss, Inc.
High Availability Enterprise Services via JBoss Clusters
Clustering allows us to run an applications on several parallel servers (a.k.a cluster nodes). The load is distributed across different servers, and even if any of the servers fails, the application is still accessible via other cluster nodes. Clustering is crucial for scalable enterprise applications, as you can improve performance by simply adding more nodes to the cluster.
The JBoss Application Server (AS) comes with clustering support out of the box. The simplest way to start a JBoss server cluster is to start several JBoss instances on the same local network, using the run -c all command for each instance. Those server instances, all started in the all configuration, detect each other and automatically form a cluster.
In the first section of this chapter, I discuss basic concepts behind JBoss’s clustering services. It is important that you understand those concepts before reading the rest of the chapter. Clustering configurations for specific types of applications are covered after this section.
1.1.1. Cluster Definition
A cluster is a set of nodes. In a JBoss cluster, a node is a JBoss server instance. Thus, to build a cluster, several JBoss instances have to be grouped together (known as a “partition”). On a same network, we may have different clusters. In order to differentiate them, each cluster must have an individual name.
Figure 1.1, “Clusters and server nodes” shows an example network of JBoss server instances divided into three clusters, with each cluster only having one node. Nodes can be added to or removed from clusters at any time.
Figure 1.1. Clusters and server nodes
Note
While it is technically possible to put a JBoss server instance into multiple clusters at the same time, this practice is generally not recommended, as it increases the management complexity.
Each JBoss server instance (node) specifies which cluster (i.e., partition) it joins in the ClusterPartition MBean in the deploy/cluster-service.xml file. All nodes that have the same ClusterPartition MBean configuration join the same cluster. Hence, if you want to divide JBoss nodes in a network into two clusters, you can just come up with two different ClusterPartition MBean configurations, and each node would have one of the two configurations depending on which cluster it needs to join. If the designated cluster does not exist when the node is started, the cluster would be created. Likewise, a cluster is removed when all its nodes are removed.
The following example shows the MBean definition packaged with the standard JBoss AS distribution. So, if you simply start JBoss servers with their default clustering settings on a local network, you would get a default cluster named DefaultPartition that includes all server instances as its nodes.
<mbean code="org.jboss.ha.framework.server.ClusterPartition"
name="jboss:service=DefaultPartition">
<! -- Name of the partition being built -->
<attribute name="PartitionName">
${jboss.partition.name:DefaultPartition}
</attribute>
<! -- The address used to determine the node name -->
<attribute name="NodeAddress">${jboss.bind.address}</attribute>
<! -- Determine if deadlock detection is enabled -->
<attribute name="DeadlockDetection">False</attribute>
<! -- Max time (in ms) to wait for state transfer to complete.
Increase for large states -->
<attribute name="StateTransferTimeout">30000</attribute>
<! -- The JGroups protocol configuration -->
<attribute name="PartitionConfig">
... ...
</attribute>
</mbean>
Here, we omitted the detailed JGroups protocol configuration for this cluster. JGroups handles the underlying peer-to-peer communication between nodes, and its configuration is discussed in Section 2.1, “JGroups Configuration”. The following list shows the available configuration attributes in the ClusterPartition MBean.
-
PartitionName is an optional attribute to specify the name of the cluster. Its default value is DefaultPartition.
-
NodeAddress is an optional attribute to specify the binding IP address of this node.
-
DeadlockDetection is an optional boolean attribute that tells JGroups to run message deadlock detection algorithms with every request. Its default value is false.
-
StateTransferTimeout is an optional attribute to specify the timeout for state replication across the cluster (in milliseconds). Its default value is 30000.
-
PartitionConfig is an element to specify JGroup configuration options for this cluster (see Section 2.1, “JGroups Configuration”).
In order for nodes to form a cluster, they must have the exact same PartitionName and the ParitionConfig elements. Changes in either element on some but not all nodes would cause the cluster to split. It is generally easier to change the ParitionConfig (i.e., the address/port) to run multiple cluster rather than changing the PartitionName due to the mulititude of places the former needs to be changed in other configuration files. However, changing the PartitionName is made easier in 4.0.2+ due to the use of the ${jboss.partition.name} property which allows the name to be change via a single jboss.partition.name system property
You can view the current cluster information by pointing your browser to the JMX console of any JBoss instance in the cluster (i.e., http://hostname:8080/jmx-console/) and then clicking on the jboss:service=DefaultPartition MBean (change the MBean name to reflect your cluster name if this node does not join DefaultPartition). A list of IP addresses for the current cluster members is shown in the CurrentView field.
Note
A cluster (partition) contains a set of nodes that work toward a same goal. Some clustering features require to sub-partition the cluster to achieve a better scalability. For example, let’s imagine that we have a 10-node cluster and we want to replicate in memory the state of stateful session beans on all 10 different nodes to provide for fault-tolerant behaviour. It would mean that each node has to store a backup of the 9 other nodes. This would not scale at all (each node would need to carry the whole state cluster load). It is probably much better to have some kind of sub-partitions inside a cluster and have beans state exchanged only between nodes that are part of the same sub-partition. The future JBoss clustering implementation will support sub-partitions and it will allow the cluster administrator to determine the optimal size of a sub-partition. The sub-partition topology computation will be done dynamically by the cluster.
1.1.2. Service Architectures
The clustering topography defined by the ClusterPartition MBean on each node is of great importance to system administrators. But for most application developers, you are probably more concerned about the cluster architecture from a client application’s point of view. JBoss AS supports two types of clustering architectures: client-side interceptors (a.k.a proxies or stubs) and load balancers.
1.1.2.1. Client-side interceptor
Most remote services provided by the JBoss application server, including JNDI, EJB, RMI and JBoss Remoting, require the client to obtain (e.g., to look up and download) a stub (or proxy) object. The stub object is generated by the server and it implements the business interface of the service. The client then makes local method calls against the stub object. The call is automatically routed across the network and invoked against service objects managed in the server. In a clustering environment, the server-generated stub object is also an interceptor that understand how to route calls to nodes in the cluster. The stub object figures out how to find the appropriate server node, marshal call parameters, un-marshall call results, return the results to the caller client.
The stub interceptors have updated knowledge about the cluster. For instance, they know the IP addresses of all available server nodes, the algorithm to distribute load across nodes (see next section), and how to failover the request if the target node not available. With every service request, the server node updates the stub interceptor with the latest changes in the cluster. For instance, if a node drops out of the cluster, each of the client stub interceptor is updated with the new configuration the next time it connects to any active node in the cluster. All the manipulations on the service stub are transparent to the client application. The client-side interceptor clustering architecture is illustrated in Figure 1.2, “The client-side interceptor (proxy) architecture for clustering”.
Figure 1.2. The client-side interceptor (proxy) architecture for clustering
Other JBoss services, in particular the HTTP web services, do not require the client to download anything. The client (e.g., a web browser) sends in requests and receives responses directly over the wire according to certain communication protocols (e.g., the HTTP protocol). In this case, a load balancer is required to process all requests and dispatch them to server nodes in the cluster. The load balancer is typically part of the cluster. It understands the cluster configuration as well as failover policies. The client only needs to know about the load balancer. The load balancer clustering architecture is illustrated in Figure 1.3, “The load balancer architecture for clustering”.
Figure 1.3. The load balancer architecture for clustering
A potential problem with the load balancer solution is that the load balancer itself is a single point of failure. It needs to be monitored closely to ensure high availability of the entire cluster services.
1.1.3. Load-Balancing Policies
Both the JBoss client-side interceptor (stub) and load balancer use load balancing policies to determine which server node to send a new request to. In this section, let’s go over the load balancing policies available in JBoss AS.
In JBoss 3.0.x, the following two load balancing options are available.
-
Round-Robin (org.jboss.ha.framework.interfaces.RoundRobin): each call is dispatched to a new node. The first target node is randomly selected from the list.
-
First Available (org.jboss.ha.framework.interfaces.FirstAvailable): one of the available target nodes is elected as the main target and is used for every call: this elected member is randomly chosen from the list of members in the cluster. When the list of target nodes changes (because a node starts or dies), the policy will re-elect a target node unless the currently elected node is still available. Each client-side interceptor or load balancer elects its own target node independently of the other proxies.
In JBoss 3.2+, three load balancing options are available. The Round-Robin and First Available options have the same meaning as the ones in JBoss AS 3.0.x.
The new load balancing option in JBoss 3.2 is “First AvailableIdenticalAllProxies” (org.jboss.ha.framework.interfaces.FirstAvailableIdenticalAllProxies). It has the same behaviour as the “First Available” policy but the elected target node is shared by all client-side interceptors of the same “family”.
In JBoss 3.2 (and later), the notion of “Proxy Family” is defined. A Proxy Family is a set of stub interceptors that all make invocations against the same replicated target. For EJBs for example, all stubs targeting the same EJB in a given cluster belong to the same proxy family. All interceptors of a given family share the same list of target nodes. Each interceptor also has the ability to share arbitrary information with other interceptors of the same family. A use case for the proxy family is give in Section 1.3.1, “Stateless Session Bean in EJB 2.x”.
1.1.4. Farming Deployment
The easiest way to deploy an application into the cluster is to use the farming service. That is to hot-deploy the application archive file (e.g., the EAR, WAR or SAR file) in the all/farm/ directory of any of the cluster member and the application is automatically duplicated across all nodes in the same cluster. If node joins the cluster later, it will pull in all farm deployed applications in the cluster and deploy them locally at start-up time. If you delete the application from one of the running cluster server node’s farm/ folder, the application will be undeployed locally and then removed from all other cluster server nodes farm folder (triggers undeployment.) You should manually delete the application from the farm folder of any server node not currently connected to the cluster.
Note
Currently, due to an implementation bug, the farm deployment service only works for hot-deployed archives. If you put an application in the farm/ directory first and then start the server, the application would not be detected and pushed across the cluster. We are working to resolve this issue.
Note
You can only put archive files, not exploded directories, in the farm directory. This way, the application on a remote node is only deployed when the entire archive file is copied over. Otherwise, the application might be deployed (and failed) when the directory is only partially copied.
Farming is enabled by default in the all configuration in JBoss AS distributions, so you will not have to set it up yourself. The configuration file is located in the deploy/deploy.last directory. If you want to enable farming in your custom configuration, simply create the XML file shown below (named it farm-service.xml) and copy it to the JBoss deploy directory $JBOSS_HOME/server/your_own_config/deploy. Make sure that you custom configuration has clustering enabled.
<?xml version="1.0" encoding="UTF-8"?>
<server>
<mbean code="org.jboss.ha.framework.server.FarmMemberService"
name="jboss:service=FarmMember,partition=DefaultPartition">
...
<attribute name="PartitionName">DefaultPartition</attribute>
<attribute name="ScanPeriod">5000</attribute>
<attribute name="URLs">farm/</attribute>
</mbean>
</server>
After deploying farm-service.xml you are ready to rumble. The required FarmMemberService MBean attributes for configuring a farm are listed below.
-
PartitionName specifies the name of the cluster for this deployed farm. Its default value is DefaultPartition.
-
URLs points to the directory where deployer watches for files to be deployed. This MBean will create this directory is if does not already exist. Also, “.” pertains to the configuration directory (i.e., $JBOSS_HOME/server/all/).
-
ScanPeriod specifies the interval at which the folder must be scanned for changes.. Its default value is 5000.
The Farming service is an extension of the URLDeploymentScanner, which scans for hot deployments in deploy/ directory. So, you can use all the attributes defined in the URLDeploymentScanner MBean in the FarmMemberService MBean. In fact, the URLs and ScanPeriod attributes listed above are inherited from the URLDeploymentScanner MBean.
1.1.5. Distributed state replication services
In a clustered server environment, distributed state management is a key service the cluster must provide. For instance, in a stateful session bean application, the session state must be synchronized among all bean instances across all nodes, so that the client application reaches the same session state no matter which node serves the request. In an entity bean application, the bean object sometimes needs to be cached across the cluster to reduce the database load. Currently, the state replication and distributed cache services in JBoss AS are provided via two ways: the HASessionState MBean and the JBoss Cache framework.
-
The HASessionState MBean provides session replication and distributed cache services for EJB 2.x stateful session beans and HTTP load balancers in JBoss 3.x and 4.x. The MBean is defined in the all/deploy/cluster-service.xml file. We will show its configuration options in the EJB 2.x stateful session bean section later.
-
JBoss Cache is a fully featured distributed cache framework that can be used in any application server environment and standalone. It gradually replaces the HASessionState service. JBoss AS integrates JBoss Cache to provide cache services for HTTP sessions, EJB 3.0 session and entity beans, as well as Hibernate persistence objects. Each of these cache services is defined in a separate MBean. We will cover those MBeans when we discuss specific services in the next several sections.
1.2. Clustered JNDI Services
JNDI is one of the most important services provided by the application server. The JBoss clustered JNDI service is based on the client-side interceptor architecture. The client must obtain a JNDI stub object (via the InitialContext object) and invoke JNDI lookup services on the remote server through the stub. Furthermore, JNDI is the basis for many other interceptor-based clustering services: those services register themselves with the JNDI so that the client can lookup their stubs and make use of their services.
The JBoss HA-JNDI (High Availability JNDI) service maintains a cluster-wide context tree. The cluster wide tree is always available as long as there is one node left in the cluster. Each JNDI node in the cluster also maintains its own local JNDI context. The server side application can bind its objects to either trees. In this section, you will learn the distinctions of the two trees and the best practices in application development. The design rational of this architecture is as follows.
-
We didn’t want any migration issues with applications already assuming that their JNDI implementation was local. We wanted clustering to work out-of-the-box with just a few tweaks of configuration files.
-
We needed a clean distinction between locally bound objects and cluster-wide objects.
-
In a homogeneous cluster, this configuration actually cuts down on the amount of network traffic.
-
Designing it in this way makes the HA-JNDI service an optional service since all underlying cluster code uses a straight new InitialContext() to lookup or create bindings.
On the server side, new InitialContext(), will be bound to a local-only, non-cluster-wide JNDI Context (this is actually basic JNDI). So, all EJB homes and such will not be bound to the cluster-wide JNDI Context, but rather, each home will be bound into the local JNDI. When a remote client does a lookup through HA-JNDI, HA-JNDI will delegate to the local JNDI Context when it cannot find the object within the global cluster-wide Context. The detailed lookup rule is as follows.
-
If the binding is available in the cluster-wide JNDI tree and it returns it.
-
If the binding is not in the cluster-wide tree, it delegates the lookup query to the local JNDI service and returns the received answer if available.
-
If not available, the HA-JNDI services asks all other nodes in the cluster if their local JNDI service owns such a binding and returns the an answer from the set it receives.
-
If no local JNDI service owns such a binding, a NameNotFoundException is finally raised.
So, an EJB home lookup through HA-JNDI, will always be delegated to the local JNDI instance. If different beans (even of the same type, but participating in different clusters) use the same JNDI name, it means that each JNDI server will have a different “target” bound (JNDI on node 1 will have a binding for bean A and JNDI on node 2 will have a binding, under the same name, for bean B). Consequently, if a client performs a HA-JNDI query for this name, the query will be invoked on any JNDI server of the cluster and will return the locally bound stub. Nevertheless, it may not be the correct stub that the client is expecting to receive!
Note
You cannot currently use a non-JNP JNDI implementation (i.e. LDAP) for your local JNDI implementation if you want to use HA-JNDI. However, you can use JNDI federation using the ExternalContext MBean to bind non-JBoss JNDI trees into the JBoss JNDI namespace. Furthermore, nothing prevents you though of using one centralized JNDI server for your whole cluster and scrapping HA-JNDI and JNP.
Note
If a binding is only made available on a few nodes in the cluster (for example because a bean is only deployed on a small subset of nodes in the cluster), the probability to lookup a HA-JNDI server that does not own this binding is higher and the lookup will need to be forwarded to all nodes in the cluster. Consequently, the query time will be longer than if the binding would have been available locally. Moral of the story: as much as possible, cache the result of your JNDI queries in your client.
If you want to access HA-JNDI from the server side, you must explicitly get an InitialContext by passing in JNDI properties. The following code shows how to access the HA-JNDI.
Properties p = new Properties();
p.put(Context.INITIAL_CONTEXT_FACTORY,
"org.jnp.interfaces.NamingContextFactory");
p.put(Context.URL_PKG_PREFIXES, "jboss.naming:org.jnp.interfaces");
p.put(Context.PROVIDER_URL, "localhost:1100"); // HA-JNDI port.
return new InitialContext(p);
The Context.PROVIDER_URL property points to the HA-JNDI service configured in the HANamingService MBean (see Section 1.2.3, “JBoss configuration”).
1.2.2. Client configuration
The JNDI client needs to be aware of the HA-JNDI cluster. You can pass a list of JNDI servers (i.e., the nodes in the HA-JNDI cluster) to the java.naming.provider.url JNDI setting in the jndi.properties file. Each server node is identified by its IP address and the JNDI port number. The server nodes are separated by commas (see Section 1.2.3, “JBoss configuration” on how to configure the servers and ports).
java.naming.provier.url=server1:1100,server2:1100,server3:1100,server4:1100
When initialising, the JNP client code will try to get in touch with each server node from the list, one after the other, stopping as soon as one server has been reached. It will then download the HA-JNDI stub from this node.
Note
There is no load balancing behavior in the JNP client lookup process. It just goes through the provider list and use the first available server. The HA-JNDI provider list only needs to contain a subset of HA-JNDI nodes in the cluster.
The downloaded smart stub contains the logic to fail-over to another node if necessary and the updated list of currently running nodes. Furthermore, each time a JNDI invocation is made to the server, the list of targets in the stub interceptor is updated (only if the list has changed since the last call).
If the property string java.naming.provider.url is empty or if all servers it mentions are not reachable, the JNP client will try to discover a bootstrap HA-JNDI server through a multicast call on the network (auto-discovery). See Section 1.2.3, “JBoss configuration” on how to configure auto-discovery on the JNDI server nodes. Through auto-discovery, the client might be able to get a valid HA-JNDI server node without any configuration. Of course, for the auto-discovery to work, the client must reside in the same LAN as the server cluster (e.g., the web servlets using the EJB servers). The LAN or WAN must also be configured to propagate such multicast datagrams.
Note
The auto-discovery feature uses multicast group address 230.0.0.4:1102.
In addition to the java.naming.provier.url property, you can specify a set of other properties. The following list shows all client side properties you can specify, when creating a new InitialContext.
-
java.naming.provier.url: Provides a list of IP addresses and port numbers for HA-JNDI provider nodes in the cluster. The client tries those providers one by one and uses the first one that responds.
-
jnp.disableDiscovery: When set to true, this property disables the automatic discovery feature. Default is false.
-
jnp.partitionName: In an environment where multiple HA-JNDI services, which are bound to distinct clusters (i.e., partitions), are started, this property allows you to configure which cluster you broadcast to when the automatic discovery feature is used. If you do not use the automatic discovery feature (e.g., you could explicitly provide a list of valid JNDI nodes in java.naming.provider.url), this property is not used. By default, this property is not set and the automatic discovery select the first HA-JNDI server that responds, independently of the cluster partition name.
-
jnp.discoveryTimeout: Determines how much time the context will wait for a response to its automatic discovery packet. Default is 5000 ms.
-
jnp.discoveryGroup: Determines which multicast group address is used for the automatic discovery. Default is 230.0.0.4.
-
jnp.discoveryPort: Determines which multicast group port is used for the automatic discovery. Default is 1102.
1.2.3. JBoss configuration
The cluster-service.xml file in the all/deploy directory includes the following MBean to enable HA-JNDI services.
<mbean code="org.jboss.ha.jndi.HANamingService"
name="jboss:service=HAJNDI">
<depends>jboss:service=DefaultPartition</depends>
</mbean>
You can see that this MBean depends on the DefaultPartition MBean defined above it (discussed in an earlier section in this chapter). In other configurations, you can put that element in the jboss-services.xml file or any other JBoss configuration files in the /deploy directory to enable HA-JNDI services. The available attributes for this MBean are listed below.
-
PartitionName is an optional attribute to specify the name of the cluster for the different nodes of the HA-JNDI service to communicate. The default value is DefaultPartition.
-
BindAddress is an optional attribute to specify the address to which the HA-JNDI server will bind waiting for JNP clients. Only useful for multi-homed computers.
-
Port is an optional attribute to specify the port to which the HA-JNDI server will bind waiting for JNP clients. The default value is 1100.
-
Backlog is an optional attribute to specify the backlog value used for the TCP server socket waiting for JNP clients. The default value is 50.
-
RmiPort determines which port the server should use to communicate with the downloaded stub. This attribute is optional. If it is missing, the server automatically assigns a RMI port.
-
AutoDiscoveryAddress is an optional attribute to specify the multicast address to listen to for JNDI automatic discovery. The default value is 230.0.0.4.
-
AutoDiscoveryGroup is an optional attribute to specify the multicast group to listen to for JNDI automatic discovery.. The default value is 1102.
-
LookupPool specifies the thread pool service used to control the bootstrap and auto discovery lookups.
-
DiscoveryDisabled is a boolean flag that disables configuration of the auto discovery multicast listener.
-
AutoDiscoveryBindAddress sets the auto-discovery bootstrap multicast bind address. If this attribute is not specified and a BindAddress is specified, the BindAddress will be used..
-
AutoDiscoveryTTL specifies the TTL (time-to-live) for autodiscovery IP multicast packets.
The full default configuration of the HANamingService MBean is as follows.
<mbean code="org.jboss.ha.jndi.HANamingService"
name="jboss:service=HAJNDI">
<depends>
jboss:service=${jboss.partition.name:DefaultPartition}
</depends>
<! -- Name of the partition to which the service is linked -->
<attribute name="PartitionName">
${jboss.partition.name:DefaultPartition}
</attribute>
<! -- Bind address of bootstrap and HA-JNDI RMI endpoints -->
<attribute name="BindAddress">${jboss.bind.address}</attribute>
<! -- Port on which the HA-JNDI stub is made available -->
<attribute name="Port">1100</attribute>
<! -- RmiPort to be used by the HA-JNDI service once bound.
0 is for auto. -->
<attribute name="RmiPort">1101</attribute>
<! -- Accept backlog of the bootstrap socket -->
<attribute name="Backlog">50</attribute>
<! -- The thread pool service used to control the bootstrap and
auto discovery lookups -->
<depends optional-attribute-name="LookupPool"
proxy-type="attribute">jboss.system:service=ThreadPool</depends>
<! -- A flag to disable the auto discovery via multicast -->
<attribute name="DiscoveryDisabled">false</attribute>
<! -- Set the auto-discovery bootstrap multicast bind address. -->
<attribute name="AutoDiscoveryBindAddress">
${jboss.bind.address}
</attribute>
<! -- Multicast Address and group port used for auto-discovery -->
<attribute name="AutoDiscoveryAddress">
${jboss.partition.udpGroup:230.0.0.4}
</attribute>
<attribute name="AutoDiscoveryGroup">1102</attribute>
<! -- The TTL (time-to-live) for autodiscovery IP multicast packets -->
<attribute name="AutoDiscoveryTTL">16</attribute>
<! -- Client socket factory to be used for client-server
RMI invocations during JNDI queries
<attribute name="ClientSocketFactory">custom</attribute>
-->
<! -- Server socket factory to be used for client-server
RMI invocations during JNDI queries
<attribute name="ServerSocketFactory">custom</attribute>
-->
</mbean>
It is possible to start several HA-JNDI services that use different clusters. This can be used, for example, if a node is part of many clusters. In this case, make sure that you set a different port or IP address for both services. For instance, if you wanted to hook up HA-JNDI to the example cluster you set up and change the binding port, the Mbean descriptor would look as follows.
<mbean code="org.jboss.ha.jndi.HANamingService"
name="jboss:service=HAJNDI">
<depends>jboss:service=MySpecialPartition</depends>
<attribute name="PartitionName">MySpecialPartition</attribute>
<attribute name="Port">56789</attribute>
</mbean>
1.3. Clustered Session EJBs
Session EJBs provide remote invocation services. They are clustered based on the client-side interceptor architecture. The client application for a clustered session bean is exactly the same as the client for the non-clustered version of the session bean, except for a minor change to the java.naming.provier.url system property to enable HA-JNDI lookup (see previous section). No code change or re-compilation is needed on the client side. Now, let’s check out how to configure clustered session beans in EJB 2.x and EJB 3.0 server applications respectively.
1.3.1. Stateless Session Bean in EJB 2.x
Clustering stateless session beans is most probably the easiest case: as no state is involved, calls can be load-balanced on any participating node (i.e. any node that has this specific bean deployed) of the cluster. To make a bean clustered, you need to modify its jboss.xml descriptor to contain a <clustered> tag.
<jboss>
<enterprise-beans>
<session>
<ejb-name>nextgen.StatelessSession</ejb-name>
<jndi-name>nextgen.StatelessSession</jndi-name>
<clustered>True</clustered>
<cluster-config>
<partition-name>DefaultPartition</partition-name>
<home-load-balance-policy>
org.jboss.ha.framework.interfaces.RoundRobin
</home-load-balance-policy>
<bean-load-balance-policy>
org.jboss.ha.framework.interfaces.RoundRobin
</bean-load-balance-policy>
</cluster-config>
</session>
</enterprise-beans>
</jboss>
Note
The <clustered>True</clustered> element is really just an alias for the <configuration-name>Clustered Stateless SessionBean</configuration-name> element.
In the bean configuration, only the <clustered> element is mandatory. It indicates that the bean works in a cluster. The <cluster-config> element is optional and the default values of its attributes are indicated in the sample configuration above. Below is a description of the attributes in the <cluster-config> element.
-
partition-name specifies the name of the cluster the bean participates in. The default value is DefaultPartition. The default partition name can also be set system-wide using the jboss.partition.name system property.
-
home-load-balance-policy indicates the class to be used by the home stub to balance calls made on the nodes of the cluster. By default, the proxy will load-balance calls in a RoundRobin fashion. You can also implement your own load-balance policy class or use the class FirstAvailable that persists to use the first node available that it meets until it fails.
-
bean-load-balance-policy Indicates the class to be used by the bean stub to balance calls made on the nodes of the cluster. Comments made for the home-load-balance-policy attribute also apply.
In JBoss 3.0.x, each client-side stub has its own list of available target nodes. Consequently, some side-effects can occur. For example, if you cache your home stub and re-create a remote stub for a stateless session bean (with the Round-Robin policy) each time you need to make an invocation, a new remote stub, containing the list of available targets, will be downloaded for each invocation. Consequently, as the first target node is always the first in the list, calls will not seemed to be load-balanced because there is no usage-history between different stubs. In JBoss 3.2+, the proxy families (i.e., the “First AvailableIdenticalAllProxies” load balancing policy, see Section 1.1.3.2, “JBoss AS 3.2+”) remove this side effect as the home and remote stubs of a given EJB are in two different families.
1.3.1.1. Handle Cluster Restart
We have covered the HA smart client architecture in Section 1.1.2.1, “Client-side interceptor”. The default HA smart proxy client can only failover as long as one node in the cluster exists. If there is a complete cluster shutdown, the proxy becomes orphanned and looses knowledge of the available nodes in the cluster. There is no way for the proxy to recover from this. The proxy needs to be looked up out of JNDI/HAJNDI when the nodes are restarted.
The 3.2.7+/4.0.2+ releases contain a RetryInterceptor that can be added to the proxy client side interceptor stack to allow for a transparent recovery from such a restart failure. To enable it for an EJB, setup an invoker-proxy-binding that includes the RetryInterceptor. Below is an example jboss.xml configuration.
<jboss>
<session>
<ejb-name>nextgen_RetryInterceptorStatelessSession</ejb-name>
<invoker-bindings>
<invoker>
<invoker-proxy-binding-name>
clustered-retry-stateless-rmi-invoker
</invoker-proxy-binding-name>
<jndi-name>
nextgen_RetryInterceptorStatelessSession
</jndi-name>
</invoker>
</invoker-bindings>
<clustered>true</clustered>
</session>
<invoker-proxy-binding>
<name>clustered-retry-stateless-rmi-invoker</name>
<invoker-mbean>jboss:service=invoker,type=jrmpha</invoker-mbean>
<proxy-factory>org.jboss.proxy.ejb.ProxyFactoryHA</proxy-factory>
<proxy-factory-config>
<client-interceptors>
<home>
<interceptor>
org.jboss.proxy.ejb.HomeInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.SecurityInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.TransactionInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.ejb.RetryInterceptor
</interceptor>
<interceptor>
org.jboss.invocation.InvokerInterceptor
</interceptor>
</home>
<bean>
<interceptor>
org.jboss.proxy.ejb.StatelessSessionInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.SecurityInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.TransactionInterceptor
</interceptor>
<interceptor>
org.jboss.proxy.ejb.RetryInterceptor
</interceptor>
<interceptor>
org.jboss.invocation.InvokerInterceptor
</interceptor>
</bean>
</client-interceptors>
</proxy-factory-config>
</invoker-proxy-binding>
1.3.2. Stateful Session Bean in EJB 2.x
Clustering stateful session beans is more complex than clustering their stateless counterparts since JBoss needs to manage the state information. The state of all stateful session beans are replicated and synchronized across the cluster each time the state of a bean changes. The JBoss AS uses the HASessionState MBean to manage distributed session states for clustered EJB 2.x stateful session beans. In this section, we cover both the session bean configuration and the HASessionState MBean configuration.
1.3.2.1. The EJB application configuration
In the EJB application, you need to modify the jboss.xml descriptor file for each stateful session bean and add the <clustered> tag.
<jboss>
<enterprise-beans>
<session>
<ejb-name>nextgen.StatefulSession</ejb-name>
<jndi-name>nextgen.StatefulSession</jndi-name>
<clustered>True</clustered>
<cluster-config>
<partition-name>DefaultPartition</partition-name>
<home-load-balance-policy>
org.jboss.ha.framework.interfaces.RoundRobin
</home-load-balance-policy>
<bean-load-balance-policy>
org.jboss.ha.framework.interfaces.FirstAvailable
</bean-load-balance-policy>
<session-state-manager-jndi-name>
/HASessionState/Default
</session-state-manager-jndi-name>
</cluster-config>
</session>
</enterprise-beans>
</jboss>
In the bean configuration, only the <clustered> tag is mandatory to indicate that the bean works in a cluster. The <cluster-config> element is optional and its default attribute values are indicated in the sample configuration above.
The <session-state-manager-jndi-name> tag is used to give the JNDI name of the HASessionState service to be used by this bean.
The description of the remaining tags is identical to the one for stateless session bean. Actions on the clustered stateful session bean’s home interface are by default load-balanced, round-robin. Once the bean’s remote stub is available to the client, calls will not be load-balanced round-robin any more and will stay “sticky” to the first node in the list.
1.3.2.2. Optimize state replication
As the replication process is a costly operation, you can optimise this behaviour by optionally implementing in your bean class a method with the following signature:
public boolean isModified ();
Before replicating your bean, the container will detect if your bean implements this method. If your bean does, the container calls the isModified() method and it only replicates the bean when the method returns true. If the bean has not been modified (or not enough to require replication, depending on your own preferences), you can return false and the replication would not occur. This feature is available on JBoss AS 3.0.1+ only.
1.3.2.3. The HASessionState service configuration
The HASessionState service MBean is defined in the all/deploy/cluster-service.xml file.
<mbean code="org.jboss.ha.hasessionstate.server.HASessionStateService"
name="jboss:service=HASessionState">
<depends>
jboss:service=${jboss.partition.name:DefaultPartition}
</depends>
<!-- Name of the partition to which the service is linked -->
<attribute name="PartitionName">
${jboss.partition.name:DefaultPartition}
</attribute>
<!-- JNDI name under which the service is bound -->
<attribute name="JndiName">/HASessionState/Default</attribute>
<!-- Max delay before cleaning unreclaimed state.
Defaults to 30*60*1000 => 30 minutes -->
<attribute name="BeanCleaningDelay">0</attribute>
</mbean>
The configuration attributes in the HASessionState MBean are listed below.
-
JndiName is an optional attribute to specify the JNDI name under which this HASessionState service is bound. The default value is /HAPartition/Default.
-
PartitionName is an optional attribute to specify the name of the cluster in which the current HASessionState protocol will work. The default value is DefaultPartition.
-
BeanCleaningDelay is an optional attribute to specify the number of miliseconds after which the HASessionState service can clean a state that has not been modified. If a node, owning a bean, crashes, its brother node will take ownership of this bean. Nevertheless, the container cache of the brother node will not know about it (because it has never seen it before) and will never delete according to the cleaning settings of the bean. That is why the HASessionState service needs to do this cleanup sometimes. The default value is 30*60*1000 milliseconds (i.e., 30 minutes).
1.3.3. Stateless Session Bean in EJB 3.0
To cluster a stateless session bean in EJB 3.0, all you need to do is to annotate the bean class withe the @Cluster annotation. You can pass in the load balance policy and cluster partition as parameters to the annotation. The default load balance policy is org.jboss.ha.framework.interfaces.RandomRobin and the default cluster is DefaultPartition. Below is the definition of the @Cluster annotation.
public @interface Clustered {
Class loadBalancePolicy() default LoadBalancePolicy.class;
String partition() default "DefaultPartition";
}
Here is an example of a clustered EJB 3.0 stateless session bean implementation.
@Stateless
@Clustered
public class MyBean implements MySessionInt {
public void test() {
// Do something cool
}
}
1.3.4. Stateful Session Bean in EJB 3.0
To cluster stateful session beans in EJB 3.0, you need to tag the bean implementation class with the @Cluster annotation, just as we did with the EJB 3.0 stateless session bean earlier.
@Stateful
@Clustered
public class MyBean implements MySessionInt {
private int state = 0;
public void increment() {
System.out.println("counter: " + (state++));
}
}
JBoss Cache provides the session state replication service for EJB 3.0 stateful session beans. The related MBean service is defined in the ejb3-clustered-sfsbcache-service.xml file in the deploy directory. The contents of the file are as follows.
<server>
<mbean code="org.jboss.ejb3.cache.tree.PassivationTreeCache"
name="jboss.cache:service=EJB3SFSBClusteredCache">
<attribute name="IsolationLevel">READ_UNCOMMITTED</attribute>
<attribute name="CacheMode">REPL_SYNC</attribute>
<attribute name="ClusterName">SFSB-Cache</attribute>
<attribute name="ClusterConfig">
... ...
</attribute>
<!-- Number of milliseconds to wait until all responses for a
synchronous call have been received.
-->
<attribute name="SyncReplTimeout">10000</attribute>
<!-- Max number of milliseconds to wait for a lock acquisition -->
<attribute name="LockAcquisitionTimeout">15000</attribute>
<!-- Name of the eviction policy class. -->
<attribute name="EvictionPolicyClass">
org.jboss.ejb3.cache.tree.StatefulEvictionPolicy
</attribute>
<!-- Specific eviction policy configurations. This is LRU -->
<attribute name="EvictionPolicyConfig">
<config>
<attribute name="wakeUpIntervalSeconds">1</attribute>
<name>statefulClustered</name>
<region name="/_default_">
<attribute name="maxNodes">1000000</attribute>
<attribute name="timeToIdleSeconds">300</attribute>
</region>
</config>
</attribute>
<attribute name="CacheLoaderFetchPersistentState">false</attribute>
<attribute name="CacheLoaderFetchTransientState">true</attribute>
<attribute name="FetchStateOnStartup">true</attribute>
<attribute name="CacheLoaderClass">
org.jboss.ejb3.cache.tree.StatefulCacheLoader
</attribute>
<attribute name="CacheLoaderConfig">
location=statefulClustered
</attribute>
</mbean>
</server>
The configuration attributes in the PassivationTreeCache MBean are essentially the same as the attributes in the standard JBoss Cache TreeCache MBean discussed in Chapter 2, JBossCache and JGroups Services. Again, we omitted the JGroups configurations in the ClusterConfig attribute (see more in Section 2.1, “JGroups Configuration”).
1.4. Clustered Entity EJBs
In a JBoss AS cluster, the entity bean instances need to replicated across all nodes. If an entity bean provides remote services, the service methods need to be load balanced as well.
To use a clustered entity bean, the application does not need to do anything special, except for looking up bean references from the clustered HA-JNDI.
1.4.1. Entity Bean in EJB 2.x
First of all, it is worth to note that clustering 2.x entity beans is a bad thing to do. Its exposes elements that generally are too fine grained for use as remote objects to clustered remote objects and introduces data synchronization problems that are non-trivial. Do NOT use EJB 2.x entity bean clustering unless you fit into the sepecial case situation of read-only, or one read-write node with read-only nodes synched with the cache invalidation services.
To cluster EJB 2.x entity beans, you need to add the <clustered> element to the application’s jboss.xml descriptor file. Below is a typical jboss.xml file.
<jboss>
<enterprise-beans>
<entity>
<ejb-name>nextgen.EnterpriseEntity</ejb-name>
<jndi-name>nextgen.EnterpriseEntity</jndi-name>
<clustered>True</clustered>
<cluster-config>
<partition-name>DefaultPartition</partition-name>
<home-load-balance-policy>
org.jboss.ha.framework.interfaces.RoundRobin
</home-load-balance-policy>
<bean-load-balance-policy>
org.jboss.ha.framework.interfaces.FirstAvailable
</bean-load-balance-policy>
</cluster-config>
</entity>
</enterprise-beans>
</jboss>
The EJB 2.x entity beans are clustered for load balanced remote invocations. All the bean instances are synchronized to have the same contents on all nodes.
However, clustered EJB 2.x Entity Beans do not have a distributed locking mechanism or a distributed cache. They can only be synchronized by using row-level locking at the database level (see <row-lock> in the CMP specification) or by setting the Transaction Isolation Level of your JDBC driver to be TRANSACTION_SERIALIZABLE. Because there is no supported distributed locking mechanism or distributed cache Entity Beans use Commit Option “B” by default (See standardjboss.xml and the container configurations Clustered CMP 2.x EntityBean, Clustered CMP EntityBean, or Clustered BMP EntityBean). It is not recommended that you use Commit Option “A” unless your Entity Bean is read-only. (There are some design patterns that allow you to use Commit Option “A” with read-mostly beans. You can also take a look at the Seppuku pattern http://dima.dhs.org/misc/readOnlyUpdates.html. JBoss may incorporate this pattern into later versions.)
Note
If you are using Bean Managed Persistence (BMP), you are going to have to implement synchronization on your own. The MVCSoft CMP 2.0 persistence engine (see http://www.jboss.org/jbossgroup/partners.jsp) provides different kinds of optimistic locking strategies that can work in a JBoss cluster.
1.4.2. Entity Bean in EJB 3.0
In EJB 3.0, the entity beans primarily serve as a persistence data model. They do not provide remote services. Hence, the entity bean clustering service in EJB 3.0 primarily deals with distributed caching and replication, instead of load balancing.
1.4.2.1. Configure the distributed cache
To avoid round trips to the database, you can use a cache for your entities. JBoss EJB 3.0 is implemented by Hibernate, which has support for a second-level cache. The Hibernate setup used for the JBoss EJB 3.0 implementation uses JBoss Cache as its underlying cache implementation. The cache provides the following functionalities.
-
If you persist a cache enabled entity bean instance to the database via the entity manager the entity will inserted into the cache.
-
If you update an entity bean instance and save the changes to the database via the entity manager the entity will updated in the cache.
-
If you remove an entity bean instance from the database via the entity manager the entity will removed from the cache.
-
If loading a cached entity from the database via the entity manager, and that entity does not exist in the database, it will be inserted into the cache.
JBoss Cache service for EJB 3.0 entity beans is configured in a TreeCache MBean (see Section 2.2, “JBossCache Configuration”) in the deploy/ejb3-entity-cache-service.xml file. The name of the cache MBean service is jboss.cache:service=EJB3EntityTreeCache. Below is the contents of the ejb3-entity-cache-service.xml file in the standard JBoss distribution. Again, we omitted the JGroups configuration element ClusterConfig.
<server>
<mbean code="org.jboss.cache.TreeCache"
name="jboss.cache:service=EJB3EntityTreeCache">
<depends>jboss:service=Naming</depends>
<depends>jboss:service=TransactionManager</depends>
<!-- Configure the TransactionManager -->
<attribute name="TransactionManagerLookupClass">
org.jboss.cache.JBossTransactionManagerLookup
</attribute>
<attribute name="IsolationLevel">REPEATABLE_READ</attribute>
<attribute name="CacheMode">REPL_SYNC</attribute>
<!--Name of cluster. Needs to be the same for all clusters,
in order to find each other -->
<attribute name="ClusterName">EJB3-entity-cache</attribute>
<attribute name="ClusterConfig">
... ...
</attribute>
<attribute name="InitialStateRetrievalTimeout">5000</attribute>
<attribute name="SyncReplTimeout">10000</attribute>
<attribute name="LockAcquisitionTimeout">15000</attribute>
<attribute name="EvictionPolicyClass">
org.jboss.cache.eviction.LRUPolicy
</attribute>
<!-- Specific eviction policy configurations. This is LRU -->
<attribute name="EvictionPolicyConfig">
<config>
<attribute name="wakeUpIntervalSeconds">5</attribute>
<!-- Cache wide default -->
<region name="/_default_">
<attribute name="maxNodes">5000</attribute>
<attribute name="timeToLiveSeconds">1000</attribute>
</region>
</config>
</attribute>
</mbean>
</server>
As we discussed in Section 2.2, “JBossCache Configuration”, JBoss Cache allows you to specify timeouts to cached entities. Entities not accessed within a certain amount of time are dropped from the cache in order to save memory. If running within a cluster, and the cache is updated, changes to the entries in one node will be replicated to the corresponding entries in the other nodes in the cluster.
Now, we have JBoss Cache configured to support distributed caching of EJB 3.0 entity beans. We still have to configure individual entity beans to use the cache service.
1.4.2.2. Configure the entity beans for cache
You define your entity bean classes the normal way. Future versions of JBoss EJB 3.0 will support annotating entities and their relationship collections as cached, but for now you have to configure the underlying hibernate engine directly. Take a look at the persistence.xml file, which configures the caching options for hibernate via its optional property elements. The following element in persistence.xml defines that caching should be enabled:
<!-- Clustered cache with TreeCache -->
<property name="cache.provider_class">
org.jboss.ejb3.entity.TreeCacheProviderHook
</property>
The following property element defines the object name of the cache to be used, and the MBean name.
<property name="treecache.mbean.object_name">
jboss.cache:service=EJB3EntityTreeCache
</property>
Next we need to configure what entities be cached. The default is to not cache anything, even with the settings shown above. We use the @Cache annotation to tag entity beans that needs to be cached.
@Entity
@Cache(usage=CacheConcurrencyStrategy.TRANSACTIONAL)
public class Customer implements Serializable {
// ... ...
}
A very simplified rule of thumb is that you will typically want to do caching for objects that rarely change, and which are frequently read. You can fine tune the cache for each entity bean in the ejb3-entity-cache-service.xml configuration file. For instance, you can specify the size of the cache. If there are too many objects in the cache, the cache could evict oldest objects (or least used objects, depending on configuration) to make room for new objects. The cache for the mycompany.Customer entity bean is /mycompany/Customer cache region.
<server>
<mbean code="org.jboss.cache.TreeCache"
name="jboss.cache:service=EJB3EntityTreeCache">
<depends>jboss:service=Naming
<depends>jboss:service=TransactionManager
... ...
<attribute name="EvictionPolicyConfig">
<config>
<attribute name="wakeUpIntervalSeconds">5</attribute>
<region name="/_default_">
<attribute name="maxNodes">5000</attribute>
<attribute name="timeToLiveSeconds">1000</attribute>
</region>
<region name="/mycompany/Customer">
<attribute name="maxNodes">10</attribute>
<attribute name="timeToLiveSeconds">5000</attribute>
</region>
... ...
</config>
</attribute>
</mbean>
</server>
If you do not specify a cache region for an entity bean class, all instances of this class will be cached in the /_default region as defined above. The EJB3 Query API provides means for you to save to load query results (i.e., collections of entity beans) from specified cache regions.
HTTP session replication is used to replicate the state associated with your web clients on other nodes of a cluster. Thus, in the event one of your node crashes, another node in the cluster will be able to recover. Two distinct functions must be performed:
State replication is directly handled by JBoss. When you run JBoss in the all configuration, session state replication is enabled by default. Just deploy your web application and its session state is already replicated across all JBoss instances in the cluster.
However, Load-balancing is a different story, it is not handled by JBoss itself and requires additional software. As a very common scenario, we will demonstrate how to setup Apache and mod_jk. This activity could be either performed by specialized hardware switches or routers (Cisco LoadDirector for example) or any other dedicated software though.
Note
A load-balancer tracks the HTTP requests and, depending on the session to which is linked the request, it dispatches the request to the appropriate node. This is called a load-balancer with sticky-sessions: once a session is created on a node, every future request will also be processed by the same node. Using a load-balancer that supports sticky-sessions without replicating the sessions allows you to scale very well without the cost of session state replication: each query will always be handled by the same node. But in the case a node dies, the state of all client sessions hosted by this node are lost (the shopping carts, for example) and the clients will most probably need to login on another node and restart with a new session. In many situations, it is acceptable not to replicate HTTP sessions because all critical state is stored in the database. In other situations, loosing a client session is not acceptable and, in this case, session state replication is the price one has to pay.
Apache is a well-known web server which can be extended by plugging modules. One of these modules, mod_jk (and the newest mod_jk2) has been specifically designed to allow forward requests from Apache to a Servlet container. Furthermore, it is also able to load-balance HTTP calls to a set of Servlet containers while maintaining sticky sessions, and this is what is actually interesting for us.
1.5.1. Download the software
First of all, make sure that you have Apache installed. You can download Apache directly from Apache web site at http://httpd.apache.org/. Its installation is pretty straightforward and requires no specific configuration. As several versions of Apache exist, we advise you to use version 2.0.x. We will consider, for the next sections, that you have installed Apache in the APACHE_HOME directory.
Next, download mod_jk binaries. Several versions of mod_jk exist as well. We strongly advise you to use mod_jk 1.2.x, as both mod_jk and mod_jk2 are deprecated, unsupported and no further developments are going on in the community. The mod_jk 1.2.x binary can be downloaded from http://www.apache.org/dist/jakarta/tomcat-connectors/jk/binaries/. Rename the downloaded file to mod_jk.so and copy it under APACHE_HOME/modules/.
1.5.2. Configure Apache to load mod_jk
Modify APACHE_HOME/conf/httpd.conf and add a single line at the end of the file:
# Include mod_jk's specific configuration file
Include conf/mod-jk.conf
Next, create a new file named APACHE_HOME/conf/mod-jk.conf:
# Load mod_jk module
# Specify the filename of the mod_jk lib
LoadModule jk_module modules/mod_jk.so
# Where to find workers.properties
JkWorkersFile conf/workers.properties
# Where to put jk logs
JkLogFile logs/mod_jk.log
# Set the jk log level [debug/error/info]
JkLogLevel info
# Select the log format
JkLogStampFormat "[%a %b %d %H:%M:%S %Y]"
# JkOptions indicates to send SSK KEY SIZE
JkOptions +ForwardKeySize +ForwardURICompat -ForwardDirectories
# JkRequestLogFormat
JkRequestLogFormat "%w %V %T"
# Mount your applications
JkMount /application/* loadbalancer
# You can use external file for mount points.
# It will be checked for updates each 60 seconds.
# The format of the file is: /url=worker
# /examples/*=loadbalancer
JkMountFile conf/uriworkermap.properties
# Add shared memory.
# This directive is present with 1.2.10 and
# later versions of mod_jk, and is needed for
# for load balancing to work properly
JkShmFile logs/jk.shm
# Add jkstatus for managing runtime data
<Location /jkstatus/>
JkMount status
Order deny,allow
Deny from all
Allow from 127.0.0.1
</Location>
Please note that two settings are very important:
-
The LoadModule directive must reference the mod_jk library you have downloaded in the previous section. You must indicate the exact same name with the “modules” file path prefix.
-
The JkMount directive tells Apache which URLs it should forward to the mod_jk module (and, in turn, to the Servlet containers). In the above file, all requests with URL path /application/* are sent to the mod_jk load-balancer. This way, you can configure Apache to server static contents (or PHP contents) directly and only use the loadbalancer for Java applications. If you only use mod_jk as a loadbalancer, you can also forward all URLs (i.e., /*) to mod_jk.
In addition to the JkMount directive, you can also use the JkMountFile directive to specify a mount points configuration file, which contains multiple Tomcat forwarding URL mappings. You just need to create a uriworkermap.properties file in the APACHE_HOME/conf directory. The format of the file is /url=worker_name. To get things started, paste the following example into the file you created:
# Simple worker configuration file
# Mount the Servlet context to the ajp13 worker
/jmx-console=loadbalancer
/jmx-console/*=loadbalancer
/web-console=loadbalancer
/web-console/*=loadbalancer
This will configure mod_jk to forward requests to /jmx-console and /web-console to Tomcat.
You will most probably not change the other settings in mod_jk.conf. They are used to tell mod_jk where to put its logging file, which logging level to use and so on.
1.5.3. Configure worker nodes in mod_jk
Next, you need to configure mod_jk workers file conf/workers.properties. This file specify where are located the different Servlet containers and how calls should be load-balanced across them. The configuration file contains one section for each target servlet container and one global section. For a two nodes setup, the file could look like this:
# Define list of workers that will be used
# for mapping requests
worker.list=loadbalancer,status
# Define Node1
# modify the host as your host IP or DNS name.
worker.node1.port=8009
worker.node1.host=node1.mydomain.com
worker.node1.type=ajp13
worker.node1.lbfactor=1
worker.node1.cachesize=10
# Define Node2
# modify the host as your host IP or DNS name.
worker.node2.port=8009
worker.node2.host= node2.mydomain.com
worker.node2.type=ajp13
worker.node2.lbfactor=1
worker.node2.cachesize=10
# Load-balancing behaviour
worker.loadbalancer.type=lb
worker.loadbalancer.balance_workers=node1,node2
worker.loadbalancer.sticky_session=1
#worker.list=loadbalancer
# Status worker for managing load balancer
worker.status.type=status
Basically, the above file configures mod_jk to perform weighted round-robin load balancing with sticky sessions between two servlet containers (JBoss Tomcat) node1 and node2 listening on port 8009.
In the works.properties file, each node is defined using the worker.XXX naming convention where XXX represents an arbitrary name you choose for one of the target Servlet container. For each worker, you must give the host name (or IP address) and port number of the AJP13 connector running in the Servlet container.
The lbfactor attribute is the load-balancing factor for this specific worker. It is used to define the priority (or weight) a node should have over other nodes. The higher this number is, the more HTTP requests it will receive. This setting can be used to differentiate servers with different processing power.
The cachesize attribute defines the size of the thread pools associated to the Servlet container (i.e. the number of concurrent requests it will forward to the Servlet container). Make sure this number does not outnumber the number of threads configured on the AJP13 connector of the Servlet container. Please review http://jakarta.apache.org/tomcat/connectors-doc/config/workers.html for comments on cachesize for Apache 1.3.x.
The last part of the conf/workers.properties file defines the loadbalancer worker. The only thing you must change is the worker.loadbalancer.balanced_workers line: it must list all workers previously defined in the same file: load-balancing will happen over these workers.
The sticky_session property specifies the cluster behavior for HTTP sessions. If you specify worker.loadbalancer.sticky_session=0, each request will be load balanced between node1 and node2. But when a user opens a session on one server, it is a good idea to always forward this user’s requests to the same server. This is called a “sticky session”, as the client is always using the same server he reached on his first request. Otherwise the user’s session data would need to be synchronized between both servers (session replication, see Section 1.5.5, “Configure HTTP session state replication”). To enable session stickiness, you need to set worker.loadbalancer.sticky_session to 1.
Note
A non-loadbalanced setup with a single node required the worker.list=node1 entry before mod_jk would function correctly.
Finally, we must configure the JBoss Tomcat instances on all clustered nodes so that they can expect requests forwarded from the mod_jk loadbalancer.
On each clustered JBoss node, we have to name the node according to the name specified in workers.properties. For instance, on JBoss instance node1, edit the JBOSS_HOME/server/all/deploy/jbossweb-tomcat50.sar/server.xml file (replace /all with your own server name if necessary). Locate the <Engine> element and add an attribute jvmRoute:
<Engine name="jboss.web" defaultHost="localhost" jvmRoute="node1">
... ...
</Engine>
Then, for each JBoss Tomcat instance in the cluster, we need to tell it to add the jvmRoute value to its session cookies so that mod_jk can route incoming requests. Edit the JBOSS_HOME/server/all/deploy/jbossweb-tomcat50.sar/META-INF/jboss-service.xml file (replace /all with your own server name). Locate the <attribute> element with a name of UseJK, and set its value to true:
<attribute name="UseJK">true</attribute>
At this point, you have a fully working Apache+mod_jk load-balancer setup that will balance call to the Servlet containers of your cluster while taking care of session stickiness (clients will always use the same Servlet container).
Note
For more updated information on using mod_jk 1.2 with JBoss Tomcat, please refer to the JBoss wiki page at http://wiki.jboss.org/wiki/Wiki.jsp?page=UsingMod_jk1.2WithJBoss.
1.5.5. Configure HTTP session state replication
In Section 1.5.3, “Configure worker nodes in mod_jk”, we covered how to use sticky sessions to make sure that a client in a session always hits the same server node in order to maintain the session state. However, that is not an ideal solution. The load might be unevenly distributed over the nodes over time and if a node goes down, all its session data is lost. A better and more reliable solution is to replicate session data across all nodes in the cluster. This way, the client can hit any server node and obtain the same session states.
The jboss.cache:service=TomcatClusteringCache MBean makes use of JBoss Cache to provide HTTP session replication service to the HTTP load balancer in a JBoss Tomcat cluster. This MBean is defined in the deploy/tc5-cluster.sar/META-INF/jboss-service.xml file.
Note
Before AS 4.0.4 CR2, the HTTP session cache configuration file is the deploy/tc5-cluster-service.xml file. Please see AS 4.0.3 documentation for more details.
Below is a typical deploy/tc5-cluster.sar/META-INF/jboss-service.xml file. The configuration attributes in the TomcatClusteringCache MBean is very similar to those in Section 2.2, “JBossCache Configuration”.
<mbean code="org.jboss.cache.aop.TreeCacheAop"
name="jboss.cache:service=TomcatClusteringCache">
<depends>jboss:service=Naming</depends>
<depends>jboss:service=TransactionManager</depends>
<depends>jboss.aop:service=AspectDeployer</depends>
<attribute name="TransactionManagerLookupClass">
org.jboss.cache.BatchModeTransactionManagerLookup
</attribute>
<attribute name="IsolationLevel">REPEATABLE_READ</attribute>
<attribute name="CacheMode">REPL_ASYNC</attribute>
<attribute name="ClusterName">
Tomcat-${jboss.partition.name:Cluster}
</attribute>
<attribute name="UseMarshalling">false</attribute>
<attribute name="InactiveOnStartup">false</attribute>
<attribute name="ClusterConfig">
... ...
</attribute>
<attribute name="LockAcquisitionTimeout">15000</attribute>
</mbean>
The detailed configuration for the TreeCache MBean is covered in Section 2.2, “JBossCache Configuration”. Below, we will just discuss several attributes that are most relevant to the HTTP cluster session replication.
-
TransactionManagerLookupClass sets the transaction manager factory. The default value is org.jboss.cache.BatchModeTransactionManagerLookup. It tells the cache NOT to participate in JTA-specific transactions. Instead, the cache manages its own transaction to support finely grained replications.
-
IsolationLevel sets the isolation level for updates to the transactional distributed cache. The valid values are SERIALIZABLE, REPEATABLE_READ, READ_COMMITTED, READ_UNCOMMITTED, and NONE. These isolation levels mean the same thing as isolation levels on the database. The default isolation of REPEATABLE_READ makes sense for most web applications.
-
CacheMode controls how the cache is replicated. The valid values are REPL_SYNC and REPL_ASYNC, which determine whether changes are made synchronously or asynchronously. Using synchronous replication makes sure changes propagated to the cluster before the web request completes. However, synchronous replication is much slower. For asyncrhonous access, you will want to enable and tune the replication queue.
-
ClusterName specifies the name of the cluster that the cache works within. The default cluster name is the the word “Tomcat-” appended by the current JBoss partition name. All the nodes should use the same cluster name. Although session replication can share the same channel (multicast address and port) with other clustered services in JBoss, replication should have it’s own cluster name.
-
The UseMarshalling and InactiveOnStartup attributes must have the same value. They must be true if FIELD level session replication is needed (see later). Otherwise, they are default to false.
-
ClusterConfig configures the underlying JGroups stack. The most import configuration elements are the muliticast adress and port, mcast_addr and mcast_port respectively, to use for clustered communication. These values should make sense for your network. Please refer to Section 2.1, “JGroups Configuration” for more information.
-
LockAcquisitionTimeout sets the maximum number of milliseconds to wait for a lock acquisition. The default value is 15000.
-
UseReplQueue determines whether to enable the replication queue when using asynchronous replication. This allows multiple cache updates to be bundled together to improve performance. The replication queue properties are controlled by the ReplQueueInterval and ReplQueueMaxElements properties.
-
ReplQueueInterval specifies the time in milliseconds JBoss Cache will wait before sending items in the replication queue.
-
ReplQueueMaxElements: specifies the maximum number of elements allowed in the replication queue before JBoss Cache will send an update.
1.5.6. Enabling session replication in your application
To enable clustering of your web application you must it as distributable in the web.xml descriptor. Here’s an example:
<?xml version="1.0"?>
<web-app xmlns="http://java.sun.com/xml/ns/j2ee"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://java.sun.com/xml/ns/j2ee
http://java.sun.com/xml/ns/j2ee/web-app_2_4.xsd"
version="2.4">
<distributable/>
<!-- ... -->
</web-app>
You can futher configure session replication using the replication-config element in the jboss-web.xml file. Here is an example:
<jboss-web>
<replication-config>
<replication-trigger>SET_AND_NON_PRIMITIVE_GET</replication-trigger>
<replication-granularity>SESSION</replication-granularity>
<replication-field-batch-mode>true</replication-field-batch-mode>
</replication-config>
</jboss-web>
The replication-trigger element determines what triggers a session replication (or when is a session is considered dirty). It has 4 options:
-
SET: With this policy, the session is considered dirty only when an attribute is set in the session. If your application always writes changed value back into the session, this option will be most optimized in term of performance. If an object is retrieved from the session and modified without being written back into the session, the change to that object will not be replicated.
-
SET_AND_GET: With this policy, any attribute that is get or set will be marked as dirty. If an object is retrieved from the session and modified without being written back into the session, the change to that object will be replicated. This option can have significant performance implications.
-
SET_AND_NON_PRIMITIVE_GET: This policy is similar to the SET_AND_GET policy except that only non-primitive get operations are considered dirty. For example, the http session request may retrieve a non-primitive object instance from the attribute and then modify the instance. If we don’t specify that non-primitive get is considered dirty, then the modification will not be replication properly. This is the default value.
-
ACCESS: This option causes the session to be marked as dirty whenever it is accessed. Since a the session is accessed during each HTTP request, it will be replicated with each request. The access time stamp in the session instance will be updated as well. Since the time stamp may not be updated in other clustering nodes because of no replication, the session in other nodes may expire before the active node if the HTTP request does not retrieve or modify any session attributes. When this option is set, the session timestamps will be synchronized throughout the cluster nodes. Note that use of this option can have a significant performance impact, so use it with caution.
The replication-granularity element controls the size of the replication units. The supported values are:
-
SESSION: Replication is per session instance. As long as it is considered modified when the snapshot manager is called, the whole session object will be serialized.
-
ATTRIBUTE: Replication is only for the dirty attributes in the session plus some session data, like, lastAccessTime. For session that carries large amount of data, this option can increase replication performance.
-
FIELD: Replication is only for data fields inside session attribute objects (see more later).
The replication-field-batch-mode element indicates whether you want to have batch update between each http request or not. Default is true.
If your sessions are generally small, SESSION is the better policy. If your session is larger and some parts are infrequently accessed, ATTRIBUTE replication will be more effective. If your application has very big data objects in session attributes and only fields in those objects are frequently modified, the FIELD policy would be the best. In the next section, let’s discuss exactly how the FIELD level replication works.
1.5.7. Use FIELD level replication
FIELD-level replication only replicates modified data fields inside objects stored in the session. It could potentially drastically reduce the data traffic between clustered nodes, and hence improve the performance of the whole cluster. To use FIELD-level replication, you have to first prepare your Java class to indicate which fields are to be replicated. This is done via JDK 1.4 style annotations embedded in JavaDocs:
To annotate your POJO, we provide two annotations: @@org.jboss.web.tomcat.tc5.session.AopMarker and @@org.jboss.web.tomcat.tc5.session.InstanceAopMarker. When you annotate your class with AopMarker, you indicate that instances of this class will be used in FIELD-level replication. For exmaple,
/*
* My usual comments here first.
* @@org.jboss.web.tomcat.tc5.session.AopMarker
*/
public class Address
{
...
}
If you annotate it with InstanceAopMarker instead, then all of its sub-class will be automatically annotated as well. For example,
/*
*
* @@org.jboss.web.tomcat.tc5.session.InstanceOfAopMarker
*/
public class Person
{
...
}
then when you have a sub-class like
public class Student extends Person
{
...
}
there will be no need to annotate Student. It will be annotated automatically because it is a sub-class of Person.
However, since we only support JDK 1.4 style annotation (provided by JBoss Aop) now, you will need to perform a pre-processing step. You need to use the JBoss AOP pre-compiler annotationc and post-compiler aopc to process the above source code before and after they are compiled by the Java compiler. Here is an example on how to invoke those commands from command line.
$ annotationc [classpath] [source files or directories]
$ javac -cp [classpath] [source files or directories]
$ aopc [classpath] [class files or directories]
Please see the JBoss AOP documentation for the usage of the pre- and post-compiler. The JBoss AOP project also provides easy to use ANT tasks to help integrate those steps into your application build process. In the next AS release, JDK 5.0 annotation support will be provided for greater transparency. But for now, it is important that you perform the pre- and post-compilation steps for your source code.
Note
Or, you can see a complete example on how to build, deploy, and validate a FIELD-level replicated web application from this page: http://wiki.jboss.org/wiki/Wiki.jsp?page=Http_session_field_level_example. The example bundles the pre- and post-compile tools so you do not need to download JBoss AOP separately.
When you deploy the web application into JBoss AS, make sure that the following configurations are correct:
-
In the server’s deploy/tc5-cluster.sar/META-INF/jboss-service.xml file, the inactiveOnStartup and useMarshalling attributes must both be true.
-
In the application’s jboss-web.xml file, the replication-granularity attribute must be FIELD.
Finally, let’s see an example on how to use FIELD-level replication on those data classes. Notice that there is no need to call session.setAttribute() after you make changes to the data object, and all changes to the fields are automatically replicated across the cluster.
// Do this only once. So this can be in init(), e.g.
if(firstTime)
{
Person joe = new Person("Joe", 40);
Person mary = new Person("Mary", 30);
Address addr = new Address();
addr.setZip(94086);
joe.setAddress(addr);
mary.setAddress(addr); // joe and mary share the same address!
session.setAttribute("joe", joe); // that's it.
session.setAttribute("mary", mary); // that's it.
}
Person mary = (Person)session.getAttribute("mary");
mary.getAddress().setZip(95123); // this will update and replicate the zip code.
Besides plain objects, you can also use regular Java collections of those objects as session attributes. JBoss cache automatically figures out how to handle those collections and replicate field changes in their member objects.
1.5.8. Monitoring session replication
If you have deployed and accessed your application, go to the jboss.cache:service=TomcatClusteringCache MBean and invoke the printDetails operation. You should see output resembling the following.
/JSESSION
/quote
/FB04767C454BAB3B2E462A27CB571330
VERSION: 6
FB04767C454BAB3B2E462A27CB571330: org.jboss.invocation.MarshalledValue@1f13a81c
/AxCI8Ovt5VQTfNyYy9Bomw**
VERSION: 4
AxCI8Ovt5VQTfNyYy9Bomw**: org.jboss.invocation.MarshalledValue@e076e4c8
This output shows two separate web sessions, in one application named quote, that are being shared via JBossCache. This example uses a replication-granularity of session. Had attribute level replication been used, there would be additional entries showing each replicated session attribute. In either case, the replicated values are stored in an opaque MarshelledValue container. There aren’t currently any tools that allow you to inspect the contents of the replicated session values. If you don’t see any output, either the application was not correctly marked as distributable or you haven’t accessed a part of application that places values in the HTTP session. The org.jboss.cache and org.jboss.web logging categories provide additional insight into session replication useful for debugging purposes.
1.5.9. Using Single Sign On
JBoss supports clustered single sign-on, allowing a user to authenticate to one application on a JBoss server and to be recognized on all applications, on that same machine or on another node in the cluster, that are deployed on the same virtual host. Authentication replication is handled by the HTTP session replication service. Although session replication does not need to be explicitly enabled for the applications in question, the tc5-cluster-service.xml file does need to be deployed.
To enable single sign-on, you must add the ClusteredSingleSignOn valve to the appropriate Host elements of the tomcat server.xml file. The valve configuration is shown here:
<Valve className="org.jboss.web.tomcat.tc5.sso.ClusteredSingleSignOn" />
1.6. Clustered JMS Services
JBoss AS 3.2.4 and above support high availability JMS (HA-JMS) services in the all server configuration. In the current production release of JBoss AS, the HA-JMS service is implemented as a clustered singleton fail-over service.
Note
If you are willing to configure HA-JMS yourself, you can get it to work with earlier versions of JBoss AS. We have a customer who uses HA-JMS successfully in JBoss AS 3.0.7. Please contact JBoss support for more questions.
1.6.1. High Availability Singleton Fail-over
The JBoss HA-JMS service (i.e., message queues and topics) only runs on a single node (i.e., the master node) in the cluster at any given time. If that node fails, the cluster simply elects another node to run the JMS service (fail-over). This setup provides redundancy against server failures but does not reduce the work load on the JMS server node.
Note
While you cannot load balance HA-JMS queues (there is only one master node that runs the queues), you can load balance the MDBs that process messages from those queues (see Section 1.6.1.3, “Load Balanced HA-JMS MDBs”).
1.6.1.1. Server Side Configuration
To use the singleton fail-over HA-JMS service, you must configure JMS services identically on all nodes in the cluster. That includes all JMS related service MBeans and all deployed JMS applications.
The JMS server is configured to persist its data in the DefaultDS. By default, that is the embedded HSQLDB. In most cluster environments, however, all nodes need to persist data against a shared database. So, the first thing to do before you start clustered JMS is to setup a shared database for JMS. You need to do the following:
-
Configure DefaultDS to point to the database server of your choice. That is to replace the deploy/hsqlsb-ds.xml file with the xxx-ds.xml file in the docs/examples/jca directory, where xxx is the name of the target shared database (e.g., mysql-ds.xml).
-
Replace the hsqldb-jdbc2-service.xml file under the server/all/deploy-hasingleton/jms directory with one tuned to the specific database. For example if you use MySQL the file is mysql-jdbc2-service.xml. Configuration files for a number of RDBMS are bundled with the JBoss AS distribution. They can be found under docs/examples/jms.
Note
There is no need to replace the hsqldb-jdbc-state-service.xml file under the server/all/deploy-hasingleton/jms directory. Despite the hsql in its name, it works with all SQL92 compliant databases, including HSQL, MySQL, SQL Server, and more. It automatically uses the DefaultDS for storage, as we configured above.
The HA-JMS client is different from regular JMS clients in two important aspects.
-
The HA-JMS client must obtain JMS connection factories from the HA-JNDI (the default port is 1100).
-
The client connection must listens for server exceptions. When the cluster fail-over to a different master node, all client operations on the current connection fails with exceptions. The client must know to re-connect.
Note
While the HA-JMS connection factory knows the current master node that runs JMS services, there is no smart client side interceptor. The client stub only knows the fixed master node and cannot adjust to server topography changes.
1.6.1.3. Load Balanced HA-JMS MDBs
While the HA-JMS queues and topics only run on a single node at a time, MDBs on multiple nodes can receive and process messages from the HA-JMS master node. The contested queues and topics result in load balancing behavior for MDBs. To enable loading balancing for MDBs, you can specify a receiver for the queue. The receiver records which node is waiting for a message and in which order the messages should be processed. JBoss provides three receiver implementations.
-
The org.jboss.mq.server.ReceiversImpl is the default implementation using a HashSet.
-
The org.jboss.mq.server.ReceiversImplArrayList is theimplementation using an ArrayList.
-
The org.jboss.mq.server.ReceiversImplLinkedList is the implementation using a LinkedList.
You can specify the receiver implementation class name as an attribute in the MBean that defines the permanent JMS Queue or DestinationManager on each node. For best load balancing performance, we suggest you to use the ReceiversImplArrayList or ReceiversImplArrayList implementations due to an undesirable implementation detail of HashSet in the JVM.
Chapter 2. JBossCache and JGroups Services
JGroups and JBossCache provide the underlying communication, node replication and caching services, for JBoss AS clusters. Those services are configured as MBeans. There is a set of JBossCache and JGroups MBeans for each type of clustering applications (e.g., the Stateful Session EJBs, the distributed entity EJBs etc.).
The JBoss AS ships with a reasonable set of default JGroups and JBossCache MBean configurations. Most applications just work out of the box with the default MBean configurations. You only need to tweak them when you are deploying an application that has special network or performance requirements.
2.1. JGroups Configuration
The JGroups framework provides services to enable peer-to-peer communications between nodes in a cluster. It is built on top a stack of network communication protocols that provide transport, discovery, reliability and failure detection, and cluster membership management services. Figure 2.1, “Protocol stack in JGroups” shows the protocol stack in JGroups.
Figure 2.1. Protocol stack in JGroups
JGroups configurations often appear as a nested attribute in cluster related MBean services, such as the PartitionConfig attribute in the ClusterPartition MBean or the ClusterConfig attribute in the TreeCache MBean. You can configure the behavior and properties of each protocol in JGroups via those MBean attributes. Below is an example JGroups configuration in the ClusterPartition MBean.
<mbean code="org.jboss.ha.framework.server.ClusterPartition"
name="jboss:service=DefaultPartition">
... ...
<attribute name="PartitionConfig">
<Config>
<UDP mcast_addr="228.1.2.3" mcast_port="45566"
ip_ttl="8" ip_mcast="true"
mcast_send_buf_size="800000" mcast_recv_buf_size="150000"
ucast_send_buf_size="800000" ucast_recv_buf_size="150000"
loopback="false"/>
<PING timeout="2000" num_initial_members="3"
up_thread="true" down_thread="true"/>
<MERGE2 min_interval="10000" max_interval="20000"/>
<FD shun="true" up_thread="true" down_thread="true"
timeout="2500" max_tries="5"/>
<VERIFY_SUSPECT timeout="3000" num_msgs="3"
up_thread="true" down_thread="true"/>
<pbcast.NAKACK gc_lag="50"
retransmit_timeout="300,600,1200,2400,4800"
max_xmit_size="8192"
up_thread="true" down_thread="true"/>
<UNICAST timeout="300,600,1200,2400,4800"
window_size="100" min_threshold="10"
down_thread="true"/>
<pbcast.STABLE desired_avg_gossip="20000"
up_thread="true" down_thread="true"/>
<FRAG frag_size="8192"
down_thread="true" up_thread="true"/>
<pbcast.GMS join_timeout="5000" join_retry_timeout="2000"
shun="true" print_local_addr="true"/>
<pbcast.STATE_TRANSFER up_thread="true" down_thread="true"/>
</Config>
</attribute>
</mbean>
All the JGroups configuration data is contained in the <Config> element under the JGroups config MBean attribute. In the next several sections, we will dig into the options in the <Config> element and explain exactly what they mean.
2.1.1. Transport Protocols
The transport protocols send messages from one cluster node to another (unicast) or from cluster node to all other nodes in the cluster (mcast). JGroups supports UDP, TCP, and TUNNEL as transport protocols.
Note
The UDP, TCP, and TUNNEL elements are mutually exclusive. You can only have one transport protocol in each JGroups Config element
2.1.1.1. UDP configuration
UDP is the preferred protocol for JGroups. UDP uses multicast or multiple unicasts to send and receive messages. If you choose UDP as the transport protocol for your cluster service, you need to configure it in the UDP sub-element in the JGroups Config element. Here is an example.
<UDP mcast_send_buf_size="32000"
mcast_port="45566"
ucast_recv_buf_size="64000"
mcast_addr="228.8.8.8"
bind_to_all_interfaces="true"
loopback="true"
mcast_recv_buf_size="64000"
max_bundle_size="30000"
max_bundle_timeout="30"
use_incoming_packet_handler="false"
use_outgoing_packet_handler="false"
ucast_send_buf_size="32000"
ip_ttl="32"
enable_bundling="false"/>
The available attributes in the above JGroups configuration are listed below.
-
ip_mcast specifies whether or not to use IP multicasting. The default is true.
-
mcast_addr specifies the multicast address (class D) for joining a group (i.e., the cluster). The default is 228.8.8.8.
-
mcast_port specifies the multicast port number. The default is 45566.
-
bind_addr specifies the interface on which to receive and send multicasts (uses the bind.address system property, if present). If you have a multihomed machine, set the bind_addr attribute to the appropriate NIC IP address. Ignored if the ignore.bind.address property is true.
-
bind_to_all_interfaces specifies whether this node should listen on all interfaces for multicasts. The default is false. It overrides the bind_addr property for receiving multicasts. However, bind_addr (if set) is still used to send multicasts.
-
ip_ttl specifies the TTL for multicast packets.
-
use_incoming_packet_handler specifies whether to use a separate thread to process incoming messages.
-
use_outgoing_packet_handler specifies whether to use a separate thread to process outgoing messages.
-
enable_bundling specifies whether to enable bundling. If it is true, the node would queue outgoing messages until max_bundle_size bytes have accumulated, or max_bundle_time milliseconds have elapsed, whichever occurs first. Then bundle queued messages into a large message and send it. The messages are unbundled at the receiver. The default is false.
-
loopback specifies whether to loop outgoing message back up the stack. In unicast mode, the messages are sent to self. In mcast mode, a copy of the mcast message is sent.
-
discard_incompatibe_packets specifies whether to discard packets from different JGroups versions. Each message in the cluster is tagged with a JGroups version. When a message from a different version of JGroups is received, it will be discarded if set to true, otherwise a warning will be logged.
-
mcast_send_buf_size, mcast_recv_buf_size, ucast_send_buf_size, ucast_recv_buf_size define receive and send buffer sizes. It is good to have a large receiver buffer size, so packets are less likely to get dropped due to buffer overflow.
Note
On Windows 2000 machines, because of the media sense feature being broken with multicast (even after disabling media sense), you need to set the UDP protocol’s loopback attribute to true.
2.1.1.2. TCP configuration
Alternatively, a JGroups-based cluster can also work over TCP connections. Compared with UDP, TCP generates more network traffic when the cluster size increases but TCP is more reliable. TCP is fundamentally a unicast protocol. To send multicast messages, JGroups uses multiple TCP unicasts. To use TCP as a transport protocol, you should define a TCP element in the JGroups Config element. Here is an example of the TCP element.
<TCP start_port="7800"
bind_addr="192.168.5.1"
loopback="true"/>
Below are the attributes available in the TCP element.
-
bind_addr specifies the binding address. It can also be set with the -Dbind.address command line option at server startup.
-
start_port, end_port define the range of TCP ports the server should bind to. The server socket is bound to the first available port from start_port. If no available port is found (e.g., because of a firewall) before the end_port, the server throws an exception.
-
loopback specifies whether to loop outgoing message back up the stack. In unicast mode, the messages are sent to self. In mcast mode, a copy of the mcast message is sent.
-
mcast_send_buf_size, mcast_recv_buf_size, ucast_send_buf_size, ucast_recv_buf_size define receive and send buffer sizes. It is good to have a large receiver buffer size, so packets are less likely to get dropped due to buffer overflow.
-
conn_expire_time specifies the time (in milliseconds) after which a connection can be closed by the reaper if no traffic has been received.
-
reaper_interval specifies interval (in milliseconds) to run the reaper. If both values are 0, no reaping will be done. If either value is > 0, reaping will be enabled.
2.1.1.3. TUNNEL configuration
The TUNNEL protocol uses an external router to send messages. The external router is known as a GossipRouter. Each node has to register with the router. All messages are sent to the router and forwarded on to their destinations. The TUNNEL approach can be used to setup communication with nodes behind firewalls. A node can establish a TCP connection to the GossipRouter through the firewall (you can use port 80). The same connection is used by the router to send messages to nodes behind the firewall. The TUNNEL configuration is defined in the TUNNEL element in the JGroups Config element. Here is an example.
<TUNNEL router_port="12001"
router_host="192.168.5.1"/>
The available attributes in the TUNNEL element are listed below.
-
router_host specifies the host on which the GossipRouter is running.
-
router_port specifies the port on which the GossipRouter is listening.
-
loopback specifies whether to loop messages back up the stack. The default is true.
2.1.2. Discovery Protocols
The cluster need to maintain a list of current member nodes at all times so that the load balancer and client interceptor know how to route their requests. The discovery protocols are used to discover active nodes in the cluster. All initial nodes are discovered when the cluster starts up. When a new node joins the cluster later, it is only discovered after the group membership protocol (GMS, see Section 2.1.5.1, “Group Membership”) admits it into the group.
Since the discovery protocols sit on top of the transport protocol. You can choose to use different discovery protocols based on your transport protocol. The discovery protocols are also configured as sub-elements in the JGroups MBean Config element.
The PING discovery protocol normally sits on top of the UDP transport protocol. Each node responds with a unicast UDP datagram back to the sender. Here is an example PING configuration under the JGroups Config element.
<PING timeout="2000"
num_initial_members="2"/>
The available attributes in the PING element are listed below.
-
timeout specifies the maximum number of milliseconds to wait for any responses.
-
num_initial_members specifies the maximum number of responses to wait for.
-
gossip_host specifies the host on which the GossipRouter is running.
-
gossip_port specifies the port on which the GossipRouter is listening on.
-
gossip_refresh specifies the interval (in milliseconds) for the lease from the GossipRouter.
-
initial_hosts is a comma-seperated list of addresses (e.g., host1[12345],host2[23456]), which are pinged for discovery.
If both gossip_host and gossip_port are defined, the cluster uses the GossipRouter for the initial discovery. If the initial_hosts is specified, the cluster pings that static list of addresses for discovery. Otherwise, the cluster uses IP multicasting for discovery.
Note
The discovery phase returns when the timeout ms have elapsed or the num_initial_members responses have been received.
The TCPGOSSIP protocol only works with a GossipRouter. It works essentially the same way as the PING protocol configuration with valid gossip_host and gossip_port attributes. It works on top of both UDP and TCP transport protocols. Here is an example.
<PING timeout="2000"
initial_hosts="192.168.5.1[12000],192.168.0.2[12000]"
num_initial_members="3"/>
The available attributes in the TCPGOSSIP element are listed below.
-
timeout specifies the maximum number of milliseconds to wait for any responses.
-
num_initial_members specifies the maximum number of responses to wait for.
-
initial_hosts is a comma-seperated list of addresses (e.g., host1[12345],host2[23456]) for GossipRouters to register with.
The TCPPING protocol takes a set of known members and ping them for discovery. This is essentially a static configuration. It works on top of TCP. Here is an example of the TCPPING configuration element in the JGroups Config element.
<TCPPING timeout="2000"
initial_hosts="192.168.5.1[7800],192.168.0.2[7800]"
port_range="2"
num_initial_members="3"/>
The available attributes in the TCPPING element are listed below.
-
timeout specifies the maximum number of milliseconds to wait for any responses.
-
num_initial_members specifies the maximum number of responses to wait for.
-
initial_hosts is a comma-seperated list of addresses (e.g., host1[12345],host2[23456]) for pinging.
-
port_range specifies the range of ports to ping on each host in the initial_hosts list. That is because multiple nodes can run on the same host. In the above example, the cluster would ping ports 7800, 7801, and 7802 on both hosts.
The MPING protocol is a multicast ping over TCP. It works almost the same way as PING works on UDP. It does not require external processes (GossipRouter) or static configuration (initial host list). Here is an example of the MPING configuration element in the JGroups Config element.
<MPING timeout="2000"
bind_to_all_interfaces="true"
mcast_addr="228.8.8.8"
mcast_port="7500"
ip_ttl="8"
num_initial_members="3"/>
The available attributes in the MPING element are listed below.
-
timeout specifies the maximum number of milliseconds to wait for any responses.
-
num_initial_members specifies the maximum number of responses to wait for.
-
bind_addr specifies the interface on which to send and receive multicast packets.
-
bind_to_all_interfaces overrides the bind_addr and uses all interfaces in multihome nodes.
-
mcast_addr, mcast_port, ip_ttl attributes are the same as related attributes in the UDP protocol configuration.
2.1.3. Failure Detection Protocols
The failure detection protocols are used to detect failed nodes. Once a failed node is detected, the cluster updates its view so that the load balancer and client interceptors know to avoid the dead node. The failure detection protocols are configured as sub-elements in the JGroups MBean Config element.
The FD discovery protocol requires each node periodically sends are-you-alive messages to its neighbor. If the neighbor fails to respond, the calling node sends a SUSPECT message to the cluster. The current group coordinator double checks that the suspect node is indeed dead and updates the cluster’s view. Here is an example FD configuration.
<FD timeout="2000"
max_tries="3"
shun="true"/>
The available attributes in the FD element are listed below.
-
timeout specifies the maximum number of milliseconds to wait for the responses to the are-you-alive messages.
-
max_tries specifies the number of missed are-you-alive messages from a node before the node is suspected.
-
shun specifies whether a failed node will be shunned. Once shunned, the node will be expelled from the cluster even if it comes back later. The shunned node would have to re-join the cluster through the discovery process.
Note
Regular traffic from a node counts as if it is a live. So, the are-you-alive messages are only sent when there is no regular traffic to the node for sometime.
The are-you-alive messages in the FD protocol could increase the network load when there are many nodes. It could also produce false suspicions. For instance, if the network is too busy and the timeout is too short, nodes could be falsely suspected. Also, if one node is suspended in a debugger or profiler, it could also be suspected and shunned. The FD_SOCK protocol addresses the above issues by suspecting node failures only when a regular TCP connection to the node fails. However, the problem with such passive detection is that hung nodes will not be detected until it is accessed and the TCP timeouts after several minutes. FD_SOCK works best in high load networks where all nodes are frequently accessed. The simplest FD_SOCK configuration does not take any attribute. You can just declare an empty FD_SOCK element in JGroups’s Config element.
<FD_SOCK/>
There is only one optional attribute in the FD_SOCK element.
The FD_SIMPLE protocol is a more tolerant (less false suspicions) protocol based on are-you-alive messages. Each node periodically sends are-you-alive messages to a randomly choosen node and wait for a response. If a response has not been received within a certain timeout time, a counter associated with that node will be incremented. If the counter exceeds a certain value, that node will be suspected. When a response to an are-you-alive message is received, the counter resets to zero. Here is an example configuration for the FD_SIMPLE protocol.
<FD_SIMPLE timeout="2000"
max_missed_hbs="10"/>
The available attributes in the FD_SIMPLE element are listed below.
-
timeout specifies the timeout (in milliseconds) for the are-you-alive message. If a response is not received within timeout, the counter for the target node is increased.
-
max_missed_hbs specifies maximum number of are-you-alive messages (i.e., the counter value) a node can miss before it is suspected failure.
2.1.4. Reliable Delivery Protocols
The reliable delivery protocols in the JGroups stack ensure that data pockets are actually delivered in the right order (FIFO) to the destination node. The basis for reliable message delivery is positive and negative delivery acknowledgments (ACK and NAK). In the ACK mode, the sender resends the message until the acknowledgment is received from the receiver. In the NAK mode, the receiver requests retransmission when it discovers a gap.
The UNICAST protocol is used for unicast messages. It uses ACK. It is configured as a sub-element under the JGroups Config element. Here is an example configuration for the UNICAST protocol.
<UNICAST timeout="100,200,400,800"/>
There is only one configurable attribute in the UNICAST element.
-
timeout specifies the retransmission timeout (in milliseconds). For instance, if the timeout is “100,200,400,800”, the sender resends the message if it hasn’t received an ACK after 100 ms the first time, and the second time it waits for 200 ms before resending, and so on.
The NAKACK protocol is used for multicast messages. It uses NAK. Under this protocol, each message is tagged with a sequence number. The receiver keeps track of the sequence numbers and deliver the messages in order. When a gap in the sequence number is detected, the receiver asks the sender to retransmit the missing message. The NAKACK protocol is configured as the pbcast.NAKACK sub-element under the JGroups Config element. Here is an example configuration.
<pbcast.NAKACK
max_xmit_size="8192"
use_mcast_xmit="true"
retransmit_timeout="600,1200,2400,4800"/>
The configurable attributes in the pbcast.NAKACK element are as follows.
-
retransmit_timeout specifies the retransmission timeout (in milliseconds). It is the same as the timeout attribute in the UNICAST protocol.
-
use_mcast_xmit determines whether the sender should send the retransmission to the entire cluster rather than just the node requesting it. This is useful when the sender drops the pocket — so we do not need to retransmit for each node.
-
max_xmit_size specifies maximum size for a bundled retransmission, if multiple packets are reported missing.
-
discard_delivered_msgs specifies whether to discard delivery messages on the receiver nodes. By default, we save all delivered messages. However, if we only ask the sender to resend their messages, we can enable this option and discard delivered messages.
2.1.5. Other Configuration Options
In addition to the protocol stacks, you can also configure JGroups network services in the Config element.
2.1.5.1. Group Membership
The group membership service in the JGroups stack maintains a list of active nodes. It handles the requests to join and leave the cluster. It also handles the SUSPECT messages sent by failure detection protocols. All nodes in the cluster, as well as the load balancer and client side interceptors, are notified if the group membership changes. The group membership service is configured in the pbcast.GMS sub-element under the JGroups Config element. Here is an example configuration.
<pbcast.GMS print_local_addr="true"
join_timeout="3000"
down_thread="false"
join_retry_timeout="2000"
shun="true"/>
The configurable attributes in the pbcast.GMS element are as follows.
-
join_timeout specifies the maximum number of milliseconds to wait for a new node JOIN request to succeed. Retry afterwards.
-
join_retry_timeout specifies the maximum number of milliseconds to wait after a failed JOIN to re-submit it.
-
print_local_addr specifies whether to dump the node’s own address to the output when started.
-
shun specifies whether a node should shun itself if it receives a cluster view that it is not a member node.
-
disable_initial_coord specifies whether to prevent this node as the cluster coordinator.
The flow control service tries to adapt the sending data rate and the receiving data among nodes. If a sender node is too fast, it might overwhelm the receiver node and result in dropped packets that have to be retransmitted. In JGroups, the flow control is implemented via a credit-based system. The sender and receiver nodes have the same number of credits (bytes) to start with. The sender subtracts credits by the number of bytes in messages it sends. The receiver accumulates credits for the bytes in the messages it receives. When the sender’s credit drops to a threshold, the receivers sends some credit to the sender. If the sender’s credit is used up, the sender blocks until it receives credits from the receiver. The flow control service is configured in the FC sub-element under the JGroups Config element. Here is an example configuration.
<FC max_credits="1000000"
down_thread="false"
min_threshold="0.10"/>
The configurable attributes in the FC element are as follows.
-
max_credits specifies the maximum number of credits (in bytes). This value should be smaller than the JVM heap size.
-
min_credits specifies the threshold credit on the sender, below which the receiver should send in more credits.
-
min_threshold specifies percentage value of the threshold. It overrides the min_credits attribute.
The state transfer service transfers the state from an existing node (i.e., the cluster coordinator) to a newly joining node. It is configured in the pbcast.STATE_TRANSFER sub-element under the JGroups Config element. It does not have any configurable attribute. Here is an example configuration.
<pbcast.STATE_TRANSFER
down_thread="false"
up_thread="false"/>
2.1.5.4. Distributed Garbage Collection
In a JGroups cluster, all nodes have to store all messages received for potential retransmission in case of a failure. However, if we store all messages forever, we will run out of memory. So, the distributed garbage collection service in JGroups periodically purges messages that have seen by all nodes from the memory in each node. The distributed garbage collection service is configured in the pbcast.STABLE sub-element under the JGroups Config element. Here is an example configuration.
<pbcast.STABLE stability_delay="1000"
desired_avg_gossip="5000"
down_thread="false"
max_bytes="250000"/>
The configurable attributes in the pbcast.STABLE element are as follows.
-
desired_avg_gossip specifies intervals (in milliseconds) of garbage collection runs. Value 0 disables this service.
-
max_bytes specifies the maximum number of bytes received before the cluster triggers a garbage collection run. Value 0 disables this service.
-
max_gossip_runs specifies the maximum garbage collections runs before any changes. After this number is reached, there is no garbage collection until the message is received.
Note
Set the max_bytes attribute when you have a high traffic cluster.
When a network error occurs, the cluster might be partitioned into several different partitions. JGroups has a MERGE service that allows the coordinators in partitions to communicate with each other and form a single cluster back again. The flow control service is configured in the MERGE2 sub-element under the JGroups Config element. Here is an example configuration.
<MERGE2 max_interval="10000"
min_interval="2000"/>
The configurable attributes in the FC element are as follows.
JGroups chooses a random value between min_interval and max_interval to send out the MERGE message.
Note
The cluster states are not merged in a merger. This has to be done by the application.
2.2. JBossCache Configuration
JBoss Cache provides distributed cache and state replication services for the JBoss cluster. A JBoss cluster can have multiple JBoss Cache MBeans (known as the TreeCache MBean), one for HTTP session replication, one for stateful session beans, one for cached entity beans, etc. A generic TreeCache MBean configuration is listed below. Application specific TreeCache MBean configurations are covered in later chapters when those applications are discussed.
<mbean code="org.jboss.cache.TreeCache"
name="jboss.cache:service=TreeCache">
<depends>jboss:service=Naming</depends>
<depends>jboss:service=TransactionManager</depends>
<! -- Configure the TransactionManager -->
<attribute name="TransactionManagerLookupClass">
org.jboss.cache.DummyTransactionManagerLookup
</attribute>
<! --
Node locking level : SERIALIZABLE
REPEATABLE_READ (default)
READ_COMMITTED
READ_UNCOMMITTED
NONE
-->
<attribute name="IsolationLevel">REPEATABLE_READ</attribute>
<! -- Valid modes are LOCAL
REPL_ASYNC
REPL_SYNC
-->
<attribute name="CacheMode">LOCAL</attribute>
<! -- Name of cluster. Needs to be the same for all clusters, in order
to find each other -->
<attribute name="ClusterName">TreeCache-Cluster</attribute>
<! -- The max amount of time (in milliseconds) we wait until the
initial state (ie. the contents of the cache) are
retrieved from existing members in a clustered environment
-->
<attribute name="InitialStateRetrievalTimeout">5000</attribute>
<! -- Number of milliseconds to wait until all responses for a
synchronous call have been received.
-->
<attribute name="SyncReplTimeout">10000</attribute>
<! -- Max number of milliseconds to wait for a lock acquisition -->
<attribute name="LockAcquisitionTimeout">15000</attribute>
<! -- Name of the eviction policy class. -->
<attribute name="EvictionPolicyClass">
org.jboss.cache.eviction.LRUPolicy
</attribute>
<! -- Specific eviction policy configurations. This is LRU -->
<attribute name="EvictionPolicyConfig">
<config>
<attribute name="wakeUpIntervalSeconds">5</attribute>
<!-- Cache wide default -->
<region name="/_default_">
<attribute name="maxNodes">5000</attribute>
<attribute name="timeToLiveSeconds">1000</attribute>
</region>
<region name="/org/jboss/data">
<attribute name="maxNodes">5000</attribute>
<attribute name="timeToLiveSeconds">1000</attribute>
</region>
<region name="/org/jboss/test/data">
<attribute name="maxNodes">5</attribute>
<attribute name="timeToLiveSeconds">4</attribute>
</region>
</config>
</attribute>
<attribute name="CacheLoaderClass">
org.jboss.cache.loader.bdbje.BdbjeCacheLoader
</attribute>
<attribute name="CacheLoaderConfig">
location=c://tmp
</attribute>
<attribute name="CacheLoaderShared">true</attribute>
<attribute name="CacheLoaderPreload">
/a/b/c,/all/my/objects
</attribute>
<attribute name="CacheLoaderFetchTransientState">false</attribute>
<attribute name="CacheLoaderFetchPersistentState">true</attribute>
<attribute name="ClusterConfig">
... JGroups config for the cluster ...
</attribute>
</mbean>
The JGroups configuration element (i.e., the ClusterConfig attribute) is omitted from the above listing. You have learned how to configure JGroups earlier in this chapter (Section 2.1, “JGroups Configuration”). The TreeCache MBean takes the following attributes.
-
CacheLoaderClass specifies the fully qualified class name of the CacheLoader implementation.
-
CacheLoaderConfig contains a set of properties from which the specific CacheLoader implementation can configure itself.
-
CacheLoaderFetchPersistentState specifies whether to fetch the persistent state from another node. The persistence is fetched only if CacheLoaderShared is false. This attribute is only used if FetchStateOnStartup is true.
-
CacheLoaderFetchTransientState specifies whether to fetch the in-memory state from another node. This attribute is only used if FetchStateOnStartup is true.
-
CacheLoaderPreload contains a list of comma-separate nodes that need to be preloaded (e.g., /aop, /productcatalogue).
-
CacheLoaderShared specifies whether we want to shared a datastore, or whether each node wants to have its own local datastore.
-
CacheMode specifies how to synchronize cache between nodes. The possible values are LOCAL, REPL_SYNC, or REPL_ASYNC.
-
ClusterName specifies the name of the cluster. This value needs to be the same for all nodes in a cluster in order for them to find each other.
-
ClusterConfig contains the configuration of the underlying JGroups stack (see Section 2.1, “JGroups Configuration”.
-
EvictionPolicyClass specifies the name of a class implementing EvictionPolicy. You can use a JBoss Cache provided EvictionPolicy class or provide your own policy implementation. If this attribute is empty, no eviction policy is enabled.
-
EvictionPolicyConfig contains the configuration parameter for the specified eviction policy. Note that the content is provider specific.
-
FetchStateOnStartup specifies whether or not to acquire the initial state from existing members. It allows for warm/hot caches (true/false). This can be further defined by CacheLoaderFetchTransientState and CacheLoaderFetchPersistentState.
-
InitialStateRetrievalTimeout specifies the time in milliseconds to wait for initial state retrieval.
-
IsolationLevel specifies the node locking level. Possible values are SERIALIZABLE, REPEATABLE_READ (default), READ_COMMITTED, READ_UNCOMMITTED, and NONE.
-
LockAcquisitionTimeout specifies the time in milliseconds to wait for a lock to be acquired. If a lock cannot be acquired an exception will be thrown.
-
ReplQueueInterval specifies the time in milliseconds for elements from the replication queue to be replicated.
-
SyncReplTimeout specifies the time in milliseconds to wait until replication ACKs have been received from all nodes in the cluster. This attribute applies to synchronous replication mode only (i.e., CacheMode attribute is REPL_SYNC).
-
UseReplQueue specifies whether or not to use a replication queue (true/false). This attribute applies to synchronous replication mode only (i.e., CacheMode attribute is REPL_ASYNC).
-
ReplQueueMaxElements specifies the maximum number of elements in the replication queue until replication kicks in.
-
TransactionManagerLookupClass specifies the fully qualified name of a class implementing TransactionManagerLookup. The default is JBossTransactionManagerLookup for the transaction manager inside the JBoss AS. There is also an option of DummyTransactionManagerLookup for simple standalone examples.
|
|
Recent Comments