How to setup a basic cluster¶
The OpenHIM Core is designed to be horizontally scalable. This means that if you need better performance, you can easily fire up more core instances. In this tutorial we will look at setting up a small, basic cluster of core instances.
The OpenHIM Core is built in Node.js and it is important to note that node uses a single threaded model. The threading model is designed for I/O bound processes - perfect for the OpenHIM - however this can lead to a single instance of core quickly becoming a bottleneck on very high transaction loads. In addition, a core single instance wouldn’t take advantage multiple cores on a CPU. On a dedicated server it thus recommended that one OpenHIM instance is used for every available CPU core. In addition it would of course be useful to be able to run multiple core instances on multiple servers as well.
Clustering on a single server¶
Luckily, since v1.2.0 of the OpenHIM, clustering on a single server is supported out of the box. So, setup is really easy. All you need to do is run the OpenHIM with a flag to let it know you want to cluster:
N is the number of instances that you want to run (eg, 2, 4 or 6) or it can be the special value of ‘auto’ where the OpenHIM will determine how many core your server has and run that many instances. Each instance that the OpenHIM starts shares the same ports and the OpenHIM will share load between all of these instances.
Clustering over multiple servers¶
The approach we’ll be taking towards scaling out core to multiple server is also really straight forward! We’ll simply fire up 4 instances on separate servers and setup load-balancing between them. For this tutorial we’ll look at using LVS on Ubuntu for load-balancing, but other options exist as well:
- NGINX - A very powerful load-balancer and web server. Note however that TCP load-balancing is only supported in their NGINX Plus commercial product. However http load-balancing can still be used, but can become more complex if you want to use https channels on the HIM.
First, install the OpenHIM on each server and grab a copy of the config file is you wish to use a non-default configuration:
npm install -g openhim-core $ wget https://raw.githubusercontent.com/jembi/openhim-core-js/master/config/default.json
Next, startup each instance on each server, you may also use the –cluster option if you choose to.
nohup openhim-core --cluster=auto
Now we can setup the load-balancer. If not already available, install the LVS Admin tool on the server that will act as you load balancer:
sudo apt-get install ipvsadm
Now we’ll setup round-robin balancing and add each node to the cluster:
sudo ipvsadm -A -t 10.0.0.1:5000 -s rr sudo ipvsadm -a -t 10.0.0.1:5000 -r 10.0.0.2:5000 -m sudo ipvsadm -a -t 10.0.0.1:5000 -r 10.0.0.3:5000 -m sudo ipvsadm -a -t 10.0.0.1:5000 -r 10.0.0.4:5000 -m sudo ipvsadm -a -t 10.0.0.1:5000 -r 10.0.0.5:5000 -m
Replace the 10.0.0.x addresses with your servers’ IP addresses. You can also follow these steps for the HTTP ports and the API ports (although another good strategy with regard to the API is to run another dedicated core instance and point the Console there). And that’s it! Try running several transactions against your server on HTTPS - they should be load-balanced between the four instances AND each instance will itself be clustered on the server that it is running! This should allow you to handle MASSIVE load with ease.