We define high availability as the ability for the system to continue functioning after failure of one or more of the servers.
A part of high availability is failover which we define as the ability for client connections to migrate from one server to another in event of server failure so client applications can continue to operate.
1. Terminology
In order to discuss both configuration and runtime behavior consistently we need to define a pair of nouns and adjectives. These terms will be used throughout the documentation, configuration, source code, and runtime logs.
1.1. Configuration
These nouns identify how the broker is configured, e.g. in broker.xml
.
- primary
-
This identifies the main broker in the high availability configuration. Oftentimes the hardware on this broker will be higher performance than the hardware on the backup broker. Typically, this broker is started before the backup and is active most of the time.
- backup
-
This identifies the broker that should take over when the primary broker fails in a high availability configuration. Oftentimes the hardware on this broker will be lower performance than the hardware on the primary broker. Typically, this broker is started after the primary and is passive most of the time.
1.2. Runtime
These adjectives describe the behavior of the broker at runtime. For example, you could have a passive primary or an active backup.
- active
-
This identifies a broker in a high-availability configuration which is accepting remote connections. For example, consider the scenario where the primary broker has failed and its backup has taken over. The backup would be described as active at that point since it is accepting remote connections.
- passive
-
This identifies a broker in a high-availability configuration which is not accepting remote connections. For example, consider the scenario where the primary broker was started and then the backup broker was started. The backup broker would be passive since it is not accepting remote connections. It is waiting for the primary to fail before it activates and begins accepting remote connections.
2. Primary/Backup Groups
Apache ActiveMQ Artemis allows servers to be linked together as primary/backup groups where each primary server can have 1 or more backup servers. A backup server is owned by only one primary server. Backup servers are not operational until failover occurs. However, one chosen backup, which will be passive, announces its status and waits to take over the primary server’s work.
Before failover, only the primary server is active, serving clients while the backup servers remain passive, awaiting to become active when the primary fails. When a primary server crashes or is brought down in the correct mode the backup server currently in passive mode will activate. If a primary server restarts after a failover then it will be passive and have priority and be the next server to become active when the current active backup server goes down. If the active backup server is configured to allow automatic failback then it will detect the primary server coming back up and automatically stop.
2.1. HA Policies
Apache ActiveMQ Artemis supports two different strategies for backing up a server:
-
shared store
-
replication
These are configured via the ha-policy
configuration element, e.g.:
<ha-policy>
<replication/>
</ha-policy>
or
<ha-policy>
<shared-store/>
</ha-policy>
As well as these 2 strategies there is also a 3rd called primary-only
.
This of course means there will be no Backup Strategy and is the default if none is provided, however this is used to configure scale-down
which we will cover in a later chapter.
The |
Only persistent message data will survive failover. Any non persistent message data will not be available after failover. |
The ha-policy
type configures which strategy a cluster should use to provide the backing up of a server’s data.
Within this configuration element we configure how a server should behave within the cluster, either as a primary (active), backup (passive) or colocated (both active and passive).
This would look something like:
<ha-policy>
<replication>
<primary/>
</replication>
</ha-policy>
or
<ha-policy>
<shared-store>
<backup/>
</shared-store>
</ha-policy>
or
<ha-policy>
<replication>
<colocated/>
</replication>
</ha-policy>
Replication allows the configuration of two new roles to enable pluggable quorum provider configuration, by using:
<ha-policy>
<replication>
<primary/>
</replication>
</ha-policy>
to configure the classic primary role, and
<ha-policy>
<replication>
<backup/>
</replication>
</ha-policy>
for the classic backup one.
If replication is configured using such new roles some additional element are required to complete configuration as detailed later.
2.2. IMPORTANT NOTE ON PLUGGABLE QUORUM VOTE FEATURE
This feature is still EXPERIMENTAL. Extra testing should be done before running this feature into production. Please raise issues eventually found to the ActiveMQ Artemis Mail Lists.
It means:
-
it’s configuration can change until declared as officially stable
2.3. Data Replication
When using replication, the primary and the backup servers do not share the same data directories, all data synchronization is done over the network. Therefore all (persistent) data received by the primary server will be duplicated to the backup.
Notice that upon start-up the backup server will first need to synchronize all existing data from the primary server before becoming capable of replacing the primary server should it fail. So unlike when using shared storage, a replicating backup will not be a fully operational backup right after start-up, but only after it finishes synchronizing the data with its primary server. The time it will take for this to happen will depend on the amount of data to be synchronized and the connection speed.
In general, synchronization occurs in parallel with current network traffic so this won’t cause any blocking on current clients.
However, there is a critical moment at the end of this process where the replicating server must complete the synchronization and ensure the replica acknowledges this completion.
This exchange between the replicating server and replica will block any journal related operations.
The maximum length of time that this exchange will block is controlled by the |
Replication will create a copy of the data at the backup. One issue to be aware of is: in case of a successful fail-over, the backup’s data will be newer than the primary’s data. If you configure your backup to allow failback to the primary then when the primary is restarted it will be passive and the active backup will synchronize its data with the passive primary before stopping to allow the passive primary to become active again. If both servers are shutdown then the administrator will have to determine which one has the latest data.
The replicating primary and backup pair must be part of a cluster. The Cluster Connection also defines how backup servers will find the remote primary servers to pair with. Refer to Clusters for details on how this is done, and how to configure a cluster connection. Notice that:
-
Both primary and backup servers must be part of the same cluster. Notice that even a simple primary/backup replicating pair will require a cluster configuration.
-
Their cluster user and password must match.
Within a cluster, there are two ways that a backup server will locate a primary server to replicate from. These are:
- specifying a node group
-
You can specify a group of primary servers that a backup server can connect to. This is done by configuring
group-name
in either theprimary
or thebackup
element of thebroker.xml
. A backup will only connect to a primary that shares the same node group name. - connecting to any live
-
This will be the behaviour if
group-name
is not configured allowing a backup server to connect to any primary server.
A
After joining the cluster the backups with The 2 backups with |
The backup will search for any primary server that it is configured to connect to. It then tries to replicate with each primary server in turn until it finds a primary server that has no current backup configured. If no primary server is available it will wait until the cluster topology changes and repeats the process.
This is an important distinction from a shared-store backup, if a backup starts and does not find a primary server, the server will just activate and start to serve client requests.
In the replication case, the backup just keeps waiting for a primary server to pair with.
Note that in replication the backup server does not know whether any data it might have is up to date, so it really cannot decide to activate automatically.
To activate a replicating backup server using the data it has, the administrator must change its configuration to make it a primary server by changing |
Much like in the shared-store case, when the primary server stops or crashes, its backup will become active and take over its duties. Specifically, the backup will become active when it loses connection to its primary server. This can be problematic because it can also happen as the result of temporary network problem.
The issue can be solved in two different ways, depending on which replication roles are configured:
-
non-pluggable replication: backup will try to determine whether it still can connect to the other servers in the cluster. If it can connect to more than half the servers, it will become active. If more than half the servers also disappeared with the primary, the backup will wait and try reconnecting with the primary. This avoids a split brain situation.
-
pluggable replication: backup relies on a pluggable quorum provider (configurable via
manager
xml element) to detect if there’s any active primary.
A backup in the pluggable replication still needs to carefully configure connection-ttl in order to promptly send a request to the quorum manager to become active before failing-over. |
2.3.1. Configuration
To configure a non-pluggable replication’s primary and backup servers to be a replicating pair, configure the primary server in ' broker.xml
to have:
<ha-policy>
<replication>
<primary/>
</replication>
</ha-policy>
...
<cluster-connections>
<cluster-connection name="my-cluster">
...
</cluster-connection>
</cluster-connections>
The backup server must be similarly configured but as a backup
<ha-policy>
<replication>
<backup/>
</replication>
</ha-policy>
To configure a pluggable quorum replication’s primary and backup use:
<ha-policy>
<replication>
<primary/>
</replication>
</ha-policy>
...
<cluster-connections>
<cluster-connection name="my-cluster">
...
</cluster-connection>
</cluster-connections>
and
<ha-policy>
<replication>
<backup/>
</replication>
</ha-policy>
2.3.2. All Replication Configuration
Primary
The following table lists all the ha-policy
configuration elements for HA strategy Replication for primary
:
- check-for-active-server
-
Whether to check the cluster for a (live) server using our own server ID when starting up. This is an important option to avoid split-brain when failover happens and the primary is restarted. Default is
false
. - cluster-name
-
Name of the cluster configuration to use for replication. This setting is only necessary if you configure multiple cluster connections. If configured then the connector configuration of the cluster configuration with this name will be used when connecting to the cluster to discover if an active server is already running, see
check-for-active-server
. If unset then the default cluster connections configuration is used (the first one configured). - group-name
-
If set, backup servers will only pair with primary servers with matching
group-name
. - initial-replication-sync-timeout
-
The amount of time the replicating server will wait at the completion of the initial replication process for the replica to acknowledge it has received all the necessary data. The default is 30,000 milliseconds.
during this interval any journal related operations will be blocked.
Backup
The following table lists all the ha-policy
configuration elements for HA strategy Replication for backup
:
- cluster-name
-
Name of the cluster configuration to use for replication. This setting is only necessary if you configure multiple cluster connections. If configured then the connector configuration of the cluster configuration with this name will be used when connecting to the cluster to discover if an active server is already running, see
check-for-active-server
. If unset then the default cluster connections configuration is used (the first one configured). - group-name
-
If set, backup servers will only pair with primary servers with matching group-name
- max-saved-replicated-journals-size
-
This option specifies how many replication backup directories will be kept when server starts as replica. Every time when server starts as replica all former data moves to 'oldreplica.{id}' directory, where id is growing backup index, this parameter sets the maximum number of such directories kept on disk.
- allow-failback
-
Whether a server will automatically stop when another places a request to take over its place. The use case is when the backup has failed over.
- initial-replication-sync-timeout
-
After failover and the backup has become active, this is set on the new active server. It represents the amount of time the replicating server will wait at the completion of the initial replication process for the replica to acknowledge it has received all the necessary data. The default is 30,000 milliseconds.
During this interval any journal related operations will be blocked.
2.3.3. Pluggable Quorum Vote Replication configurations
Pluggable Quorum Vote replication configuration options are a bit different from classic replication, mostly because of its customizable nature.
Apache curator is used by the default quorum provider.
Below some example configurations to show how it works.
For primary
:
<ha-policy>
<replication>
<primary>
<manager>
<class-name>org.apache.activemq.artemis.quorum.zookeeper.CuratorDistributedPrimitiveManager</class-name>
<properties>
<property key="connect-string" value="127.0.0.1:6666,127.0.0.1:6667,127.0.0.1:6668"/>
</properties>
</manager>
</primary>
</replication>
</ha-policy>
And backup
:
<ha-policy>
<replication>
<backup>
<manager>
<class-name>org.apache.activemq.artemis.quorum.zookeeper.CuratorDistributedPrimitiveManager</class-name>
<properties>
<property key="connect-string" value="127.0.0.1:6666,127.0.0.1:6667,127.0.0.1:6668"/>
</properties>
</manager>
<allow-failback>true</allow-failback>
</backup>
</replication>
</ha-policy>
The configuration of class-name
as follows
<class-name>org.apache.activemq.artemis.quorum.zookeeper.CuratorDistributedPrimitiveManager</class-name>
isn’t really needed, because Apache Curator is the default provider, but has been shown for completeness.
The properties
element:
<properties>
<property key="connect-string" value="127.0.0.1:6666,127.0.0.1:6667,127.0.0.1:6668"/>
</properties>
can specify a list of property
elements in the form of key-value pairs, appropriate to what is supported by the specified class-name
provider.
Apache Curator’s provider allows the following properties:
-
connect-string
: (no default) -
session-ms
: (default is 18000 ms) -
session-percent
: (default is 33); should be <= default, see https://cwiki.apache.org/confluence/display/CURATOR/TN14 for more info -
connection-ms
: (default is 8000 ms) -
retries
: (default is 1) -
retries-ms
: (default is 1000 ms) -
namespace
: (no default)
Configuration of the Apache ZooKeeper ensemble is the responsibility of the user, but there are few suggestions to improve the reliability of the quorum service:
-
broker
session_ms
must be>= 2 * server tick time
and<= 20 * server tick time
as by ZooKeeper 3.6.3 admin guide: it directly impacts how fast a backup can failover to an isolated/killed/unresponsive live; the higher, the slower. -
GC on broker machine should allow keeping GC pauses within 1/3 of
session_ms
in order to let the ZooKeeper heartbeat protocol work reliably. If that is not possible, it is better to increasesession_ms
, accepting a slower failover. -
ZooKeeper must have enough resources to keep GC (and OS) pauses much smaller than server tick time: please consider carefully if broker and ZooKeeper node should share the same physical machine, depending on the expected load of the broker
-
network isolation protection requires configuring >=3 ZooKeeper nodes
Peer or Multi Primary
With coordination delegated to the quorum service, roles are less important. It is possible to have two peer servers compete for activation; the winner activating as live, the looser taking up a backup role. On restart, 'any' peer server with the most up to date journal can activate. The instances need to know in advance, what identity they will coordinate on. In the replication 'primary' ha policy we can explicitly set the 'coordination-id' to a common value for all peers in a cluster.
For multi primary
:
<ha-policy>
<replication>
<primary>
<manager>
<class-name>org.apache.activemq.artemis.quorum.zookeeper.CuratorDistributedPrimitiveManager</class-name>
<properties>
<property key="connect-string" value="127.0.0.1:6666,127.0.0.1:6667,127.0.0.1:6668"/>
</properties>
</manager>
<coordination-id>peer-journal-001</coordination-id>
</primary>
</replication>
</ha-policy>
the string value provided will be converted internally into a 16 byte UUID, so it may not be immediately recognisable or human-readable, however it will ensure that all 'peers' coordinate. |
2.4. Shared Store
When using a shared store both primary and backup servers share the same entire data directory using a shared file system. This means the paging directory, journal directory, large messages and binding journal.
When failover occurs and a backup server takes over, it will load the persistent storage from the shared file system and clients can connect to it.
This style of high availability differs from data replication in that it requires a shared file system which is accessible by both the primary and backup nodes. Typically this will be some kind of high performance Storage Area Network (SAN). We do not recommend you use Network Attached Storage (NAS), e.g. NFS mounts to store any shared journal (NFS is slow).
The advantage of shared-store high availability is that no replication occurs between the primary and backup nodes. This means it does not suffer any performance penalties due to the overhead of replication during normal operation.
The disadvantage of shared store replication is that it requires a shared file system, and when the backup server activates it needs to load the journal from the shared store which can take some time depending on the amount of data in the store.
If you require the highest performance during normal operation then acquire access to a fast SAN and deal with a slightly slower failover (depending on amount of data).
2.4.1. Configuration
To configure the primary and backup servers to share their store use the ha-policy
configuration in broker.xml
:
<ha-policy>
<shared-store>
<primary/>
</shared-store>
</ha-policy>
...
<cluster-connections>
<cluster-connection name="my-cluster">
...
</cluster-connection>
</cluster-connections>
The backup server must also be configured as a backup.
<ha-policy>
<shared-store>
<backup/>
</shared-store>
</ha-policy>
In order for primary/backup groups to operate properly with a shared store, both servers must have configured the location of journal directory to point to the same shared location (as explained in Configuring the message journal)
todo write something about GFS |
Also each node, primary and backups, will need to have a cluster connection defined even if not part of a cluster. The Cluster Connection info defines how backup servers announce there presence to its primary server or any other nodes in the cluster. Refer to Clusters for details on how this is done.
2.5. Failing Back to Primary Server
After a primary server has failed and a backup taken has taken over its duties, you may want to restart the primary server and have clients fail back.
2.5.1. Shared Store
In case of "shared disk" you have a couple of options:
-
Simply restart the primary and kill the backup. You can do this by killing the process itself.
-
Alternatively you can set
allow-fail-back
totrue
on the backup which will force the backup that has become active to automatically stop. This configuration would look like:<ha-policy> <shared-store> <backup> <allow-failback>true</allow-failback> </backup> </shared-store> </ha-policy>
It is also possible, in the case of shared store, to cause failover to occur on normal server shutdown, to enable this set the following property to true in the ha-policy
configuration on either the primary
or backup
like so:
<ha-policy>
<shared-store>
<primary>
<failover-on-shutdown>true</failover-on-shutdown>
</primary>
</shared-store>
</ha-policy>
By default this is set to false, if by some chance you have set this to false but still want to stop the server normally and cause failover then you can do this by using the management API as explained at Management
You can also force the active backup to shutdown when the primary comes back up allowing the primary to take over automatically by setting the following property in the broker.xml
configuration file as follows:
<ha-policy>
<shared-store>
<backup>
<allow-failback>true</allow-failback>
</backup>
</shared-store>
</ha-policy>
2.5.2. Replication
As with shared storage the allow-failback
option can be set for both non-pluggable and pluggable replication.
Non-Pluggable
<ha-policy>
<replication>
<backup>
<allow-failback>true</allow-failback>
</backup>
</replication>
</ha-policy>
With non-pluggable replication you need to set an extra property check-for-active-server
to true
in the primary
configuration.
If set to true
then during start-up the primary server will first search the cluster for another active server using its nodeID.
If it finds one it will contact this server and try to "fail-back".
Since this is a remote replication scenario the primary will have to synchronize its data with the backup server running with its ID. Once they are in sync it will request the other server (which it assumes it is a backup that has assumed its duties) to shutdown in order for it to take over.
This is necessary because otherwise the primary server has no means to know whether there was a fail-over or not, and if there was, if the server that took its duties is still running or not.
To configure this option at your broker.xml
configuration file as follows:
<ha-policy>
<replication>
<primary>
<check-for-active-server>true</check-for-active-server>
</primary>
</replication>
</ha-policy>
For Non-Pluggable Replication
Be aware that if you restart a primary server after failover has occurred then |
Pluggable
One key difference between pluggable replication and non-pluggable replication is that with non-pluggable replication if the primary cannot reach any active server with its nodeID then it activates unilaterally. With pluggable replication the responsibilities of coordination are delegated to the quorum provider. There are no unilateral decisions. The primary will only activate when it knows that it has the most up to date version of the journal identified by its nodeID.
In short: a primary cannot become active without consensus when using pluggable replication.
Here’s an example configuration:
<ha-policy>
<replication>
<manager>
<!-- some meaningful configuration -->
</manager>
<primary>
<!-- no need to check-for-active-server anymore -->
</primary>
</replication>
</ha-policy>
2.5.3. All Shared Store Configuration
Primary
The following lists all the ha-policy
configuration elements for HA strategy shared store for primary
:
- failover-on-shutdown
-
If set to
true
then when this server is stopped normally the backup will become active assuming failover. If false then the backup server will remain passive. Note that iffalse
and you want failover to occur then you can use the management API as explained at Management. - wait-for-activation
-
If set to true then server startup will wait until it is activated. If set to false then server startup will be done in the background. Default is
true
.
Backup
The following lists all the ha-policy
configuration elements for HA strategy Shared Store for backup
:
- failover-on-shutdown
-
In the case of a backup that has become active then when set to
true
and this server is stopped normally the passive primary will become active assuming failover. Iffalse
then the primary server will remain passive. Note that iffalse
and you want failover to occur then you can use the management API as explained at Management. - allow-failback
-
Whether a server will automatically stop when another places a request to take over its place. The use case is when the backup has failed over.
2.5.4. Colocated Backup Servers
It is also possible when running standalone to colocate backup servers in the same JVM as another primary server.
Primary Servers can be configured to request another primary server in the cluster to start a backup server in the same JVM either using shared store or replication.
The new backup server will inherit its configuration from the primary server creating it apart from its name, which will be set to colocated_backup_n
where n is the number of backups the server has created, and any directories and its Connectors and Acceptors which are discussed later on in this chapter.
A primary server can also be configured to allow requests from backups and also how many backups a primary server can start.
This way you can evenly distribute backups around the cluster.
This is configured via the ha-policy
element in the broker.xml
file like so:
<ha-policy>
<replication>
<colocated>
<request-backup>true</request-backup>
<max-backups>1</max-backups>
<backup-request-retries>-1</backup-request-retries>
<backup-request-retry-interval>5000</backup-request-retry-interval>
<primary/>
<backup/>
</colocated>
</replication>
</ha-policy>
the above example is configured to use replication, in this case the primary
and backup
configurations must match those for normal replication as in the previous chapter.
shared-store
is also supported
2.5.5. Configuring Connectors and Acceptors
If the HA Policy is colocated
then connectors
and acceptors
will be inherited from the primary server creating it and offset depending on the setting of backup-port-offset
configuration element.
If this is set to say 100 (which is the default) and a connector is using port 61616 then this will be set to 61716 for the first server created, 61816 for the second, and so on.
for INVM connectors and Acceptors the id will have |
2.5.6. Remote Connectors
It may be that some of the Connectors configured are for external servers and hence should be excluded from the offset.
for instance a connector used by the cluster connection to do quorum voting for a replicated backup server, these can be omitted from being offset by adding them to the ha-policy
configuration like so:
<ha-policy>
<replication>
<colocated>
...
<excludes>
<connector-ref>remote-connector</connector-ref>
</excludes>
...
</colocated>
</replication
</ha-policy>
2.5.7. Configuring Directories
Directories for the Journal, Large messages and Paging will be set according to what the HA strategy is. If shared store the requesting server will notify the target server of which directories to use. If replication is configured then directories will be inherited from the creating server but have the new backups name appended.
The following table lists all the ha-policy
configuration elements for colocated policy:
- request-backup
-
If true then the server will request a backup on another node
- backup-request-retries
-
How many times the primary server will try to request a backup,
-1
means for ever. - backup-request-retry-interval
-
How long to wait for retries between attempts to request a backup server.
- max-backups
-
How many backups a primary server can create
- backup-port-offset
-
The offset to use for the Connectors and Acceptors when creating a new backup server.
2.6. Scaling Down
An alternative to using primary/backup groups is to configure scaledown. When configured for scale down a server can copy all its messages and transaction state to another active server. The advantage of this is that you don’t need full backups to provide some form of HA, however there are disadvantages with this approach the first being that it only deals with a server being stopped and not a server crash. The caveat here is if you configure a backup to scale down.
Another disadvantage is that it is possible to lose message ordering. This happens in the following scenario, say you have 2 active servers and messages are distributed evenly between the servers from a single producer, if one of the servers scales down then the messages sent back to the other server will be in the queue after the ones already there, so server 1 could have messages 1,3,5,7,9 and server 2 would have 2,4,6,8,10, if server 2 scales down the order in server 1 would be 1,3,5,7,9,2,4,6,8,10.
The configuration for an active server to scale down would be something like:
<ha-policy>
<primary-only>
<scale-down>
<connectors>
<connector-ref>server1-connector</connector-ref>
</connectors>
</scale-down>
</primary-only>
</ha-policy>
In this instance the server is configured to use a specific connector to scale down, if a connector is not specified then the first INVM connector is chosen, this is to make scale down from a backup server easy to configure. It is also possible to use discovery to scale down, this would look like:
<ha-policy>
<primary-only>
<scale-down>
<discovery-group-ref discovery-group-name="my-discovery-group"/>
</scale-down>
</primary-only>
</ha-policy>
2.6.1. Scale Down with groups
It is also possible to configure servers to only scale down to servers that belong in the same group. This is done by configuring the group like so:
<ha-policy>
<primary-only>
<scale-down>
...
<group-name>my-group</group-name>
</scale-down>
</primary-only>
</ha-policy>
In this scenario only servers that belong to the group my-group
will be scaled down to
2.6.2. Scale Down and Backups
It is also possible to mix scale down with HA via backup servers.
If a backup is configured to scale down then after failover has occurred, instead of starting fully the backup server will immediately scale down to another active server.
The most appropriate configuration for this is using the colocated
approach.
It means that as you bring up primary servers they will automatically be backed up, and as they are shutdown their messages are made available on another active server.
A typical configuration would look like:
<ha-policy>
<replication>
<colocated>
<backup-request-retries>44</backup-request-retries>
<backup-request-retry-interval>33</backup-request-retry-interval>
<max-backups>3</max-backups>
<request-backup>false</request-backup>
<backup-port-offset>33</backup-port-offset>
<primary>
<group-name>purple</group-name>
<check-for-active-server>true</check-for-active-server>
<cluster-name>abcdefg</cluster-name>
</primary>
<backup>
<group-name>tiddles</group-name>
<max-saved-replicated-journals-size>22</max-saved-replicated-journals-size>
<cluster-name>33rrrrr</cluster-name>
<restart-backup>false</restart-backup>
<scale-down>
<!--a grouping of servers that can be scaled down to-->
<group-name>boo!</group-name>
<!--either a discovery group-->
<discovery-group-ref discovery-group-name="wahey"/>
</scale-down>
</backup>
</colocated>
</replication>
</ha-policy>
2.6.3. Scale Down and Clients
When a server is stopping and preparing to scale down it will send a message to all its clients informing them which server it is scaling down to before disconnecting them. At this point the client will reconnect however this will only succeed once the server has completed the scaledown process. This is to ensure that any state such as queues or transactions are there for the client when it reconnects. The normal reconnect settings apply when the client is reconnecting so these should be high enough to deal with the time needed to scale down.
3. Client Failover
Apache ActiveMQ Artemis clients can be configured to receive knowledge of all primary and backup servers, so that in event of connection failure the client will detect this and reconnect to the backup server. The backup server will then automatically recreate any sessions and consumers that existed on each connection before failover, thus saving the user from having to hand-code manual reconnection logic. For further details see Client Failover
3.1. Handling Blocking Calls During Failover
If the client code is in a blocking call to the server, waiting for a response to continue its execution, when failover occurs, the new session will not have any knowledge of the call that was in progress. This call might otherwise hang for ever, waiting for a response that will never come.
To prevent this, Apache ActiveMQ Artemis will unblock any blocking calls that were in progress at the time of failover by making them throw a javax.jms.JMSException
(if using JMS), or a ActiveMQException
with error code ActiveMQException.UNBLOCKED
.
It is up to the client code to catch this exception and retry any operations if desired.
If the method being unblocked is a call to commit(), or prepare(), then the transaction will be automatically rolled back and Apache ActiveMQ Artemis will throw a javax.jms.TransactionRolledBackException
(if using JMS), or a ActiveMQException
with error code ActiveMQException.TRANSACTION_ROLLED_BACK
if using the core API.
3.2. Handling Failover With Transactions
If the session is transactional and messages have already been sent or acknowledged in the current transaction, then the server cannot be sure that messages sent or acknowledgements have not been lost during the failover.
Consequently the transaction will be marked as rollback-only, and any subsequent attempt to commit it will throw a javax.jms.TransactionRolledBackException
(if using JMS), or a ActiveMQException
with error code ActiveMQException.TRANSACTION_ROLLED_BACK
if using the core API.
The caveat to this rule is when XA is used either via JMS or through the core API.
If 2 phase commit is used and prepare has already been called then rolling back could cause a |
It is up to the user to catch the exception, and perform any client side local rollback code as necessary. There is no need to manually rollback the session - it is already rolled back. The user can then just retry the transactional operations again on the same session.
Apache ActiveMQ Artemis ships with a fully functioning example demonstrating how to do this, please see the examples chapter.
If failover occurs when a commit call is being executed, the server, as previously described, will unblock the call to prevent a hang, since no response will come back. In this case it is not easy for the client to determine whether the transaction commit was actually processed before failure occurred.
If XA is being used either via JMS or through the core API then an |
To remedy this, the client can simply enable duplicate detection (Duplicate Message Detection) in the transaction, and retry the transaction operations again after the call is unblocked. If the transaction had indeed been committed successfully before failover, then when the transaction is retried, duplicate detection will ensure that any durable messages resent in the transaction will be ignored on the server to prevent them getting sent more than once.
By catching the rollback exceptions and retrying, catching unblocked calls and enabling duplicate detection, once and only once delivery guarantees can be provided for messages in the case of failure, guaranteeing 100% no loss or duplication of messages. |
3.2.1. Handling Failover With Non Transactional Sessions
If the session is non transactional, messages or acknowledgements can be lost in the event of a failover.
If you wish to provide once and only once delivery guarantees for non transacted sessions too, enable duplicate detection, and catch unblock exceptions as described in Handling Blocking Calls During Failover
3.2.2. Use client connectors to fail over
Apache ActiveMQ Artemis clients retrieve the backup connector from the topology updates that the cluster brokers send.
If the connection options of the clients don’t match the options of the cluster brokers the clients can define a client connector that will be used in place of the connector in the topology.
To define a client connector it must have a name that matches the name of the connector defined in the cluster-connection
of the broker, i.e. supposing to have a primary broker with the cluster connector name node-0
and a backup broker with the cluster-connector
name node-1
the client connection url must define 2 connectors with the names node-0
and node-1
:
Primary broker config:
<connectors>
<!-- Connector used to be announced through cluster connections and notifications -->
<connector name="node-0">tcp://localhost:61616</connector>
</connectors>
...
<cluster-connections>
<cluster-connection name="my-cluster">
<connector-ref>node-0</connector-ref>
...
</cluster-connection>
</cluster-connections>
Backup broker config
<connectors>
<!-- Connector used to be announced through cluster connections and notifications -->
<connector name="node-1">tcp://localhost:61617</connector>
</connectors>
<cluster-connections>
<cluster-connection name="my-cluster">
<connector-ref>node-1</connector-ref>
...
</cluster-connection>
</cluster-connections>
Client connection url
(tcp://localhost:61616?name=node-0,tcp://localhost:61617?name=node-1)?ha=true&reconnectAttempts=-1
3.3. Getting Notified of Connection Failure
JMS provides a standard mechanism for getting notified asynchronously of connection failure: java.jms.ExceptionListener
.
Please consult the JMS javadoc or any good JMS tutorial for more information on how to use this.
The Apache ActiveMQ Artemis core API also provides a similar feature in the form of the class org.apache.activemq.artemis.core.client.SessionFailureListener
Any ExceptionListener or SessionFailureListener instance will always be called by ActiveMQ Artemis on event of connection failure, irrespective of whether the connection was successfully failed over, reconnected or reattached, however you can find out if reconnect or reattach has happened by either the failedOver
flag passed in on the connectionFailed
on SessionfailureListener
or by inspecting the error code on the javax.jms.JMSException
which will be one of the following:
JMSException error codes:
- FAILOVER
-
Failover has occurred and we have successfully reattached or reconnected.
- DISCONNECT
-
No failover has occurred and we are disconnected.
3.4. Application-Level Failover
In some cases you may not want automatic client failover, and prefer to handle any connection failure yourself, and code your own manually reconnection logic in your own failure handler. We define this as application-level failover, since the failover is handled at the user application level.
To implement application-level failover, if you’re using JMS then you need to set an ExceptionListener
class on the JMS connection.
The ExceptionListener
will be called by Apache ActiveMQ Artemis in the event that connection failure is detected.
In your ExceptionListener
, you would close your old JMS connections, potentially look up new connection factory instances from JNDI and creating new connections.
For a working example of application-level failover, please see the Application-Layer Failover Example.
If you are using the core API, then the procedure is very similar: you would set a FailureListener
on the core ClientSession
instances.