Making Domino Clusters a Key Part of Your Enterprise Disaster

Download Report

Transcript Making Domino Clusters a Key Part of Your Enterprise Disaster

Making Domino
Clusters a Key Part of
Your Enterprise
Disaster Recovery
Architecture
Andy Pedisich
Technotics
© 2011 Wellesley Information Services. All rights reserved.
In This Session …
•
•
Clustering is one of my favorite Domino features
 Introduced with Release 4.5 in December of 1996
 We have seen an increased adoption in clustering Domino
servers over the past two years
 Much of that was for disaster recovery
Some enterprises have been using Domino clustering for disaster
recovery for years
 This session shares as much information as possible about the
topic based on what I’ve seen, heard about, and discovered
creating Domino-based disaster recovery solutions
 And most importantly, the session speaks to the all important
issues about managing failovers
1
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
2
Setting a Baseline for Concepts
•
•
•
First, lets make sure that everyone knows when you see the
letters DR, we’re talking about disaster recovery
Disaster recovery is part of a larger concept referred to as
business continuity
 That’s what we plan to do to keep our businesses running while
disruptive events occur
 These events could be extremely local, like a power outage in
a building
 Or they could be regional, like an earthquake or tornado, or a
disaster caused by humans doing their thing
Disaster recovery has its focus on the technology that supports
business operations
3
Why Use Domino for Disaster Recovery
•
•
•
Domino clustering has been accepted by many organizations as
an important part of their DR infrastructure
 Although that’s not a total endorsement as corporations do
wacky things
Clustering works with just about all things Lotus Notes/Domino
such as
 Email, calendaring, Traveler, BlackBerry, and roaming user
Clustering is included as part of your enterprise server licenses
 Clusters clearly should be exploited in every enterprise
 And they should have a solid role in any DR solution for
messaging and collaboration
4
Important Facts to Help You Sell Clustering for DR
•
•
Here are some facts to back you up when you start making plans
to use Domino clusters for DR
 Clustering is a shrink wrapped solution
 It’s not new. As a matter of fact, it’s been burned in for years.
 It works really well
 It’s automatic
 It’s easy to set up and maintain
 It’s easy to test
The biggest drawback for most companies using clusters is the
increase in storage requirements since their data size is double
 Domino Attachment and Object Services (DAOS) can
reduce that size by 30% to 40%
5
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
6
Basic Clustering Requirements
•
•
All servers in the cluster must:
 Run on Domino Enterprise or Domino Utility server
 Be connected using high-speed local area network (LAN) or a
high-speed wide area network (WAN)
 Use TCP/IP and be on the same Notes named network
 Be in the same Domino domain
 Share a common Domino directory
 Have plenty of CPU power and memory
 It’s safe to say that clustered servers need more power and
more disk resources than unclustered servers
A server can be a member of only one cluster at a time
7
Technical Considerations for Disaster Recovery Clusters
•
•
A few guidelines regarding configuring clusters, especially for
clusters used for disaster recovery
 Servers in a cluster should not use the same power source
 Nor should they use the same network subnet or routers
 They should not use the same disk storage array
 They should not be in the same building
Your clustered DR solution should ideally be in different cities
 Never in the same room
8
Keep Your Distance
•
•
•
•
•
Good DR cluster designs should take into account both local and
regional problems
Consider a financial company who had clustered servers in two
separate buildings across the street from each other in Manhattan
 This firm now has primary servers in offices in New York City
 And failover servers are thousands of miles away
Another firm has primary servers in Chicago
 With a failover server in the UK
A college has primary and failover servers separated by 200 miles
Another company we know is just starting out with DR and has
servers 25 miles from each other
 A good start, but they really want more distance
9
Servers Accessed Over a Wide Area Network
•
•
•
If servers are as far away as they are supposed to be, there might
be some latency in the synchronization of the data
 If this is the case, users might not find the failover server is up
to date with their primary server during a failover
Everyone must be aware that this is a possibility
 Expectations must be set
 Or management needs to provide budgets for better
networking
Work out all of these details in advance with management
 Get them written down and approved so there are no surprises
10
Most Common DR Cluster Configuration
•
The most common DR cluster configuration is active/passive
 Servers on site are active
 Servers at the failover site are passive, waiting for failover
 Sometimes domains use this failover server as a place to do
backups or journaling
 The number of servers in the cluster usually vary


There could be 1 active and 1 passive
Or 2 or 3 active and 1 passive
11
The Active/Passive Clustered Servers
•
•
Active/Passive servers
 Cluster has two servers: one active, the other not generally
used or used for backing up
 Very common disaster recovery setup
Each server holds replicas of all the files from the other server
 During failover all users flip to the other server
12
The 3 Server Cluster for DR
•
•
Three or more servers in the cluster
 There are two replicas for each cluster supported application or
mail file
Each primary server holds the mail files of the users assigned to
that server
 Replicas of mail file from both primary servers are on the
failover
13
3 Server Cluster with One Primary Server Down
•
If a primary server goes down, users from that server go to the
failover server
 Easy to understand, and you save yourself a server
 You’ll still need twice the total disk space of Mail1 and Mail2
 What happens when both Mail1 and Mail2 are unavailable
14
Both Primary Servers Are Down
•
•
If both primary servers are down, the last server in the cluster has
to support everyone
 Remember that generally speaking only about 60% to 70% of
users assigned to a server are on there concurrently
 Still that has to be a pretty strong server with fast disks
Some sites have remote servers as primary and failover happens
at the home office data center
15
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
16
Understanding Cluster Replication
•
•
Cluster replication is event driven
 It doesn’t run on a schedule
 The cluster replicator detects a change in a database
and immediately pushes the change to other replicas
in the cluster
If a server is down or there is significant network latency, the
cluster replicator stores changes in memory so it can push them
out when it can
 If a change to the same application happens before a previous
change has been sent, the CLREPL gathers them and sends
them all together
17
Streaming Cluster Replication
•
•
R8 introduced Streaming Cluster Replication (SCR)
 This newer functionality reduces replicator overhead
 Provides reduction in cluster replicator latency
As changes occur to databases, they are captured and
immediately queued to other replicas in the same cluster
 This makes cluster replication more efficient
18
SCR Only Works with R8
•
•
If one server in the cluster is R8 another is not, Domino will
attempt SCR first
 When that doesn’t work, it will fall back on standard cluster
replication
If you’d like to turn off SCR entirely to ensure compatibility, use
this parameter
 DEBUG_SCR_DISABLED=1
 This must be used on all cluster mates
19
Only One Cluster Replicator by Default
•
•
When a cluster is created, each server has only a single cluster
replicator instance
 If there have been a significant number of changes to many
applications, a single cluster replicator can fall behind
 Databases synchronization won’t be up to date
If a server fails when database synch has fallen behind, users will
think their mail file or app is “missing data”
 They won’t understand why all the meetings they made this
morning are not there
 They think their information is gone forever!
 Users need their cluster insurance!
20
Condition Is Completely Manageable
•
•
•
•
Adding a cluster replicator will help fix this problem
You can load cluster replicators manually using the following
console command
 Load CLREPL
 Note that a manually loaded cluster replicator will not be
there if the server is restarted after manually loading a
cluster replicator
Add cluster replicators permanently to a server
 Use this parameter in the NOTES.INI
 CLUSTER_REPLICATORS=#
I always use at least two cluster replicators
21
When to Add Cluster Replicators
•
•
•
But how do you tell if there’s a potential problem?
 Do you let it fail and then wait for the phone to ring?
 No!
You look at the cluster stats and get the data you need to make an
intelligent decision
 Adding too many will have a negative effect on server
performance
Here are some important statistics to watch
22
Key Stats for Vital Information About Cluster Replication
Statistic
What It Tells You
Acceptable values
Replica.Cluster.
SecondsOnQueue
Total seconds that last DB
replicated spent on work queue
< 15 sec – light load
< 30 sec – heavy
Replica.Cluster.
SecondsOnQueue.Avg
Average seconds a DB spent on Use for trending
work queue
Replica.Cluster.
SecondsOnQueue.Max
Maximum seconds a DB spent
on work queue
Use for trending
Replica.Cluster.
WorkQueueDepth
Current number of databases
awaiting cluster replication
Usually zero
Replica.Cluster.
WorkQueueDepth.Avg
Average work queue depth
since the server started
Use for trending
Replica.Cluster.
WorkQueueDepth.Max
Maximum work queue depth
since the server started
Use for trending
23
What to Do About Stats Over the Limit
•
•
Acceptable Replica.Cluster.SecondsOnQueue
 Queue is checked every 15 seconds, so under light load,
should be less than 15
 Under heavy load, if the number is larger than 30, another
cluster replicator should be added
If the above statistic is low and Replica.Cluster.WorkQueueDepth
is constantly higher than 10 …
 Perhaps your network bandwidth is too low
 Consider setting up a private LAN for cluster
replication traffic
24
Stats That Have Meaning but Have Gone Missing
•
There aren’t any views in Lotus version of Statrep that let you see
these important statistics
 Matter of fact, the Cluster view is pretty worthless
25
The Documents Have More Information
•
The cluster documents have much better information
 You can actually use the data in the docs
 But they still lack key stats, though they’re in each doc
26
Stats That Have Meaning but Have Gone Missing
•
But there is a view like that in the Technotics R8.5 Statrep.NTF
 It shows the key stats you need
 To help track and adjust your clusters
 It is included on the CD for this conference
27
My Column Additions to Statrep
Column Title
Formula
Formatting
Min on Q
Replica.Cluster.SecondsOnQueue / 60
Fixed (One
Decimal Place)
Min/Q Av
Replica.Cluster.SecondsOnQueue.Avg / 60
Fixed (One
Decimal Place)
Min/Q Mx
Replica.Cluster.SecondsOnQueue.Max / 60
Fixed (One
Decimal Place)
WkrDpth
Replica.Cluster.WorkQueueDepth
General
WD Av
Replica.Cluster.WorkQueueDepth.Avg
General
WD Mx
Replica.Cluster.WorkQueueDepth.Max
General
28
Use a Scheduled Connection Document Also
•
Back up your clustered replication with a scheduled connection
document between servers
 Have it replicate at least once per hour
 You’ll always be assured to have your servers in sync even if
one has been down for a few days
 And it replicates deletion stubs too!
29
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
30
Busy Clusters Might Require a Private LAN
•
•
A private LAN separates the network traffic the cluster creates for
replication and server probes
 And will probably leave more room on your primary LAN
Start by installing an additional network interface card (NIC) for
each server in the cluster
 Connect the NICs through a private hub or switch
31
Setting Up the Private LAN
•
•
Assign second IP address to additional NIC
Assign host names to the addresses in the local HOSTS file on
each server
 Using DNS is a best practice
 10.200.100.1 mail1_clu.domlab.com
 10.200.100.2 mail2_clu.domlab.com
 Test by pinging the new hosts from each server
32
Modify Server Document
•
For each server in the cluster, edit the server document to enable
the new port
33
Set Parameters so Servers Use Private LAN
•
Make your clusters use the private LAN for cluster traffic by
establishing the ports in the server NOTES.INI with these
parameters
 CLUSTER=TCP,0,15,0
 PORTS=TCPIP,CLUSTER
 CLUSTER_TCPIPADDRESS=0,10.200.100.2:1352
 You will use the address of your NIC card
34
Parameters to Make the Cluster Use the Port
•
•
Use the following parameter to ensure Domino uses the port for
cluster traffic
 SERVER_CLUSTER_DEFAULT_PORT=CLUSTER
Use this parameter just in case the CLUSTER port you’ve
configured isn’t available
 SERVER_CLUSTER_AUXILIARY_PORTS=*
 This allows clustering to use any port if the one you’ve
defined isn’t available
35
Keep Users Off the Private LAN
•
•
To keep users from grabbing on to the private LAN port, take the
following steps
 Create a group called ClusterServers
 Add the servers in the cluster to this group
Add the following parameter to the NOTES.INI of both servers
 It will keep users from connecting through the CLUSTER port
 Allow_Access_Cluster=ClusterServers
36
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
37
Respect the Users
•
•
Clustering provides outstanding service levels for users
 But the process of failing over is sometimes hard on users
 Failover is actually the most difficult moment for users
And sometimes errors in network configuration might prevent
successful failover
 For example, the common name of the server should be listed
as an alias in DNS to ensure users can easily open their
application on the servers
 If the server is not in DNS, the clients won’t know how to get
to the failover servers
38
Best Practice for Cluster Management
•
•
Best Practice:
 Don’t take a clustered server down during working hours
unless it is absolutely necessary
 A non-planned server outage, such as a crash or power
failure, is a legitimate reason to fail over
Resist the urge to take a server down because you know it’s
clustered
 You could probably do it, but the risk of a hard failover will
probably cause unwanted help desk calls
39
Easiest Cluster Configuration to Manage
•
•
The Active/Passive model of clustering is by far the easiest to
manage
Use parameters in the NOTES.INI file on the servers in the cluster
that allow users on the primary one server
 But don’t allow them on the failover server
40
Check for Server Availability
•
The parameters we use are thresholds that check the cluster
statistic Server.AvailabilityIndex (SAI)
 This statistic shows how busy the server is
 100 means it’s not busy at all
 0 means it’s crazy busy
41
Adjusting the Threshold
•
•
Setting the parameter Server_availablity_threshold controls
whether users can access the server
 50 means if the SAI is above 50, then failover users to another
server
 A setting like this can be used for load balancing
 100 means the SAI must be 100, which means the server must
be 100% available
 This translates into “nobody is allowed on the server”
 0 means that load balancing and checking the SAI is turned off
These thresholds can come in handy
42
Setting Up Active/Passive Servers in a Cluster
•
•
•
•
Let’s look at the following scenario
 Mail1 is the active primary; Mail2 is the passive failover
To allow users to access their primary server, use this parameter
in the NOTES.INI of Mail1
 Server_availablity_threshold = 0
 Use this console command:
 Set config server_availability_threshold=0
Prevent users from accessing failover server Mail2; use this
parameter
 Server_availablity_threshold = 100
 Use this console command:
 Set config server_availability_threshold=100
Administrators will still be able to access this server
43
Mail1 Is Crashing
•
•
•
If Mail1 crashes, the Notes client will disregard our setting of 100
on Mail2 and users will be permitted
To help stabilize the system use this parameter on Mail2
 Server_availability_threshold = 0
 Let all users aboard
While Mail1 is down, enter this parameter into the NOTES.INI to
prevent users from connecting:
 Server_restricted=2
 2 will keep the setting after a server restart
 Setting it to 1 also keeps them off, but the setting will be
disabled with a 0 after restarting the server
44
Recovering After a Crash
•
•
•
•
When Mail1 is brought back up after the crash, no one will be
permitted to access it except administrators
 That’s because of the Server_restricted=2 setting
 Leave it that way until the end of the day
The ugliest part about failing over is the client part
 Clients are working just fine on Mail2
 By the way, iMap and POP3 users still have access to Mail1
At the day’s end, switch the Server_availability_threshold back to
100 on Mail2 and 0 on Mail1
Issue this console command on Mail2
 Drop all
45
Taking a Clustered Server Down Voluntarily
•
•
•
If you must take clustered Mail1 server down, set
Server_restricted=2, then drop all at the console
 Remember that POP3 and iMap users won’t like you very much
Set Mail2 to Server_availability_threshold to 0
Don’t forget to set the Server_Restricted back to 0 when the crisis
has passed
 I know someone who forgot to do this a couple of times
 This person could access the server and work because he
was in an administrator’s group
 However nobody else could get on the server and he made
all the users very angry
46
Triggering Failover
•
•
You can set the maximum number of concurrent NRPC users
allowed to connect to a server
 Server_MaxUsers NOTES.INI variable
 Set variable to a number determined in planning stage
Set variable using console command
 Or use NOTES.INI tab in server configuration document
 Set config Server_MaxUsers = desired maximum number of
active concurrent users
 Additional users will fail over to the other members of
the cluster
47
Load Balancing
•
•
If you’d like to load balance your servers, determine what your
comfort range is for how busy your servers are and set the
Server_availablity_threshold accordingly
 Perhaps start with a value of 60
 Users should fail over when the SAI goes below 60
Pay close attention to the SAI in the STATREP.NSF, which is
listed under the Av Inx column
 Some hardware can produce inaccurate SAI reading and cause
users to failover when it’s not necessary
48
SAI Is Unreliable in Some Cases
•
Note how in this case, the Server Availability Index never seems
to get much above 50 consistently
 Users would be failing over constantly
 And if both servers had the issue, users would be bouncing
back and forth between the clustered servers
49
The Expansion Factor
•
Cluster members determine their workload based on the
expansion factor
 This is calculated based on response times for recent requests
 Server compares recent response time to minimum response
time that the server has completed
 Example: Server currently averages 12ms for DBOpen
requests; minimum time was 4ms
 Expansion factor = 3 (current time/fastest time)
 This is averaged over different types of transactions
 Fastest time is stored in memory and in LOADMON.NCF
 LOADMON.NCF is read each time server starts
50
The Expansion Factor (cont.)
•
•
But sometimes Domino has a difficult time calculating the
expansion factor
 The result is that the Server_AvailabilityIndex is not a reliable
measure of how busy the server is
 This can happen with extremely high performing servers
If you see a very low Server_AvailabilityIndex at a time you know
servers are supposed to be idle and you are trying to load
balance, there is something you can do to correct it
 And Domino can help!
51
Changing Expansion Factor Calculation
•
•
Use this parameter to change how the Expansion Factor is
calculated
 SERVER_TRANSINFO_RANGE=n
To determine the optimal value for this variable:
 After the server has experienced heavy usage use this console
command:
 Show AI
 This means, show the availability index calculation

It has nothing to do with that 2001 Steven Spielberg movie about the
robot that looks like a child and tries to become a real boy
52
An Easy Way to Find the Parameter Value
•
Show AI is a console command that has been around since
Domino Release 6
 It runs some computations on the server
 And suggests a SERVER_TRANSINFO_RANGE for you
53
Events to Monitor When Using Clusters
•
Use event monitoring to look for certain problems that could ruin
your clustering by preventing replication
 Look for the following phrases
 “cannot replicate database”
 “folder is larger than supported”
 They both mean that users are going to hate you in the
event of a failover

Because their databases will not be in synch
54
Disaster Recovery Is a Special Case
•
•
Many enterprises configure their clusters for manual failover
 They don’t want the user to fail over unless they permit it
To keep users off of the failover servers, use the following
parameters in the NOTES.INI of the server
 Server_restricted=2
 This will keep off all users until you set it to a zero
 Server_availability_threshold=100
 The server looks busy and keeps everyone on the primary
 Set it to zero to allow users on the server
 Keep in mind, if the primary servers are down, the users
won’t have a server to work on unless you manually reset
these parameters
55
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
56
Things That Do Not Cluster with Domino Clustering
•
•
•
Only traffic over NRPC port 1352 will fail over
Other TCP/IP services will not fail over, including:
 IMAP
 POP3
 HTTP
This means iNotes users will not fail over
 But that doesn’t mean you can’t make it work
 In the aftermath of Katrina, some companies found that
browser-based email was a savior
57
High Availability Is Not Disaster Recovery
•
•
•
IBM/Lotus has several recommendations for achieving high
availability with iNotes
For example, Internet Cluster Manager
 This can redirect the initial URL for an application to one of
several back-end mail servers
 However, if the mail server becomes unavailable after a
session is started, there is a way to recover and switch to
another server that has a replica of the mail file
IBM/Lotus also recommends an HTTP load balancer to help with
high availability, and with DR to a certain extent
 The issue is, what happens when the load balancer is
unavailable?
58
iNotes Users Accessing a Hub Server
•
•
The IWAREDIR.NSF is an application that helps direct browser
requests to a user’s mail server
 Any user on either server Mail1 or Mail2 connects to either
iNotes server that is deployed using the IWAREDIR.NSF
 The IWAREDIR.NSF looks up their name in the Domino
directory and “redirects” them to their mail server
 But it is not designed to take them to a failover server
However, it is possible to have a second version of this file with
code changes that will take a user to their failover mail server
59
Open the IWAREDIR.NSF Using the Designer Client
•
Open the AutoLogin form using the Notes Designer Client
 At the very bottom, you’ll find a field called $$HTMLHead
 This contains the code that discovers the user’s mail server
and mail file name
60
Section of Code That Can Be Modified
•
In this field is code that can be modified to point the user to the
failover server rather than to their primary mail server
 There’s a link at the end of this presentation to some extreme
programming you can do with this field
61
Code to Point Users to Failover Server
•
If the address book says the user’s mail is on Mail1, but you want
to take them to server Mail1F, the failover box, the code can be
changed to something like this
 This is an @if statement that essentially says, if the directory
states that the MailServer for the user is Mail1, take them to
Mail1F, the failover server
 This failover version of the IWAREDIR.NSF that could be
called IWAREDIRF.NSF
62
During Normal Operation
•
During a normal day with no need for failover, the IWAREDIR.NSF
would be specified as the home URL for the server
 Users would be taken to their home servers and their mail files
automatically
63
During a Failover Situation
•
When required, this configuration can be changed to use the
failover version, the IWAREDIRF.NSF
 After this change is made, HTTP must be restarted
64
Restarting the HTTP Task
•
•
Although the HTTP task can be restarted, the most reliable
method is to shut down HTTP and load it again
From the Domino console use these commands
 Tell HTTP quit
 Load HTTP
65
iNotes Is There When You Need It
•
When there are no fat Notes clients, iNotes can have extreme
value and maintain communications in your organization
 It’s worth planning it out so that when users call the help desk
during an emergency they can point the users to any DR server
that is configured to use the IWAREDIRF.NSF
 That’s the failover version of your redirector
66
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
67
7 Rules for Clustering Domino Servers for Disaster
Recovery
1.
2.
3.
4.
5.
6.
7.
Get agreement from all parties before you start configuring
Be sure everyone agrees on what failover means and what kind of
disaster you are preparing for, such as local or regional outages
Keep primary and failover servers as far apart as possible
Actually test the failover scenarios so that there is no doubt that
the configuration works
Consider manual failover to prevent users from accessing servers
over a wide area network or slow connections
Review cluster statistics regularly to ensure there are enough
cluster replicators
Review the CLUDBDIR.NSF to make sure there is a failover replica
for each database on the primary server
68
What We’ll Cover …
•
•
•
•
•
•
•
•
Convincing management that DR clusters rock
Exploring the choices for a clustered DR architecture
Mastering cluster replication
Setting up a private LAN for cluster traffic
Managing cluster failover and load balancing
Understanding the role of iNotes in your DR solution
Reviewing the 7 important rules for configuring DR clusters
Wrap-up
69
Resources
•
•
•
Achieving high availability with IBM Lotus iNotes
 www10.lotus.com/ldd/dominowiki.nsf/dx/Achieving_high_availability
_with_IBM_Lotus_iNotes
ServersLookup Form for IWAReder to return User replicas for
Load Balancers and Reverse Proxies to use
 www10.lotus.com/ldd/dominowiki.nsf/dx/ServersLookup_Form_for_I
WAReder_to_return_User_replicas_for_Load_Balancers_and_R
everse_Proxies_to_use
Understanding IBM Lotus Domino server clustering
 www.ibm.com/developerworks/lotus/documentation/d-lsdominoclustering/index.html
70
Resources (cont.)
•
How to test cluster failover on one Domino database (application)
 www304.ibm.com/support/docview.wss?rs=899&uid=swg21280021
71
7 Key Points to Take Home
•
•
•
•
Disaster recovery is part of the bigger concept known as business
continuity
The biggest drawback when using DR clusters is disk space
consumption, which can be decreased by implementing Domino
Attachment and Object Service
A three-server cluster, with two primary servers and one failover
they share, is an excellent way to conserve server resources and
licensing costs
Make sure your clustered servers are configured with scheduled
connection documents because deletion stubs don’t replicate via
cluster replication
72
7 Key Points to Take Home (cont.)
•
•
•
Use the Allow_Access_Portname parameter to prevent users from
accidently using the private LAN between clustered servers
Failing over is the hardest part for users
iNotes is an excellent way to keep users connected during a
disaster
73
Your Turn!
How to contact me:
Andy Pedisich
[email protected]
HTTP://www.andypedisich.com
74