Background
"Look Ma, No SAN..."Occasionally the need arises to demonstrate the functionality of a cluster that is built using Microsoft Clustering Service (MSCS) or Failover Clustering (new name in 2008/2008 R2) utilizing an environment without a SAN. This might arise from needing to demonstrate basic functionality to a client (using a reasonably powerful laptop/projector) or performing a specialized test case against a certain set of conditions or cluster/machine state. This is also useful for demonstrating proof-of-concept failover/recovery scenarios. Although this solution may be useful for creating failover nodes on multiple Hyper-V host systems, my intention is to show how a clustered solution can be built using a single Hyper-V node.
Note that this does not create a highly available or fault tolerant solution because the VMs are hosted on a single Hyper-V host on a single piece of hardware. The intention is to use the environment for a specific test (possibly involving a patch or new software release) or in a consulting/sales environment to demonstrate the functionality of a product or provide a proof of concept prior to implementation. Additionally, this does not create a clustered VM that is compatible with other Hyper-V 2008 R2 features like Live Migration.
Although I create this solution using Hyper-V, the same solution can be easily implemented on any other virtualization platform that supports guest networking (VMWare, Citrix XenServer, etc).
One of the motivations behind this post is that a virtual disk in Hyper-V can only be mapped to a single Hyper-V guest. Trying to map a single disk to multiple hosts gives the following error: "Error Applying Hard Drive Changes: Failed to add device ' Microsoft Virtual Hard Disk', Attachment <Path> failed to open because of this error: 'The process cannot access the file because it is being used by another process.'"
Since a MSCS/Failover cluster needs shared storage, this requires a centralized network-based storage solution. Most SANs are expensive (and don't fit in a typical laptop bag). Solutions on other platforms exist (such as OpenFiler), but on the Microsoft platform, enter the Microsoft iSCSI Software Target. This is a tool that is currently only compatible with Windows Server 2008 R2 and Windows Server 2008 R2 SP1 and allows the creation of a virtual disk that acts as an iSCSI target for other initiators (VMs and non-VMs).
Designing the Solution
To demonstrate the solution, I designed/implemented a cluster using Microsoft SQL Server 2008 R2 on two Windows Server 2008 R2 VMs. I also built a DC that also acts as a storage server for the iSCSI targets and provides a domain environment. This fits my purposes for this post because I typically build a new environment for each group of posts that I write and it allows me to build smaller specific-purpose demonstration environments using less hardware. I also don't need to install any software or add any unnecessary roles/features to the Hyper-V host.
From an architectural perspective, placing the shared storage on the Hyper-V host instead of a guest will provide the best performance because the overhead of accessing the storage through the host will be less than accessing the storage through the VM (and subsequently the host). Otherwise, resource allocations involving memory and CPU should provide the VMs with sufficient resources to run the hosted application(s) with sufficient performance (whatever that means in your specific case).
Implementing the Solution
Implementation of the clustered SQL Instance involves three main steps:- Initially configure VMs and perform iSCSI configuration
- Establish the Failover cluster with the nodes
- Perform a SQL cluster preparation/configuration on the first node and join the second node to the established cluster
To implement, install the iSCSI software target on the storage server.
Then create the virtual disks that the iSCSI initiators will access:
To set up access to the volumes, I usually use iSCSI Qualified Names (IQNs). The easiest way to get these to appear is to configure the iSCSI client on the initiating hosts and scan the IP address of the storage server. Verify via the Discovery tab that the storage server is being used to discover possible targets.
Now, back on the storage server, configure the targets and associate them with the volumes. Note that when it comes time to add the IQNs for the target, use the "advanced" button to add multiple IQNs.
After the targets for the Quorum drive and the SQL volume have been created, add the targets to the volume configuration.
Back on the servers, connect to the targets and configure devices.
At this point, set up the filesystems on the two volumes, otherwise the cluster will be created as a node majority cluster and a single node failing will cause the cluster to fail. If you forget to do this, you can use the "Configure Cluster Quorum Wizard" later on and configure the shared disk resources manually.
Now the cluster setup and SQL configuration can begin. Ensure the Failover Clustering feature is added through server manager, and add the two virtual SQL nodes to the cluster.
Once the cluster is created and validated, the next step is to prepare/install the SQL cluster.
After getting through the preparation and completion wizards, a single node cluster on the first node is set up. To add the second node, complete the "Add node to a SQL Server failover cluster" wizard on the second node.
Verification of the Solution
Verification of the solution is fairly straightforward. There are immediately three main things to test with the clustered solution:- Connection to the SQL Instance virtual hostname/IP via SQL Server Management Studio
- Failover of SQL Cluster resources on node failure (simulated by hard power off of one of the VMs)
- Automatic re-establishment of connection to the cluster after the connection is dropped due to failover
Run a query and fail the primary node...
Watch the failover to the secondary node.
Next, verify via SSMS that a query can be completed after failover.
And that concludes a demonstration of SQL clustering. This setup was configured on a single Hyper-V host with a virtual two-node SQL cluster. The back end storage could not be supplied using VHDs in Hyper-V, so the Microsoft iSCSI software target was used to create shared storage for the cluster nodes.
Hi, great post. So if I understand correctly,u can use iSCSI with internal storage/disk array?
ReplyDeleteYes, you are creating a virtual disk that acts as the iSCSI target. This can potentially be hosted on a VM or on the host system. There is also a wide range of possibilities regarding where to put the vhd file, including direct atttached storage, flash drives, USB/eSATA/Firewire external drives. Like everything else though, there are good ideas and bad ideas about how to architect this type of solution and it is mainly applicable for a sales/consulting or proof-of-concept environment.
ReplyDeleteIt may have been some time since this post was posted. Firstly, great detailed description, well done. Secondly - I just want to confirm that I have understood correctly...
ReplyDeleteThis setup involves using:
- 1 Physical Computer
- 2 Virtual Machines (Clustered)
I come to a roadblock trying to understand the concept behind iSCSI. I know that a shared volume/s are required for a cluster to work, but for testing purposes in this case, can the shared volume simply be a virtual disk or a physical partition on the same 1 physical computer being used?
In my case, the setup involved 3 VMs
Delete- 1 DC that also hosted the storage volume
- 2 cluster nodes sharing the SQL instance
For your setup, it sounds like you'll want to have the iSCSI software target installed on the physical server and have network connectivity between the VMs and the physical server.
iSCSI is really just the transport protocol that allows a server to access a remote storage device over TCP/IP, in the case of the Microsoft iSCSI Software Target, the remote storage device is a VHD stored on a server somewhere that the cluster nodes can access over the network.
If you were to use other storage that supports iSCSI natively (say EMC or Dell EqualLogic, there are also tons of others), then you would configure the storage accordingly.
Back to the iSCSI software target, the VHD needs to be a file, so the physical partition would need to have a volume (preferably NTFS) to create the VHD that will act as the remote storage device for the cluster nodes.
Make sense?