StarWind Virtual SAN® is a native Windows hypervisor-centric hardware-less VM storage solution. It creates a fully fault-tolerant and high performing storage pool built for the virtualization workloads by mirroring existing server’s storage and RAM between the participating storage cluster nodes. The mirrored storage resource is then connected to all cluster nodes and treated just as a local storage by all hypervisors and clustered applications. High Availability is achieved by providing multipath access to all storage nodes. StarWind Virtual SAN® delivers supreme performance compared to any dedicated SAN solution since it runs locally on the hypervisor and all I/O is processed by local RAM, SSD cache, and disks. This way, it never gets bottlenecked by storage fabric.
This guide is written for the experienced Windows Server users or system administrators. It provides the detailed instructions on how to set up a 3-node hyperconverged scenario with Hyper-V Cluster on Windows Server 2016 with StarWind Virtual SAN® as a storage provider.
A full set of up-to-date technical documentation can always be found here, or by pressing the Help button in the StarWind Management Console.
Preconfiguring the Servers
The network diagram of the configuration described in this guide is shown in the picture below.
NOTE: Synchronization links can be connected either to the redundant switches or directly between the nodes (recommended). Also, additional network connections may be necessary, depending on the cluster setup and applications requirements. For any technical help with configuring the additional networks, please, do not hesitate to contact StarWind support department via online community forum, or via support form (depends on your support plan).
1. Make sure to have a domain controller and servers added to the domain.
2. Install Failover Clustering and Multipath I/O features, as well as the Hyper-V role on all servers. This can be done through Server Manager (Add Roles and Features menu item).
3. Configure network interfaces on each node to make sure that Synchronization and iSCSI/StarWind Heartbeat interfaces are in different subnets and connected according to the network diagram above. In this document, 172.16.10.x and 172.16.11.x subnets are used for iSCSI/StarWind Heartbeat traffic, while 172.16.110.x, 172.16.111.x and 172.16.112.x subnets are used for the Synchronization traffic.
4. In order to allow iSCSI Initiators discover all StarWind Virtual SAN interfaces, the StarWind configuration file (StarWind.cfg) should be changed after stopping the StarWind service on the node where it will be edited. Locate the StarWind Virtual SAN configuration file (the default path is “C:\Program Files\StarWind Software\StarWind\StarWind.cfg”) and open it via WordPad as Administrator. Find the <iScsiDiscoveryListInterfaces value=”0”/> string and change the value from 0 to 1 (should look as follows: <iScsiDiscoveryListInterfaces value=”1”/>). Save the changes and exit Wordpad. Once StarWind.cfg is changed and saved, the StarWind service can be restarted.
Enabling Multipath Support
5. Open the MPIO manager: Start->Administrative Tools->MPIO.
7. When prompted to restart the server, click Yes to proceed.
NOTE: Repeat the procedures on all the servers.
Configuring Shared Storage
8. Download the StarWind setup executable file from the StarWind website: https://www.starwind.com/registration-starwind-virtual-san
NOTE: The setup file is the same for x86 and x64 systems, as well as for all Virtual SAN deployment scenarios.
Install StarWind Virtual SAN on all nodes, using default settings. Apply a time-limited fully functional evaluation key or a FREE version key. Alternatively, use the already purchased StarWind Virtual SAN commercial license key. After the successful installation, launch the StarWind Management Console: double-click the StarWind tray icon.
If the StarWind service and Management Console are installed on the same server, Management Console will automatically add the local StarWind instance to the console tree after the first launch. Afterwards, the Management Console will automatically connect to it using default credentials. To add remote StarWind VSAN servers to the console, use the Add Server button on the control panel.
9. StarWind Management Console will ask you to specify the default storage pool on the server you are connecting to for the first time. Please configure the storage pool to use one of the volumes you prepared earlier. All devices created through the Add Device wizard will be stored on it. Should you decide to use an alternative storage path for your StarWind virtual disks, please use the Add Device (advanced) menu item.
Press the Yes button to configure the storage pool. Should you require to change the storage pool destination, press Choose path… and point the browser to the necessary disk.
NOTE: Each array that will be used by StarWind Virtual SAN to store virtual disk images has to meet the following requirements:
- Initialized as GPT
- Have a single NTFS-formatted partition
- Have a drive letter assigned
10. Select the StarWind server where you wish to create the device.
11. Press the Add Device (advanced) button on the toolbar.
12. Add Device Wizard will appear. Select Hard disk device and click Next.
13. Select Virtual disk and click Next.
14. Specify the virtual disk location and size. Click Next.
15. Specify the virtual disk options. Click Next.
16. Define the caching policy and specify the cache size (in GB). Click Next.
NOTE: The basic recommendation is to assign 1 GB of L1 cache in Write-Back or Write-Through mode per 1 TB of storage capacity. Yet, the cache size should correspond to the storage working set of the servers.
17. Define L2 caching policy and size if necessary. Click Next to continue.
NOTE: The recommended size of the L2 cache is 10% of the initial StarWind device capacity.
18. Specify the target parameters.
Select the Target Name checkbox to enter a custom name of a target.
Otherwise, the name will be generated automatically based on the target alias.
19. Click Create to add a new device and attach it to the target. Then click Close to close the wizard.
20. Right-click on the Server field and select Add Server. Add new StarWind server, which will be used as a second HA node. Click OK to continue.
21. Right-click on the Server field and select Add Server again. Add new StarWind server which will be used as a third HA node. Click OK to continue.
22. Right-click on the device you just created and select Replication Manager. Then the Replication Manager window will appear. Press the Add Replica button.
23. Select Synchronous Two-Way Replication. Click Next to proceed.
24. Specify the partner server IP address.
Default StarWind management port is 3261. If a different port has been configured, please type it in the Port number field. Click Next.
25. Select the failover strategy for the HA device. For the purposes of this document, the Heartbeat failover strategy is used. Click Next to continue.
26. Choose Create new Partner Device. Click Next.
27. Specify the partner device location if necessary. You can also modify the target name of the device. Click Next.
28. The following screen allows selecting the Synchronization and Heartbeat channels for the HA device. The ALUA settings and the target name of the device can also be modified. Click Change Network Settings to continue.
29. Specify the interfaces for the Synchronization and Heartbeat. Click OK. Then click Next.
NOTE: It is recommended to configure the Heartbeat and iSCSI channels on the same interfaces to avoid the split-brain issue. If the Synchronization and Heartbeat interfaces are located on the same network adapter, it is recommended to assign one more Heartbeat interface to a separate adapter.
30. Select Synchronize from existing Device for the partner device initialization mode. Click Next.
31. Press the Create Replica button. Then click Close to close the wizard.
32. Click Add replica again. Select Synchronous “Two-Way” Replication as a replication mode. Click Next to proceed.
33. Specify the partner server IP address. The default StarWind management port is 3261. If a different port has been configured, type it in the Port Number field. Click Next to continue.
34. Choose Create new Partner Device. Click Next.
35. Specify the partner device location if necessary and/or modify the target name of the device. Click Next.
36. Select the Synchronization and Heartbeat networks for the HA device by clicking Change Network Settings.
37. Specify the interfaces for Synchronization and Heartbeat. Click OK. Then click Next.
38. Select Synchronize from existing Device as a partner device initialization mode. Click Next.
39. Press Create Replica button. Then click Close to close the wizard.
40. The successfully added devices will appear in StarWind Management Console.
Repeat the HA device creation steps for any virtual disks that will be further used as a Cluster Shared Volumes.
Once all devices are created, Management Console should look similar to the screenshot below.
Discovering Target Portals
This part describes how to discover Target Portals on each StarWind node.
41. Launch Microsoft iSCSI Initiator: Start > Administrative Tools > iSCSI Initiator or iscsicpl from the command line interface. The iSCSI Initiator Properties window appears.
42. Navigate to the Discovery tab.
43. Click the Discover Portal button. Discover Target Portal dialog appears. Type in 127.0.0.1.
44. Click on the Advanced button. Select Microsoft ISCSI Initiator as your Local adapter and select your Initiator IP (keep default 127.0.0.1). Confirm the actions to complete the Target Portal discovery.
45. Click the Discover Portal button again.
46. In Discover Target Portal dialog, type in the iSCSI interface IP address of the partner node that will be used to connect the StarWind provisioned targets. Click Advanced.
47. Select Microsoft iSCSI Initiator as the Local adapter, select the Initiator IP in the same subnet as the IP address of the partner server from the previous step. Confirm the actions to complete the Target Portal discovery.
48. Click the Discover Portal button once again.
49. Discover Target Portal dialog appears. Type in the second IP address of the partner node you will use to connect to the secondary mirrors of the HA devices. Click Advanced.
50. Select Microsoft ISCSI Initiator as your Local adapter, select the Initiator IP in the same subnet as the IP address on the partner server from the previous step. Click OK. Then click OK again to complete the Target Portal discovery.
51. Discover Target Portal dialog will appear. Type in the IP address of the partner node you will use to connect to the third mirrors of HA devices.
52. Select Microsoft ISCSI Initiator as your Local adapter, select the Initiator IP in the same subnet as the IP address on the second partner server from the previous step. Click OK. Then click OK again to complete the Target Portal discovery.
53. Click the Discover Portal… button again. Type in the IP address of the partner node you will use to connect to the secondary mirrors of HA devices. Click Advanced.
54. Select Microsoft ISCSI Initiator as your Local adapter, select the Initiator IP in the same subnet as the IP address on the partner server from the previous step. Click OK. Then click OK again to complete the Target Portal discovery.
55. All target portals are added on the first node.
56. Complete the steps (41-55) for the second and third nodes.
57. All target portals are added on the second node.
58. All target portals are added on the third node.
59. Click on the Targets tab. The previously created targets are listed under the Discovered Targets section.
NOTE: If the created targets are not listed, check the firewall settings of the StarWind server as well as the list of networks served by the StarWind server (go to StarWind Management Console -> Configuration -> Network). Alternatively, check the Access Rights tab on the corresponding StarWind VSAN server in StarWind Management Console for any restrictions implemented.
60. Connecting CSV1. Select CSV1 target located on the local server and click Connect.
61. Enable checkboxes as in the image below. Press Advanced.
62. Select Microsoft iSCSI Initiator in the Local adapter text field.
In the Target portal IP, select 127.0.0.1. Click OK.
63. Select the partner target from the second StarWind node and click Connect.
64. Enable checkboxes as in the image below. Press Advanced.
65. Select Microsoft iSCSI Initiator in the Local adapter text field. Select the IP address in the Initiator IP field. Select the corresponding portal IP from the same subnet in the Target portal IP. Click OK.
66. Select the same target as in the previous step and click Connect.
67. Enable all the checkboxes and press Advanced.
68. This time, in Target portal IP, in dropdown list, select another data channel IP addresses for initiator and target portal. Click OK. Click OK again to connect to the target.
69. Select the partner target from the third node and click Connect again.
70. Enable the checkboxes as in the image below. Press Advanced.
71. This time, in Target portal IP, from the dropdown list, select data channel IP addresses from the other subnet for the initiator and target portals. Click OK. Click OK again to connect to the target.
72. Select the same target as in the previous step and click Connect.
73. Enable all the checkboxes and press Advanced.
74. This time, in Target portal IP, from the dropdown list, select another data channel IP addresses for the initiator and target portal. Click OK, then click OK again to connect to the target.
75. Repeat the actions described in the steps (59-74) for the rest of the HA devices.
76. Repeat steps 59-75 of this section on the second and third StarWind nodes, specifying corresponding local and data channel IP addresses. The result should look like in the screenshot below.
NOTE: It is recommended to configure different MPIO policies depending on iSCSI channel throughput. For 1 Gbps iSCSI channel throughput, it is recommended to set Failover Only or Least Queue Depth MPIO load balancing policy. For 10 Gbps iSCSI channel throughput, it is recommended to set Round Robin or Least Queue Depth MPIO load balancing policy.
77. Configure the MPIO policy for each device specifying localhost (127.0.0.1) as the active path. Select a target of Witness located on the local server and click Devices.
78. Devices dialog box will appear. Click MPIO.
79. For the Witness target, set the load balance policy to Failover Only.
NOTE: In case the Failover Only MPIO policy is used, be sure to check that the local path (127.0.0.1) is set to Active, while the partner connection is set to Standby.
80. Check that 127.0.0.1 is the active path by selecting it from the list and clicking Details.
81. Repeat the same steps for all devices on the second and third nodes.
82. Initialize the disks and create partitions on them using computer management snap-in. It is required that disk devices are visible on all three nodes to create the cluster.
NOTE: It is recommended to initialize disks as GPT.
Creating a Cluster
NOTE: To avoid issues during the cluster validation configuration, it is recommended to install the latest Microsoft updates on each node.
83. Open Server Manager. Select the Failover Cluster Manager item from the Tools menu.
84. Click the Create Cluster link in the Actions section of the Failover Cluster Manager.
85. Specify the servers to be added to the cluster. Click Next to continue.
86. Validate the configuration by passing the cluster validation tests: select Yes… and click Next to continue.
87. Specify a cluster name.
NOTE: If the cluster servers get IP addresses over DHCP, the cluster also gets its IP address over DHCP, though this scenario is not advisable. If the IP addresses are set statically, setting the cluster IP address manually is required. Click Next to continue.
88. Make sure that all settings are correct. Click Previous to make any changes. Click Next to proceed.
NOTE: If the Add all eligible storage to the cluster checkbox is enabled, the wizard will try to add all disks to the cluster automatically. The smallest device will be assigned as the cluster witness disk. It is recommended to uncheck this checkbox before clicking Next and assign the cluster witness disk later manually.
89. Upon cluster creation completion, the system displays a report with the detailed information that can be viewed by clicking on the View Report button. Alternatively, click Finish to close the wizard.
Adding Cluster Shared Volumes
Follow the steps below to add the Cluster Shared Volumes (CSV) that are necessary for working with Hyper-V virtual machines.
90. In the Failover Cluster Manager, go to Cluster -> Storage -> Disks. Click Add Disk in the Actions panel, choose StarWind disks from the list and confirm the selection.
91. To configure the cluster witness disk, right-click on Cluster and proceed to More Actions -> Configure Cluster Quorum Settings.
92. Follow the wizard and use the Select the quorum witness option. Click Next.
93. Select Configure a disk witness. Click Next.
94. Select the Witness disk to be assigned as the cluster witness disk. Click Next and press Finish to complete the operation.
95. In Failover Cluster Manager, select a disk. Right-click on it and select Add to Cluster Shared Volumes.
96. Perform the steps above for any other disk in Failover Cluster Manager. The resulting list of disks will look similar to the screenshot below.
97. Once the disks are added to the Cluster Shared Volumes and the network preferences are set, highly available virtual machines can be created on the Cluster Shared Volumes. Select Roles and in the Action tab, click Virtual Machines -> New Virtual Machine. Complete the wizard.
NOTE: To avoid the unnecessary CSV overhead, configure each CSV to be owned by different cluster node. This node should also be the preferred owner of the VMs running on that CSV.
The Windows Failover Cluster will increase availability of services or applications running on it, as well as exclude the downtime during maintenance.
In case of any additional questions, visit the StarWind Support Forum and StarWind Tech Support engineers will gladly assist.