StarWind Virtual SAN® for vSphere 2-Node Hyperconverged Scenario with VMware vSphere 7

Introduction to StarWind Virtual SAN for vSphere

StarWind Virtual SAN for vSphere comes as a prepackaged Linux virtual machine to be installed as a VM on vSphere. It creates a VM-centric and high performing storage pool for a VMware сluster.
This guide describes the deployment and configuration process of the StarWind Virtual SAN with VMware vSphere.

StarWind Virtual SAN for vSphere VM requirements

Prior to installing StarWind Virtual SAN Virtual Machines, please make sure that the system meets the requirements, which are available via the following link: https://www.starwindsoftware.com/system-requirements

Recommended RAID settings for HDD and SSD disks:
https://knowledgebase.starwindsoftware.com/guidance/recommended-raid-settings-for-hdd-and-ssd-disks/

Please read StarWind Virtual SAN Best Practices document for additional information: https://www.starwindsoftware.com/resource-library/starwind-virtual-san-best-practices

Pre-Configuring the Servers

The diagram below illustrates the network and storage configuration of the solution:

StarWind VSAN

1. ESXi hypervisor should be installed on each host.

2. StarWind Virtual SAN for vSphere VM should be deployed on each  ESXi  host from an OVF template, downloaded on this page: https://www.starwindsoftware.com/release-notes-build-vsan-for-vmware-vsphere

3. The network interfaces on each node for Synchronization and iSCSI/StarWind heartbeat interfaces should be in different subnets and connected directly according to the network diagram above. Here, the 172.16.10.x subnet is used for the iSCSI/StarWind heartbeat traffic, while the 172.16.20.x subnet is used for the Synchronization traffic.

NOTE: Do not use ISCSI/Heartbeat and Synchronization channels over the same physical link. Synchronization and iSCSI/Heartbeat links and can be connected either via redundant switches or directly between the nodes.

vCenter Server can be deployed separately on another host or as VCSA on StarWind VSAN highly-available storage, created in this guide.
StarWind HyperConverged Appliance is a turnkey hyper-converged hardware platform fitted into a small two-node footprint. You don’t need anything else to build a budget-friendly new IT infrastructure or upgrade an existing one. All your systems will be “babysitted” by StarWind 24/7/365, troubleshooting any concerns without your involvement. Everything’s operated through a neat web UI. We’ll also migrate your workloads at no extra cost.
Dramatically decrease your CapEx, OpEx, and IT management costs, while visibly increasing return on investment (ROI) with hyperconvergence for ROBO, SMB & Edge from StarWind. Hyperconverged Appliance from StarWind HCA Datasheet

Preparing Environment for StarWind VSAN Deployment

Configuring Networks

Configure network interfaces on each node to make sure that Synchronization and iSCSI/StarWind heartbeat interfaces are in different subnets and connected physically according to the network diagram above. All actions below should be applied to each ESXi server.

NOTE: Virtual Machine Port Group should be created for both iSCSI/ StarWind Heartbeat and the Synchronization vSwitches. VMKernel port should be created only for iSCSI traffic. Static IP addresses should be assigned to VMKernel ports.

NOTE: It is recommended to set MTU to 9000 on vSwitches and VMKernel ports for iSCSI and Synchronization traffic. Additionally, vMotion can be enabled on VMKernel ports.

1. Using VMware ESXi web console, create two standard  vSwitches: one for the iSCSI/ StarWind Heartbeat channel (vSwitch1) and the other one for the Synchronization channel (vSwitch2).

Create vSwitch

2. Create a VMKernel port for the iSCSI/ StarWind Heartbeat channel.

Add VMKernel

3. Add a Virtual Machine Port Groups on the vSwitch for iSCSI traffic (vSwtich1) and on the vSwitch for Synchronization traffic (vSwitch2).

Add port group
4. Repeat steps 1-3 for any other links intended for Synchronization and iSCSI/Heartbeat traffic on ESXi hosts.

Deploying StarWind Virtual SAN for vSphere

1. Download zip archive that contains StarWind Virtual SAN for vSphere: https://www.starwindsoftware.com/starwind-virtual-san#download
2. Extract virtual machine files from the downloaded archive.
3. Deploy a virtual machine on each ESXi host using the “Create/Register VM” button. Select “Deploy a virtual machine from an OVF or OVA file” in the Select creation type section and press Next.

Select_creation_type
4. Specify the name for the virtual machine with StarWind VSAN, drag and drop the extracted files to the wizard, and press Next.

Select_OVF_and_VMDK_filess
5. Specify the location for the StarWind Virtual SAN VM and press Next.

Select_storage
6. Read and accept the license agreements by pressing on “I agree” button. Press Next to continue.

License_agreement
7. Choose the proper networks for the VM by their purpose and other options.

Deployment_options
8. Review the settings and click Finish to start the deployment process.

OVF_Ready_to_Complete

12. Repeat all the steps from this section on the other ESXi hosts.

NOTE: In some cases, it’s recommended to reserve memory for StarWind VSAN VM.

NOTE: When using StarWind with the synchronous replication feature inside of a Virtual Machine, it is recommended not to make backups and/or snapshots of the Virtual Machine with the StarWind VSAN service installed, as this could pause the StarWind Virtual Machine. Pausing the Virtual Machines while the StarWind VSAN service in under load may lead to split-brain issues in synchronous replication devices, thus to data corruption.

Configuring StarWind VMs Startup/Shutdown

1. Setup the VMs startup policy on both ESXi hosts from Manage -> System tab in the ESXi web console. In the appeared window, check Yes to enable the option and choose the stop action as Shut down. Click Save to proceed.VM_Auto_Start

2. To configure a VM autostart, right-click on the VM, navigate to Autostart and click Enable.

VM_Auto_Start_Enable3. Complete the actions above on StarWind VM located on all ESXi hosts.
4. Start the virtual machines on all ESXi hosts.

Configuring StarWind Virtual SAN VM settings

By default, the StarWind Virtual SAN virtual machine receives an IP address automatically via DHCP. It is recommended to create a DHCP reservation and set a static IP address for this VM. In order to access StarWind Virtual SAN VM from the local network, the virtual machine must have access to the network. In case there is no DHCP server, the connection to the VM can be established using the VMware console and static IP address can be configured manually.

1. Open a web browser and enter the IP address of the VM, which it had received via DHCP (or had it assigned manually), and log in to StarWind Virtual SAN for vSphere using the following default credentials:

Username: user
Password: rds123RDS
NOTE:  Make sure to tick Reuse my password for privileged tasks check box.Login_screen2. After the successful login, on the left sidebar, click Accounts.
3. Select a user and click Set Password.
Set_password
4.
 On the left sidebar, click Networking.
Networking

Here, the Management IP address of the StarWind Virtual SAN Virtual Machine, as well as IP addresses for iSCSI and Synchronization networks can be configured.
In case the Network interface is inactive, click on the interface, turn it on, and set it to “Connect automatically“.Networking_auto

5. Click on Automatic (DHCP) to set the IP address (DNS and gateway – for Management).
Networking_IPv4_Settings

6. The result should look like on the picture below:Networking_result
NOTE: It is recommended to set MTU to 9000 on interfaces, dedicated for iSCSI and Synchronization traffic. Change Automatic to 9000, if required.Networking_MTU

6. Alternatively, log in to the VM via the VMware console and assign a static IP address by editing the configuration file of the interface located by the following path: /etc/sysconfig/network-scripts
Networking_manual

7. Open the file, corresponding to the Management interface  using text editor, for example:
sudo nano /etc/sysconfig/network-scripts/ifcfg-ens192
8. Edit the file:
Change the line BOOTPROTO=dhcp to: BOOTPROTO=static

Add the IP settings needed to the file:
IPADDR=192.168.12.10
NETMASK=255.255.255.0
GATEWAY=192.168.12.1
DNS1=192.168.1.1
By default, the Management link should have an ens192 interface name. The configuration file should look as follows:
Networking_edit
9.
Restart interface using the following cmdlet: sudo ifdown ens192 sudo ifup ens192 or restart the VM.
10. Change the Host Name from the System tab by clicking on it:
Host_Name
11. Change System time and NTP settings if required:
System_Time
12.  Repeat the steps above on each StarWind VSAN VM.

Configuring StarWind Management Console

1. Install StarWind Management Console on a workstation with Windows OS (Windows 7 or higher, Windows Server 2008 R2 and higher) using the installator available here.
NOTE: StarWind Management Console and PowerShell Management Library components are required.
2. Select the appropriate option to apply the StarWind License key.
Once the appropriate license key has been received, it should be applied to StarWind Virtual SAN service via Management Console or PowerShell.
3. Open StarWind Management Console and click Add Server.
Add Server4. Type the IP address of the StarWind Virtual SAN in the pop-up window and click OK.
Add new StarWind server5. Select the server and click Connect.
6. Click Apply Key… on the pop-up window.

Apply key7. Select Load license from file and click the Load button.
8. Select the appropriate license key.
As an alternative, PowerShell can be used. Open StarWind InstallLicense.ps1 script with PowerShell ISE as administrator. It can be found here:
C:\Program Files\StarWind Software\StarWind\StarWindX\Samples\powershell\InstallLicense.ps1
Type the IP address of StarWind Virtual SAN VM and credentials of StarWind Virtual SAN service (defaults login: root, password: starwind).
Add the path to the license key.
Configuring StarWind Management Console9. After the license key is applied, StarWind devices can be created.
NOTE: In order to manage StarWind Virtual SAN service (e.g. create ImageFile devices, VTL devices, etc.), StarWind Management Console can be used.

Configuring Storage

StarWind Virtual SAN for vSphere can work on top of Hardware RAID or Linux Software RAID (MDADM) inside of the Virtual Machine.
Please select the required option:

Configuring StarWind Storage on Top of Hardware RAID

1. Add a new virtual disk to the StarWind Virtual SAN VM by editing its settings. Make sure it is Thick Provisioned Eager Zeroed. Virtual Disk should be located on the datastore provided by hardware RAID.New_Hard_disk

NOTE: Alternatively, the disk can be added to StarWind VSAN VM as RDM. The link to VMware documentation is below:
https://docs.vmware.com/en/VMware-vSphere/7.0/com.vmware.vsphere.vm_admin.doc/GUID-4236E44E-E11F-4EDD-8CC0-12BA664BB811.html

NOTE: If a separate RAID controller is available, it can be used as dedicated storage for StarWind VM, and RAID controller can be added to StarWind VM as a PCI device. In this case RAID volume will be available as a virtual disk in the Drives section in the Web console. Follow the instructions in the section below on how to add RAID controller as PCI device to StarWind VM.

2. Login to StarWind VSAN VM web console and find in the Storage section under Drives the Virtual Disk that was recently added and choose it.

Storage
3. The added disk does not have any partitions and filesystem. Press on Format button to create the partition and format it.

Format

4. Create the XFS partition. Specify the name and erase option. The mount point should be as following: /mnt/%yourdiskname% . Click Format.

Format_Mount_Point

5. On the storage page of the disk, navigate to the Filesystem tab. Click Mount.

Disk_Mount

6. Using StarWind Management Console, connect to StarWind Virtual SAN VM and configure storage pool (default storage for StarWind devices) by clicking Yes.

StarWind Management Console 7. Select the disk which was recently mounted.

select the mounted disk

Configuring StarWind Storage on Top of Software RAID

Make sure that the prerequisites for deploying Software RAID with StarWind Virtual SAN are met:

  • The ESXi hosts have all the drives connected through HBA or RAID controller in HBA mode.
  • StarWind Virtual SAN for vSphere VM is deployed on the ESXi server and turned off.
  • StarWind Virtual SAN VM is installed on a separate storage device available to the ESXi host (e.g. SSD, HDD etc.).
  • HBA or RAID controller can be added via a DirectPath I/O passthrough device to a StarWind Virtual SAN VM without affecting ESXi host work.

PCI Device Configuration
1. Login to the ESXi host where StarWind Virtual SAN VM is installed.
2. In the Navigator, go to Manage, and in the Hardware tab, select PCI Devices.
3. Locate the HBA/RAID Controller of the ESXi host. Check the box on the appropriate PCI device. Click Toggle passthrough.
Enable_passthrough

4. Restart ESXi host to make PCI device available to VMs.
5. Right-click on the StarWind Virtual SAN VM to Edit Settings.
7. Click ADD NEW DEVICE. Select PCI Device.
Add_PCI_Device
8. Add HBA/RAID Controller to the VM. Reserve memory for the StarWind Virtual Machine. Click OK.
9. Boot StarWind Virtual SAN VM.
10. Repeat steps 1-8 for all hosts where StarWind Virtual SAN for vSphere is deployed.
11. Login to StarWind Virtual SAN VM via IP. The default credentials:
Login: user
Password: rds123RDS
NOTE:
Please make sure that the default password is changed.Login_screen

12. Go to the Storage page. The Drives section shows the drives connected to HBA/RAID Controller (if available).Storage
13. Click “+” in the RAID Devices section to create Software RAID. (In the current example, RAID 10 will be created with 4 HDD drives). The RAID configuration depends on the number of disks, chunk size, and array level are shown in the table below:

RAID Level Chunk size for HDD Arrays Chunk size for SSD Arrays

0

Disk quantity * 4Kb

Disk quantity * 8Kb

5

(Disk quantity – 1) * 4Kb

(Disk quantity – 1) * 8Kb

6

(Disk quantity – 2) * 4Kb

(Disk quantity – 2) * 8Kb

10 (Disk quantity * 4Kb)/2 (Disk quantity * 8Kb)/2

StarWind Software RAID recommended settings can be found here:
https://knowledgebase.starwindsoftware.com/guidance/recommended-raid-settings-for-hdd-and-ssd-disks/

14.
Select the drives to add to the array.
Select the drives to add to the array15. After the synchronization is finished, find the RAID array created.
NOTE: The disk created will not have any partitions and file system. Click Format.
RAID array created16. Create the XFS partition. Mount point should be as follows: /mnt/%yourdiskname% . Select the Custom mounting option and type noatime. Click Format.
Mount point17. On the storage page of the disk, navigate to the Filesystem tab. Click Mount.
Filesystem tab18. Connect to StarWind Virtual SAN from StarWind Management Console or from Web Console. Click Yes.
StarWind Management Console 19. Select the disk recently mounted.
mounted disk

Creating StarWind devices

1. In the StarWind Management Console click to Add Device (advanced) button and open Add Device (advanced) Wizard.

Creating HA StarWind
2. Select Hard Disk Device as the type of device to be created.

StarWind Hard Disk Device
3. Select Virtual Disk.

Add Device Wizard StarWind
4. Specify a virtual disk Name, Location, and Size.

Device Wizard
5. Select the Thick provisioned disk type and block size.
NOTE: Use 4096 sector size for targets, connected on Windows-based systems and 512 bytes sector size for targets, connected on Linux-based systems (ESXi/Xen/KVM).

6. Define a caching policy and specify a cache size (in MB). Also, the maximum available cache size can be specified by selecting the appropriate checkbox. Optionally, define the L2 caching policy and cache size.

Device Wizard
7. Specify Target Parameters. Select the Target Name checkbox to enter a custom target name. Otherwise, the name is generated automatically in accordance with the specified target alias.

Device Wizard StarWind
8. Click Create to add a new device and attach it to the target.

Device Wizard StarWind
9. Click Close to finish the device creation.

10. The successfully added devices appear in the StarWind Management Console.

 

Select the Required Replication Mode

The replication can be configured in one of two modes:

Synchronous “Two-Way” Replication
Synchronous or active-active replication ensures real-time synchronization and load balancing of data between two or three cluster nodes. Such a configuration tolerates the failure of two out of three storage nodes and enables the creation of an effective business continuity plan. With synchronous mirroring, each write operation requires control confirmation from both storage nodes. It guarantees the reliability of data transfers but is demanding in bandwidth since mirroring will not work on high-latency networks.

Asynchronous “One-Way” Replication
Asynchronous Replication is used to copy data over a WAN to a separate location from the main storage system. With asynchronous replication, confirmation from each storage node is not required during the data transfer. Asynchronous replication does not guarantee data integrity in case of storage or network failure; hence, some data loss may occur, which makes asynchronous replication a better fit for backup and disaster recovery purposes where some data loss is acceptable. The Replication process can be scheduled in order to prevent the main storage system and network channels overloads.
Please select the required option:

Synchronous “Two-Way” replication

1. Right-click the recently created device and select Replication Manager from the shortcut menu.

2. Select the Add Replica button in the top menu.

StarWind Replication Manager

3. Select Synchronous “Two-Way” replication as a replication mode.

4. Specify a partner Host name or IP address and Port Number.

Selecting the Failover Strategy

StarWind provides 2 options for configuring a failover strategy:

Heartbeat
The Heartbeat failover strategy allows avoiding the “split-brain” scenario when the HA cluster nodes are unable to synchronize but continue to accept write commands from the initiators independently. It can occur when all synchronization and heartbeat channels disconnect simultaneously, and the partner nodes do not respond to the node’s requests. As a result, StarWind service assumes the partner nodes to be offline and continues operations on a single-node mode using data written to it.
If at least one heartbeat link is online, StarWind services can communicate with each other via this link. The device with the lowest priority will be marked as not synchronized and get subsequently blocked for the further read and write operations until the synchronization channel resumption. At the same time, the partner device on the synchronized node flushes data from the cache to the disk to preserve data integrity in case the node goes down unexpectedly. It is recommended to assign more independent heartbeat channels during the replica creation to improve system stability and avoid the “split-brain” issue.
With the heartbeat failover strategy, the storage cluster will continue working with only one StarWind node available.

Node Majority
The Node Majority failover strategy ensures the synchronization connection without any additional heartbeat links. The failure-handling process occurs when the node has detected the absence of the connection with the partner.
The main requirement for keeping the node operational is an active connection with more than half of the HA device’s nodes. Calculation of the available partners is based on their “votes”.
In case of a two-node HA storage, all nodes will be disconnected if there is a problem on the node itself, or in communication between them. Therefore, the Node Majority failover strategy requires the addition of the third Witness node or file share (SMB) which participates in the nodes count for the majority, but neither contains data on it nor is involved in processing clients’ requests. In case an HA device is replicated between 3 nodes, no Witness node is required.
With Node Majority failover strategy, failure of only one node can be tolerated. If two nodes fail, the third node will also become unavailable to clients’ requests.
Please select the required option:

Heartbeat

1. Select Failover Strategy.
2. Select Create new Partner Device and click Next.
3. Select a partner device Location.
4. Click Change Network Settings.
5. Specify the interfaces for Synchronization and Heartbeat Channels. Click OK and then click Next.
6. In Select Partner Device Initialization Mode, select Synchronize from existing Device and click Next.
7. Click Create Replica. Click Finish to close the wizard.
The successfully added device appears in StarWind Management Console.
8. Follow the same procedure for the creation of other virtual disks that will be used as storage repositories.

Node Majority

There are two ways to configure Witness for 2-nodes StarWind HA device, created with Node Majority Failover Strategy: File Share (SMB) as Witness and additional server as Witness Node.

– Creating HA device with File SHare(SMB) as Witness:

SMB Witness is a file, located on SMB share, which can be accessed by both nodes and help them to eliminate the split-brain issue in case of synchronization connection interruption between the nodes. To set up the SMB file share as a Witness for 2-nodes HA device with Node Majority Failover Strategy, perform the actions, described on this page:
https://www.starwindsoftware.com/help/ConfiguringFileShareSMBasWitness.html

– Creating HA device with Witness Node:

1. Select the Node Majority failover strategy and click Next.
Node Majority2. Choose Create new Partner Device and click Next.
3. Specify the partner device Location and modify the target name if necessary. Click Next.
4. In Network Options for Replication, press the Change network settings button and select the synchronization channel for the HA device.
5. In Specify Interfaces for Synchronization Channels, select the checkboxes with the appropriate networks and click OK. Then click Next.
6. Select Synchronize from existing Device as the partner device initialization mode.
7. Press the Create Replica button and close the wizard.
8. The added devices will appear in StarWind Management Console.
Repeat the steps above to create other virtual disks if necessary.

Adding Witness Node
Witness node can be configured on a separate host or as a virtual machine in a cloud. It requires StarWind Virtual SAN service installed on it.
NOTE: Since the device created in this guide is replicated between 2 active nodes with the Node Majority failover strategy, a Witness node must be added to it.
1. Open StarWind Management Console, right-click on the Servers field and press the Add Server button. Add a new StarWind Server which will be used as the Witness node and click OK.
witness node2. Right-click on the HA device with the configured Node Majority failover policy and select Replication Manager and press the Add Replica button.
3. Select Witness Node.
witness node
4. Specify the Witness node Host Name or IP address. The default Port Number is 3261.
Ad partner node
5. In Partner Device Setup, specify the Witness device Location. Optionally, modify the target name by clicking the appropriate button.
6. In Network Options for Replication, select the synchronization channel with the Witness node by clicking the Change Network Settings button.
7. Specify the interface for Synchronization and Heartbeat and click OK.
8. Click Create Replica and then close the wizard.
9. Repeat the steps above to create other virtual disks if necessary.

NOTE: To extend an Image File or a StarWind HA device to the required size, please check the article below:

How to extend Image File or High Availability device

Preparing Datastores

Adding Discover Portals

1. To connect the previously created devices to the ESXi host, click on the Storage -> Adapters -> Software iSCSI and in the appeared window choose the Enabled option to enable Software iSCSI storage adapter. Push the Save configuration button.

Enable_Software_iSCSI

2. In the Configure iSCSI window, under Dynamic Targets, click on the Add dynamic target button to specify iSCSI interfaces.

Add_Dynamic_Target

3. Enter the iSCSI IP addresses of all StarWind nodes for the iSCSI traffic.

Add dynamic target ESXI

Add dynamic target ESXI

Confirm the actions by pressing Save configuration.

4. The result should look like in the image below.

Configure iSCSI

5. Click on the Rescan button to rescan storage.

ESXi_storage_devices

6. Now, the previously created StarWind devices are visible to the system.

7. Repeat all the steps from this section on the other ESXi host, specifying corresponding IP addresses for the iSCSI subnet.

Creating Datastores

1. Open the Storage tab on one of your hosts and click on New Datastore.
Select creation type

2. Specify the Datastore name, select the previously discovered StarWind device, and click Next.
Select device

3. Enter datastore size and click Next.
Enter datastore size

4. Verify the settings and click Finish.
5. Add another Datastore (DS2) in the same way but select the second device for the second datastore.
6. Verify that your storages (DS1, DS2) are connected to both hosts. Otherwise, rescan the storage adapter.
storages


NOTE: 
Path Selection Policy changing for Datastores from Most Recently Used (VMware) to Round Robin (VMware) is added into the Rescan Script, and this action is performed automatically. For checking and changing this parameter manually, the hosts should be connected to vCenter.
Multipathing configuration can be checked only from vCenter. To check it, click the Configure button, choose the Storage Devices tab, select the device, and click the Edit Multipathing button.

StarWind edit multipathStarWind Edit multipath

Configuring an Automatic Storage Rescan

1. Open the Terminal page.
2. Edit file /opt/StarWind/StarWindVSA/drive_c/StarWind/hba_rescan.ps1 with the following command:
sudo nano /opt/StarWind/StarWindVSA/drive_c/StarWind/hba_rescan.ps1
Modify_Rescan_Script
3. In the appropriate lines, specify the IP address and login credentials of the ESXi host (see NOTE below) on which the current StarWind VM is stored and running:
$ESXiHost = “IP address”
$ESXiUser = “Login”
$ESXiPassword = “Password”

NOTE:  In some cases the rescan script can be changed and storage rescan added for another ESXi host. Appropriate lines should be duplicated and changed with properly edited variables if required.

NOTE: In some cases, it makes sense to create a separate ESXi user for storage rescans. To create the user, please follow the steps below:

Log in to ESXi with the VMware Host Client. Click Manage, and under Security & users tab, in the Users section click Add user button. In the appeared window, enter a user name, and a password.
ESXi_Add_User
Create a new Role, under Roles section, and click New Role button.
Type a name for the new role. Select privileges for the role and click OK.
The following privileges might be assigned: Host – Inventory, Config, Local Cim, and Global – Settings.
ESXi_Add_Role

Assign permission to the storage rescan user for an ESXi host  – right-click Host in the VMware Host Client inventory and click Permissions. In the appeared window click Add user.
Click the arrow next to the Select a user text box and select the user that you want to assign a role to. Click the arrow next to the Select a role text box and select a role from the list.
(Optional) Select Propagate to all children or Add as group. Click Add user and click Close.

ESXi_Manage_Permissions_Add_User_for_Host

Make sure that rescan script is working and execute it from the VM: sudo /opt/StarWind/StarWindVSA/drive_c/StarWind/hba_rescan.ps1

4. Repeat all steps from this section on the other ESXi hosts.

Performance Tweaks

1. Click on the Configuration tab on all of the ESXi hosts and choose Advanced Settings.

Advanced Settings
2. Select Disk and change the Disk.DiskMaxIOSize parameter to 512.
Disk.DiskMaxIOSize

3. To optimize  performance change I/O scheduler options according to the article below:
https://knowledgebase.starwindsoftware.com/guidance/starwind-vsan-for-vsphere-changing-linux-i-o-scheduler-to-optimize-storage-performance/ 

NOTE: Changing Disk.DiskMaxIOSize to 512 might cause startup issues with Windows-based VMs, located on the datastore where specific ESX builds are installed. If the issue with VMs start appears, leave this parameter as default or update the ESXi host to the next available build.

NOTE: To provide high availability for clustered VMs, deploy vCenter and add ESXi hosts to the cluster.
Click on Cluster -> Configure -> Edit and check the turn on vSphere HA option if it’s licensed.
vSphere HA option

Save your time finding all the answers to your questions in one place!
Have a question? Doubt something? Or just want to know an independent opinion?
StarWind Forum exists for you to “trust and verify” any issue already discussed and solved