StarWind Virtual SAN (VSAN):
Configuration Guide for Microsoft Hyper-V Server [Hyper-V], VSAN Deployed as a Controller Virtual Machine (CVM) using Web UI

Applies To: Windows Server 2016, Windows Server 2019, Windows Server 2022

Introduction to StarWind Virtual SAN CVM

StarWind Virtual SAN® Controller Virtual Machine (CVM) comes as a prepackaged Linux Virtual Machine (VM) to be deployed on any industry-standard hypervisor. It creates a VM-centric and high-performing storage pool for a VM cluster.
This guide describes the deployment and configuration process of the StarWind Virtual SAN CVM.

StarWind VSAN System Requirements

Prior to installing StarWind Virtual SAN, please make sure that the system meets the requirements, which are available via the following link: https://www.starwindsoftware.com/system-requirements

Recommended RAID settings for HDD and SSD disks:
https://knowledgebase.starwindsoftware.com/guidance/recommended-raid-settings-for-hdd-and-ssd-disks/

Please read StarWind Virtual SAN Best Practices document for additional information: https://www.starwindsoftware.com/resource-library/starwind-virtual-san-best-practices

Pre-configuring the Windows Server Hosts

The diagram below illustrates the network and storage configuration of the solution:

2-node Hyper-V CVM
1. Make sure that a domain controller is configured and the servers are added to the domain.

NOTE: Please follow the recommendation in KB article on how to place a DC in case of StarWind Virtual SAN usage.
2. Deploy Windows Server on each server and install Failover Clustering and Multipath I/O features, as well as the Hyper-V role on both servers. This can be done through Server Manager (Add Roles and Features menu item).
3. Define at least 2x network interfaces on each node that will be used for the Synchronization and iSCSI/StarWind heartbeat traffic. Do not use ISCSI/Heartbeat and Synchronization channels over the same physical link. Synchronization and iSCSI/Heartbeat links can be connected either via redundant switches or directly between the nodes (see diagram above).
4. Separate external  Virtual Switches should be created for iSCSI and Synchronization traffic based on the selected before iSCSI and Synchronization interfaces. Using Hyper-V Manager open Virtual Switch Manager and create two external Virtual Switches: one for the iSCSI/StarWind Heartbeat channel (iSCSI) and another one for the Synchronization channel (Sync).

5. Configure and set the IP address on each virtual switch interface. In this document, the 172.16.10.x subnet is used for iSCSI/StarWind heartbeat traffic, while 172.16.20.x subnet is used for the Synchronization traffic.

NOTE: In case NIC supports SR-IOV, enable it for the best performance. An additional internal switch is required for iSCSI Connection. Contact support for additional details.
6. Set MTU size to 9000 on iSCSI and Sync interfaces using the following Powershell script.

It will apply MTU 9000 to all iSCSI and Sync interfaces if they have iSCSI or Sync as part of their name.

NOTE: MTU setting should be applied on the adapters only if there is no live production running through the NICs.
Enabling Multipath Support
7. Open the MPIO Properties manager: Start -> Windows Administrative Tools -> MPIO. Alternatively, run the following PowerShell command :

8. In the Discover Multi-Paths tab, select the Add support for iSCSI devices checkbox and click Add.
wp-image-13680
9. When prompted to restart the server, click Yes to proceed.
10. Repeat the same procedure on the other server.

Installing File Server Roles

Please follow the steps below if file shares configuration is required

Scale-Out File Server (SOFS) for application data

1. Open Server Manager: Start -> Server Manager
2. Select: Manage -> Add Roles and Features
3. Follow the installation wizard steps to install the roles selected in the screenshot below:

StarWind - Select Server Role

4. Restart the server after installation is completed and perform steps above on the each server.

File Server for general use with SMB share

1. Open Server Manager: Start -> Server Manager
2. Select: Manage -> Add Roles and Features
3. Follow the installation wizard steps to install the roles selected in the screenshot below:

StarWind - Select Server Role

4. Restart the server after installation is completed and perform steps above on each server.

File Server for general use with NFS share

1. Open Server Manager: Start -> Server Manager
2. Select: Manage -> Add Roles and Features
3. Follow the installation wizard steps to install the roles selected in the screenshot below:

StarWind Select Role NFS

4. Restart the server after installation is completed and perform steps above on each server.

Deploying StarWind Virtual SAN CVM

1. Download the zip archive that contains StarWind Virtual SAN CVM
https://www.starwindsoftware.com/vsan#download

2. Extract the virtual machine files.

3. Deploy the control virtual machine to the Microsoft Hyper-V Server using the “Import Virtual Machine” wizard in Hyper-V Manager.

4. On the second page of the wizard, point to the location of the VM template. Select the VM folder and click Next.

5. Click Next on the “Select Virtual Machine” step.

6. Select the “Copy the virtual machine” import type and click Next.

7. Specify new or existing folders to store virtual machine files, such as configuration, snapshots, smart paging, and virtual disk. Click Next.

8. In the second step of the wizard, the “VM import” wizard will validate the network.

The default naming for virtual switches:

  • the Management virtual switch is “Management vSwitch”,
  • the iSCSI virtual switch is “Data/iSCSI vSwitch”,
  • the Synchronization virtual switch is “Replication/Sync vSwitch “.

If existing virtual switches have different names, specify corresponding network connections. Click Next.

9. Review the import configuration and click Finish to complete the import.

10. Repeat the VM deployment on each partner server which is used for configuring 2-node or 3-node highly available storage according to your licensing.

Initial Configuration Wizard

1. Start StarWind Virtual SAN CVM.

2. Launch VM console to see the VM boot process and get the IPv4 address of the Management network interface.
Note: in case VM has no IPv4 address obtained from a DHCP server, use the Text-based User Interface (TUI) to set up a Management network.

3. Using the web browser, open a new tab and enter the VM IPv4 address to open StarWind vSAN Web Interface. Click “Advanced” and then “Continue to…

4. StarWind vSAN web UI welcomes you, and the “Initial Configuration” wizard will guide you through the deployment process.

5. In the following step, upload the license file.

6. Read and accept the End User License Agreement to proceed.

7. Review or edit the Network settings and click Next.
Note: Static network settings are recommended for the configuration.

8. Specify the hostname for the virtual machine and click Next.

9. Create an administrator account. Click Next.

10. Review your settings selection before setting up StarWind vSAN.

11. Please standby until the Initial Configuration Wizard configures StarWind vSAN for you.

12. The appliance is set and ready. Click on the Done button to install the StarWind vCenter Plugin right now or uncheck the checkbox to skip this step and proceed to the Login page.

13. Repeat the initial configuration on other StarWind CVMs that will be used to create 2-node or 3-node HA shared storage.

Add Appliance

To create 2-way or 3-way synchronously replicated highly available storage, add partner appliances that use the same license key.

1. Add StarWind appliance(s) in the web console, on the Appliances page

NOTE: The newly added appliance will be linked to already connected partners.

2. Provide credentials of partner appliance.

3. Wait for connection and validation of settings.

4. Review the summary and click “Add appliance”.

Configure HA networking

1. Launch the “Configure HA Networking” wizard.

2. Select appliances for network configuration.

NOTE: the number of appliances to select is limited by your license, so can be either two or three appliances at a time.

3. Configure the “Data” network. Select interfaces to carry storage traffic, configure them with static IP addresses in unique networks, and specify subnet masks.

  • Assign and configure at least one interface on each node
  • For redundant configuration, select two interfaces on each node
  • Ensure interfaces are connected to client hosts directly or through redundant switches

4. Assign MTU value to all selected network adapters, e.g. 1500 or 9000. Ensure the switches have the same MTU value set.

5. Click Next to validate Data network settings.

6. Configure the “Replication” network. Select interfaces to carry storage traffic, configure them with static IP addresses in unique networks, and specify subnet masks.

  • Assign and configure at least one interface on each node
  • For redundant configuration, select two interfaces on each node
  • Ensure interfaces are connected to client hosts directly or through redundant switches

7. Assign MTU value to all selected network adapters, e.g. 1500 or 9000. Ensure the switches have the same MTU value set.

8. Click Next to validate the Replication network settings completion.

9. Review the summary and click Configure.

Add physical disks

Attach storage to StarWind Virtual SAN Controller VM.

NOTE: In order to make RDM disks available for StarWind devices in StarWind CVM Version 20231016 (build 15260), please follow the steps below.

  • Stop service:

  • Get VMDK/RDM/ device letter using lsblk command:

  • Edit config file

  • Add lines to the file, previously setting the disk letters to config (e.g. sdb, sdc) :

  • Start service:

Create Storage Pool

1. Click the “Add” button to create a storage pool.

2. Select two storage nodes to create a storage pool on them simultaneously.

3. Select physical disks to include in the storage pool name and click the “Next” button.

NOTE: Select identical type and number of disks on each storage node to create identical storage pools.

4. Select one of the preconfigured storage profiles or create a redundancy layout for the new storage pool manually according to your redundancy, capacity, and performance requirements.

Hardware RAID, Linux Software RAID, and ZFS storage pools are supported and integrated into the StarWind CVM web interface. To make easier the storage pool configuration, the preconfigured storage profiles are provided to configure the recommended pool type and layout according to the direct-attached storage:

  • Hardware RAID – configures Hardware RAID’s virtual disk as a storage pool. It is available only if a hardware RAID controller is passed through to the CVM.
  • High performance – creates Linux Software RAID-10 to maximize storage performance while maintaining redundancy
  • High capacity – creates Linux Software RAID-5 to maximize storage capacity while maintaining
    redundancy
  • Better redundancy – creates ZFS Stripped RAID-Z2 (RAID 60)) to maximize redundancy while maintaining high storage capacity
  • Manual – allows users to configure any storage pool type and layout with attached storage.

5. Review “Summary” and click the “Create” button to create the pools on storage servers simultaneously.

Create Volume

1. To create volumes, click the “Add” button.

2. Select two identical storage pools to create a volume simultaneously.

3. Specify volume name and capacity.

4. Select the Standard volume type.

5. Review “Summary” and click the “Create” button to create the pool.

Create HA LUN

The LUN availability for StarWind LUN can be Standalone and High availability (2-way or 3-way replication) and is narrowed by your license.

1. To create a virtual disk, click the Add button.

2. Select the protocol.

3. Choose the “High availability” LUN availability type.

4. Select the appliances that will host the LUN. Partner appliances must have identical hardware configurations, including CPU, RAM, storage, and networking.

5. Select a volume to store the LUN data. Selected volumes must have identical storage configurations.

6. Select the “Heartbeat” failover strategy.

NOTE:  To use the Node witness or the File share witness failover strategies, the appliances should have these features licensed.

7. Specify the HA LUN settings, e.g. name, size, and block size. Click Next.

8. Review “Summary” and click the “Create” button to create the LUN.

Creating StarWind HA LUNs using PowerShell

1. Open PowerShell ISE as Administrator.
2.  Open StarWindX sample CreateHA_2.ps1 using PowerShell ISE. It can be found here:
C:\Program Files\StarWind Software\StarWind\StarWindX\Samples\

2.  Configure script parameters according to the following example:

Detailed explanation of script parameters:

-addr, -addr2 — partner nodes IP address.
Format: string. Default value: 192.168.0.1, 192.168.0.1
allowed values: localhost, IP-address
-port, -port2 — local and partner node port.
Format: string. Default value: 3261
-user, -user2 — local and partner node user name.
Format: string. Default value: root
-password, -password2 — local and partner node user password.
Format: string. Default value: starwind

#common
-initMethod
Format: string. Default value: Clear
-size – set size for HA-devcie (MB)
Format: integer. Default value: 12
-sectorSize – set sector size for HA-device
Format: integer. Default value: 512
allowed values: 512, 4096
-failover – set type failover strategy
Format: integer. Default value: 0 (Heartbeat)
allowed values: 0, 1 (Node Majority)
-bmpType – set bitmap type, is set for both partners at once
Format: integer. Default value: 1 (RAM)
allowed values: 1, 2 (DISK)
-bmpStrategy – set journal strategy, is set for both partners at once
Format: integer. Default value: 0
allowed values: 0, 1 – Best Performance (Failure), 2 – Fast Recovery (Continuous)

#primary node
-imagePath – set path to store the device file
Format: string. Default value: My computer\C\starwind”. For Linux the following format should be used: “VSA Storage\mnt\mount_point”
-imageName – set name device
Format: string. Default value: masterImg21
-createImage – set create image file
Format: boolean. Default value: true
-targetAlias – set alias for target
Format: string. Default value: targetha21
-poolName – set storage pool
Format: string. Default value: pool1
-aluaOptimized – set Alua Optimized
Format: boolean. Default value: true
-cacheMode – set type L1 cache (optional parameter)
Format: string. Default value: wb
allowed values: none, wb, wt
-cacheSize – set size for L1 cache in MB (optional parameter)
Format: integer. Default value: 128
allowed values: 1 and more
-syncInterface – set sync channel IP-address from partner node
Format: string. Default value: “#p2={0}:3260”
-hbInterface – set heartbeat channel IP-address from partner node
Format: string. Default value: “”
-createTarget – set creating target
Format: string. Default value: true
Even if you do not specify the parameter -createTarget, the target will be created automatically.
If the parameter is set as -createTarget $false, then an attempt will be made to create the device with existing targets, the names of which are specified in the -targetAlias (targets must already be created)
-bmpFolderPath – set path to save bitmap file
Format: string.

#secondary node
-imagePath2 – set path to store the device file
Format: string. Default value: “My computer\C\starwind”. For Linux the following format should be used: “VSA Storage\mnt\mount_point”
-imageName2 – set name device
Format: string. Default value: masterImg21
-createImage2 – set create image file
Format: boolean. Default value: true
-targetAlias2 – set alias for targetFormat: string.
Default value: targetha22
-poolName2 – set storage pool
Format: string. Default value: pool1
-aluaOptimized2 – set Alua Optimized
Format: boolean. Default value: true
-cacheMode2 – set type L1 cache (optional parameter)
Format: string. Default value: wb
allowed values: wb, wt
-cacheSize2 – set size for L1 cache in MB (optional parameter)
Format: integer. Default value: 128
allowed values: 1 and more
-syncInterface2 – set sync channel IP-address from partner node
Format: string. Default value: “#p1={0}:3260”
-hbInterface2 – set heartbeat channel IP-address from partner node
Format: string. Default value: “”
-createTarget2 – set creating target
Format: string. Default value: true
Even if you do not specify the parameter -createTarget, the target will be created automatically.If the parameter is set as -createTarget $false, then an attempt will be made to create the device with existing targets, the names of which are specified in the -targetAlias (targets must already be created)
-bmpFolderPath2 – set path to save bitmap file
Format: string.

Selecting the Failover Strategy

StarWind provides 2 options for configuring a failover strategy:

Heartbeat
The Heartbeat failover strategy allows avoiding the “split-brain” scenario when the HA cluster nodes are unable to synchronize but continue to accept write commands from the initiators independently. It can occur when all synchronization and heartbeat channels disconnect simultaneously, and the partner nodes do not respond to the node’s requests. As a result, StarWind service assumes the partner nodes to be offline and continues operations on a single-node mode using data written to it.
If at least one heartbeat link is online, StarWind services can communicate with each other via this link. The device with the lowest priority will be marked as not synchronized and get subsequently blocked for the further read and write operations until the synchronization channel resumption. At the same time, the partner device on the synchronized node flushes data from the cache to the disk to preserve data integrity in case the node goes down unexpectedly. It is recommended to assign more independent heartbeat channels during the replica creation to improve system stability and avoid the “split-brain” issue.
With the heartbeat failover strategy, the storage cluster will continue working with only one StarWind node available.

Node Majority
The Node Majority failover strategy ensures the synchronization connection without any additional heartbeat links. The failure-handling process occurs when the node has detected the absence of the connection with the partner.
The main requirement for keeping the node operational is an active connection with more than half of the HA device’s nodes. Calculation of the available partners is based on their “votes”.
In case of a two-node HA storage, all nodes will be disconnected if there is a problem on the node itself, or in communication between them. Therefore, the Node Majority failover strategy requires the addition of the third Witness node or file share (SMB) which participates in the nodes count for the majority, but neither contains data on it nor is involved in processing clients’ requests. In case an HA device is replicated between 3 nodes, no Witness node is required.
With Node Majority failover strategy, failure of only one node can be tolerated. If two nodes fail, the third node will also become unavailable to clients’ requests.
Please select the required option:

Provisioning StarWind HA Storage to Windows Server Hosts

1. Launch Microsoft iSCSI InitiatorStart -> Windows Administrative Tools -> iSCSI Initiator. Alternatively, launch it using the command below in the command line interface:

2. Navigate to the Discovery tab.
iscsi3. Click the Discover Portal button. The Discover Target Portal dialog appears. Type 172.16.10.10. Discover Portal
4. Click the Advanced button. Select Microsoft iSCSI Initiator as a Local adapter and select Initiator IP. Confirm the actions to complete the Target Portal discovery.
Advanced Settings5. Click the Discover Portal… button once again.
6. In Discover Target Portal dialog, type in the iSCSI interface IP address of the partner node that will be used to connect the StarWind provisioned targets. Click Advanced.
discover target portal partner7. Select Microsoft iSCSI Initiator as the Local adapter, select the Initiator IP in the same subnet as the IP address of the partner server from the previous step. Confirm the actions to complete the Target Portal discovery.
Advanced Settings

8. Now, all the target portals are added on the first node.iSCSI Initiator Properties
9. Repeat the steps 1-8 on the partner node.

Connecting Targets
1. Click the Targets tab. The previously created targets are listed in the Discovered Targets section.
NOTE: If the created targets are not listed, check the firewall settings of the StarWind Server as well as the list of networks served by the StarWind Server (go to StarWind Management Console -> Configuration -> Network). Alternatively, check the Access Rights tab on the corresponding StarWind VSAN server in StarWind Management Console for any restrictions.
initiator properties 32. Select the Witness target from the local server and click Connect.
3. Enable checkboxes as shown in the image below. Click Advanced.
Connect To Target4. Select Microsoft iSCSI Initiator in the Local adapter dropdown menu. In the Initiator IP field, select the IP address for the iSCSI channel. In the Target portal IP, select the corresponding portal IP from the same subnet. Confirm the actions. iSCSI device connection
5. Repeat the steps 2-4 to connect to partner node.
6. Select the CSV1 target discovered from the local server and click Connect.
7. Enable checkboxes as shown in the image below. Click Advanced.
connect to target multipath8. Select Microsoft iSCSI Initiator in the Local adapter dropdown menu. In Target portal IP, select 172.16.10.10. Confirm the actions.
9. Select the partner target from the other StarWind node and click Connect.
10. Repeat the step 6.
11. Select Microsoft iSCSI Initiator in the Local adapter dropdown menu. In the Initiator IP field, select the IP address for the iSCSI channel. In the Target portal IP, select the corresponding portal IP from the same subnet. Confirm the actions.iSCSI device connection 2
11. Repeat the steps 1-10 for all remaining HA device targets.
12. Repeat the steps 1-11 on the other StarWind node, specifying corresponding data channel IP addresses.

Configuring Multipath
NOTE: It is recommended to configure the different MPIO policies depending on iSCSI channel throughput. For 1 Gbps iSCSI channel throughput, it is recommended to set Failover Only or Least Queue Depth MPIO load balancing policy. For 10 Gbps iSCSI channel throughput, it is recommended to set Round Robin or Least Queue Depth MPIO load balancing policy.
1. Configure the MPIO policy for each target with the load balance policy of choice. Select the Target located on the local server and click Devices.
2. In the Devices dialog, click MPIO.
devices 3. Select the appropriate load balancing policy.
least queue depth4. Repeat the steps 1-3 for configuring the MPIO policy for each remaining device on the current node and on the partner node.

Connecting Disks to Servers
1. Open the Disk Management snap-in. The StarWind disks will appear as unallocated and offline.
disk management 2. Bring the disks online by right-clicking on them and selecting the Online menu option.
3. Select the CSV disk (check the disk size to be sure) and right-click on it to initialize.
4. By default, the system will offer to initialize all non-initialized disks. Use the Select Disks area to choose the disks. Select GPT (GUID Partition Style) for the partition style to be applied to the disks. Press OK to confirm.
initialize disk5. Right-click on the selected disk and choose New Simple Volume.
6. In New Simple Volume Wizard, indicate the volume size. Click Next.
7. Assign a drive letter to the disk. Click Next.
assign drive letter 8. Select NTFS in the File System dropdown menu. Keep Allocation unit size as Default. Set the Volume Label of choice. Click Next.
NTFS volume9. Press Finish to complete.
10. Complete the steps 1-9 for the Witness disk. Do not assign any drive letter or drive path for it.
do not assign drive letter11. On the partner node, open the Disk Management snap-in. All StarWind disks will appear offline. If the status is different from the one shown below, click Action->Refresh in the top menu to update the information about the disks.
12. Repeat step 2 to bring all the remaining StarWind disks online.

Creating a Failover Cluster in Windows Server

NOTE: To avoid issues during the cluster validation configuration, it is recommended to install the latest Microsoft updates on each node.
NOTE: Server Manager can be opened on the server with desktop experience enabled (necessary features should be installed). Alternatively, the Failover cluster can be managed with  Remote Server Administration Tools: https://docs.microsoft.com/en-us/windows-server/remote/remote-server-administration-tools
NOTE: For converged deployment (SAN & NAS running as a dedicated storage cluster) the Microsoft Failover Cluster is deployed on separate computing nodes. Additionally, for the converged deployment scenario, the storage nodes that host StarWind SAN & NAS as CVM or bare metal do not require a domain controller and Failover Cluster to operate.

1. Open Server Manager. Select the Failover Cluster Manager item from the Tools menu.

failover manager
2. Click the Create Cluster link in the Actions section of Failover Cluster Manager.
failover cluster ws 2016
3. Specify the servers to be added to the cluster. Click Next to continue.
select servers
4. Validate the configuration by running the cluster validation tests: select Yes… and click Next to continue.
validation warning
5. Specify the cluster name.
NOTE: If the cluster servers get IP addresses over DHCP, the cluster also gets its IP address over DHCP. If the IP addresses are set statically, set the cluster IP address manually.

cluster ip
6. Make sure that all settings are correct. Click Previous to make any changes or Next to proceed.
confirmation
NOTE: If checkbox Add all eligible storage to the cluster is selected, the wizard will add all disks to the cluster automatically. The device with the smallest storage volume will be assigned as a Witness. It is recommended to uncheck this option before clicking Next and add cluster disks and the Witness drive manually.
7. The process of the cluster creation starts. Upon the completion, the system displays the summary with the detailed information. Click Finish to close the wizard.
summary
Adding Storage to the Cluster
1. In Failover Cluster Manager, navigate to Cluster -> Storage -> Disks. Click Add Disk in the Actions panel, choose StarWind disks from the list and confirm the selection.
add disks
2. To configure the cluster witness disk, right-click on Cluster and proceed to More Actions -> Configure Cluster Quorum Settings.
configure cluster quorum
3. Follow the wizard and use the Select the quorum witness option. Click Next.
select quorum
4. Select Configure a disk witness. Click Next.
select disk witness
5. Select the Witness disk to be assigned as the cluster witness disk. Click Next and press Finish to complete the operation.
configure storage witness
6. In Failover Cluster Manager, Right-click the disk and select Add to Cluster Shared Volumes.
add csv
7. If renaming of the cluster shared volume is required, right-click on the disk and select Properties. Type the new name for the disk and click Apply followed by OK.
CSV properties
8. Perform the steps 6-7 for any other disk in Failover Cluster Manager. The resulting list of disks will look similar to the screenshot below.
Failover Cluster disks
Configuring Cluster Network Preferences
1. In the Networks section of the Failover Cluster Manager, right-click on the network from the list. Set its new name if required to identify the network by its subnet. Apply the change and press OK.

NOTE: Please double-check that cluster communication is configured with redundant networks: https://docs.microsoft.com/en-us/windows-server/failover-clustering/smb-multichannel


2. Rename other networks as described above, if required.

3. In the Actions tab, click Live Migration Settings. Uncheck the synchronization network, while the iSCSI network can be used if it is 10+ Gbps. Apply the changes and click OK.
live migration

The cluster configuration is completed and it is ready for virtual machines deployment. Select Roles and in the Action tab, click Virtual Machines -> New Virtual Machine. Complete the wizard.

 

Configuring File Shares

Please follow the steps below if file shares should be configured on cluster nodes.

Configuring the Scale-Out File Server Role

1. To configure the Scale-Out File Server Role, open Failover Cluster Manager
2. Right-click the cluster name, then click Configure Role and click Next to continue

StarWind Configuring Server Roles

3. Select the File Server item from the list in High Availability Wizard and click Next to continue

StarWind Select Role

4. Select Scale-Out File Server for application data and click Next

Select Server Type

5. On the Client Access Point page, in the Name text field, type the NetBIOS name that will be used to access a Scale-Out File Server

StarWind File Server Name

Click Next to continue.

6. Check whether the specified information is correct. Click Next to continue or Previous to change the settings

File Server Confirmation

7. Once the installation is finished successfully, the Wizard should now look like the screenshot below.

Click Finish to close the Wizard.

File Server Summary

8. The newly created role should now look like the screenshot below.

File Server Confirmation

NOTE: If the role status is Failed and it is unable to Start, please, follow the next steps:

FIle Server Error

  • Open Active Directory Users and Computers;
  • Enable the Advanced view if it is not enabled;
  • Edit the properties of the OU containing the cluster computer object (in this case – Production);
  • Open the Security tab and click Advanced;
  • In the appeared window, press Add (the Permission Entry dialog box opens), click Select a principal;
  • In the appeared window, click Object Types, select Computers, and click OK;
  • Enter the name of the cluster computer object (in this case – Production);

Select User. Computer

  • Go back to Permission Entry dialog, scroll down, and select Create Computer Objects

Permission Entry for Computers

  • Click OK on all opened windows to confirm the changes.
  • Open Failover Cluster Manager, right-click SOFS role and click Start Role

Configuring File Share

To Add File Share:

1. Open Failover Cluster Manager.
2. Expand the cluster and then click Roles.
3. Right-click the file server role and then press Add File Share.
4. On the Select the profile for this share page, click SMB ShareApplications and then click Next.

SMB Share

5. Select a CSV to host the share. Click Next to proceed

Selecting server for share

6. Type in the file share name and click Next

Spacify Share name

7. Make sure that the Enable Continuous Availability box is checked. Click Next to proceed

Configuring share settings

8. Specify the access permissions for the file share.

Permissions to Control Access

NOTE:

  • For the Scale-Out File Server for Hyper-V, all Hyper-V computer accounts, the SYSTEM account, and all Hyper-V administrators must be provided with the full control on the share and file system;
  • For the Scale-Out File Server on Microsoft SQL Server, the SQL Server service account must be granted full control on the share and the file system.

9. Check whether specified settings are correct. Click Previous to make any changes or click Create to proceed.

Confirm Selections SOFS

10. Check the summary and click Close to close the Wizard.

SOFS results

To Manage Created File Shares:

1. Open Failover Cluster Manager.
2. Expand the cluster and click Roles.
3. Choose the file share role, select the Shares tab, right-click the created file share, and select Properties:

Failover Cluster manager Shares

Configuring the File Server for General Use Role

NOTE: To configure File Server for General Use, the cluster should have available storage

1. To configure the Scale-Out File Server role, open Failover Cluster Manager
2. Right-click on the cluster name, then click Configure Role and click Next to continue

StarWind Configuring Server Roles

3. Select the File Server item from the list in High Availability Wizard and click Next to continue

StarWind Select Role

4. Select File Server for general use and click Next

File Server Type

 

5. On the Client Access Point page, in the Name text field, type the NETBIOS name that will be used to access the File Server and IP for it

File Server Client Access Point

Click Next to continue

6. Select the Cluster disk and click Next

Select Storage

7. Check whether the specified information is correct. Click Next to proceed or Previous to change the settings

File Server General Use Confirmation

8. Once the installation has been finished successfully, the Wizard should now look like the screenshot below

Click Finish to close the Wizard.

File Server Summary

9. The newly created role should now look like the screenshot below

File Server

NOTE: If the role status is Failed and it is unable to Start, please, follow the next steps:

  • Open Active Directory Users and Computers;
  • Enable the Advanced view if it is not enabled;
  • Edit the properties of the OU containing the cluster computer object (in this case – Production);
  • Open the Security tab and click Advanced;
  • In the appeared window, press Add (the Permission Entry dialog box opens), click Select a principal;
  • In the appeared window, click Object Types, select Computers, and click OK;
  • Enter the name of the cluster computer object (in this case – Production);

Select User. Computer

  • Go back to Permission Entry dialog, scroll down, and select Create Computer Objects

Permission Entry for Computers

  • Click OK on all opened windows to confirm the changes.
  • Open Failover Cluster Manager, right-click File Share role and click Start Role

Configuring SMB File Share

To Add SMB File Share:

1. Open Failover Cluster Manager
2. Expand the cluster and then click Roles
3. Right-click the File Server role and then press Add File Share
4. On the Select the profile for this share page, click SMB ShareQuick and then click Next

Profile for SMB Share

5. Select available storage to host the share. Click Next to continue

Server and path for share

6. Type in the file share name and click Next

Share Name

7. Make sure that the Enable Continuous Availability box is checked. Click Next to continue

Configure share settings

8.Specify the access permissions for the file share

Permissions

9. Check whether specified settings are correct. Click Previous to make any changes or Next/Create to continue

Confirmation

10. Check the summary and click Close

View results

To manage created SMB File Shares:

11. Open Failover Cluster Manager
12. Expand the cluster and click Roles
13. Choose the File Share role, select the Shares tab, right-click the created file share, and select Properties

Failover cluster settings

 

Configuring NFS file share

To Add NFS File Share:

1. Open Failover Cluster Manager.
2. Expand the cluster and then click Roles.
3. Right-click the File Server role and then press Add File Share.
4. On the Select the profile for this share page, click NFS Share – Quick and then click Next

Profile for this share

5. Select available storage to host the share. Click Next to continue

Server and path for share

6. Type in the file share name and click Next

Share Name

7. Specify the Authentication. Click Next and confirm the message in pop-up window to continue

Autentication method

8. Click Add and specify Share Permissions

Share Permissions

Add Permission

9. Specify the access permissions for the file share

Permissions

 

10. Check whether specified settings are correct. Click Previous to make any changes or click Create to continue

Confirm Selections

11. Check a summary and click Close to close the Wizard

NFS Share results

To manage created NFS File Shares:

1. Open Failover Cluster Manager.
2. Expand the cluster and click Roles.
3. Choose the File Share role, select the Shares tab, right-click the created file share, and select Properties.

Failover cluster settings

 
Save your time finding all the answers to your questions in one place!
Have a question? Doubt something? Or just want to know an independent opinion?
StarWind Forum exists for you to “trust and verify” any issue already discussed and solved
Power Your Business With StarWind!
Book Free Online Demo:

Thank you for your application!
We will reply to you shortly.