Microsoft Failover Cluster Manager (MSFCM) on Windows server 2012/2016
Since Microsoft Windows 2012, Microsoft have reconfigured the way in which you can manage High Availability with print services. They introduced the ability to use Hyper-V and failover clusteringClustering allows your organisation to ensure your services are not affected if anything should happen to your main server. PaperCut is a cluster compatible application and is supported under Windows (Microsoft Cluster Server / MSCS, Microsoft Failover Cluster Manager / MSFCM, Vetitas Cluster Server / VCS) and Linux (Novell Cluster Services / NCS, Linux-HA) at all levels of the application, including: clustering at the print spooler service layer by integrating with clustering services, failover based clustering at the Application Server layer using clustering services, and at the database layer by utilising cluster aware databases such as Microsoft SQL Server, PostgreSQL, or Oracle. to make your print serverA print server is a system responsible for hosting print queues and sharing printer resources to desktops. Users submit print jobs to a print server rather then directly to the printer itself. A print server can be a dedicated server but on many networks this server also performs other tasks, such as file serving a highly available virtual machine. This solution provides full server failover options and can be implemented with the PaperCut NG Application ServerAn Application Server is the primary server program responsible for providing the PaperCut user interface, storing data, and providing services to users. PaperCut uses the Application Server to manage user and account information, manage printers, calculate print costs, provide a web browser interface to administrators and end users, and much more., Secondary Server, and/or Site ServerSite Servers take over the role of a Primary Application Server in the event of network outages. Key roles taken over include authentication, copy and print tracking and Find-Me printing. Site Servers ensure continuous availability of printing resources to support key business functions over unreliable network links or during unplanned network disruptions..
There are several steps involved in the setup of this failover solution:
- Before you Begin
- Step 1: Configure roles / features on nodes for high availability
- Step 2: Connecting to iSCI network drives
- Step 3: Create the failover cluster
- Step 4: Create a high availability virtual machine
- Step 5: Set up your print server
- 2+ physical servers
- MS Server 2012/2016 with Hyper-V capabilities
- iSCI SAN with 2 drives configured
- Drive 1: 5GB (to be used for the Quorum if using only 2 nodes)
- Drive 2: VM server storage
You will also require the following:
- Administrative rights to be able to join machines to the domain
- IP address to connect to your SAN
- IP Addresses for the following:
- 2 x IP addresses for the Physical Servers
- 1 x IP address for the Cluster
- 1 x IP address for the VM
- Install MS Windows Server 2012 or MS Server 2016 onto your two server nodes.
In the Server Manager > Dashboard, click Add roles and features.
The Before you begin window is displayed.
The Select installation type window is displayed.
- Select Role-based or feature-based installation.
The Select destination server window is displayed.
- Select the server that you are logged in on.
The Select server roles window is displayed.
In the Roles list, select Hyper-V.
The Add feature that are required for Hyper-V window is displayed. This window shows the dependencies that will be installed.
- Click Add Features.
The Select features window is displayed.
In the Features list, select Failover Clustering.
The Add features that are required for Failover Clustering window will pop up and now be displayed. This window shows the dependencies that will be installed with this feature.
Click Add Features.
The Hyper-V window is displayed.
The Create Virtual Switches window is displayed.
- In Network adapters, select the network that you want your virtual machine to use for the cluster. If you have multiple NIC’s listed here, choose the appropriate one/s.
The Virtual Machine Migration window is displayed.
- Select the Allow this server to send and receive live migrations of virtual machines checkbox. This allows the VM to transfer between your nodes where required.
The Default Stores window is displayed. Do not change the default stores.
The Confirm installation selections window is displayed.
- Check the setting to ensure you are happy with the selected items to be installed.
The chosen Role and Feature will be installed on your machine.
- When the installation is complete, restart the machine to finalize the installation.
- Repeat the above steps on any additional nodes you want to include in the failover cluster.
- Start the iSCI initiator then connect to your two SAN drives:
Click Start > Windows Administrative Tools > iSCI Initiator.Note:
One of the drives is configured for your Quorum and the other is configured for the virtual machines. 5GB is required for the Quorum and the second drive will be used to store your Virtual machines. Therefore, ensure you have provided enough disk space on this drive to handle the virtual machine storage.
- Ensure the drives are set to allow simultaneous connections. This is configured on your SAN. Make sure you have granted access to your two cluster servers.
- Connect to the SAN drives:
- Open iSCSI initiator.
Click the Targets tab.
- In Target, enter the IP address for the SAN.
Click Quick Connect.
Your advertised drives are displayed.
- Highlight a drive; then click connect to connect to each drive.
- When a drive is connected, click Done.
- Click the Volumes and Devices tab.
- Click Auto Configure.
- Click OK.
When you are connected on the first machine, click Start > WIndows Administrative tools > Computer Management > Disk Management.
Your two disks appear as Unknown and Offline. They also display the size you configured on the SAN (eg: 5GB and 150GB).
- Bring the disks online:
- Right-click the disk number.
- Select Online.
- Repeat the above step for the second drive.
- Initialize the disks:
- Right-click the disk number next to one of the new drives.
- select Initialize disk.
- In the box that appears, ensure that both your new disks have a tick mark next to them.
- Set as MBR.
- Click OK.
- Configure the new drive:
Right-click the first drive.
- Select Create New Simple Volume.
- Leave the defaults in place.
- Choose a drive letter to assign.
- Label your drives:
- 5GB drive—label the drive as Quorum
- 150GB (larger drive)—label it as ClusterStorage
Repeat the above steps for the second drive.
Once this is done for your first node (Server), repeat these steps on your additional nodes (Servers). For example, from the two listed servers above (Server1 and Server2), you must connect the iSCSI drives on both systems before they will be available for your cluster.
As the simple volume is now configured, you only need to initialise the disks on the other Nodes, not recreate the Volume.
From the OS of any of the nodes:
Click Start > Windows Administrative tools > Failover Cluster Manager to launch the Failover Cluster Manager.
Click Create Cluster.
The Before you Begin window is displayed.
The Select Servers window is displayed.
- Enter the server names that you want to add to the cluster. Alternatively, you can locate them via Browse.
- Click Add.
The Validation Warning window is displayed.
- Select Yes to allow verification of the cluster services.
The Validate a Configuration Wizard is displayed. This wizard validates the server configuration.
The Testing Options window is displayed.
- Select Run all tests (recommended).
The Confirmation window is displayed. This window lists all of the tests that will be run.
The Validating window is displayed while all of the clustering tests are being run. This process may take several minutes depending on your network infrastructure and the number of nodes you have chosen to add to your cluster.
When the tests have completed, check the report then fix any configuration errors. The cluster setup will fail if any errors exist.Note:
A common error is the NIC/s used on the two nodes. The NIC/s must have the same name on both servers for the cluster to use the NIC in the configuration.
The Access Point for Administering the Cluster window in the Create Cluster wizard is displayed.
- In Cluster Name, enter a name for your cluster.
In the available network provide an IP address for the cluster.Note:
This name and IP address will be registered in your DNS
The Confirmation window is displayed. This window lists the settings to be applied to your new cluster.
- Select the Add all eligible storage to the cluster check box.
The system will now try to assign any storage it can find.
The system attempts to create the new cluster in your domain. This may take a while as there are several checks that must take place and tests that are conducted while the system is configured.
When the process is complete, the Summary window is displayed stating that the cluster wizard completed successfully.
- Click Finish.
- Check to confirm that the cluster is configured correctly:
In the Failover Cluster Manager, navigate to Nodes.
- Check that all nodes in the cluster are online. If they are not, go to the server that is offline and bring the system online to join the cluster.
Navigate to Storage > Disks.
The system detects the SCSI drives and displays them here. If you were setting this up with only two nodes, then the 5GB Quorum cluster would have been assigned as Disk Witness in Quorum.Note:
If no disks are displayed, check that you have correctly completed the steps on all nodes detailed in Connecting to ISCSI network drives.
The configured storage space is assigned to Available Storage.
- Configure this storage to be part of the cluster:
Right-click the disk assigned to available storage; then select Add to Cluster Shared Volumes.
- Check the Cluster Events folder for any reported issues with the cluster. If there are no issues, you can configure your virtual machine in the cluster environment.
The cluster is now assigned to Cluster Shared Volume.
- In the Failover Cluster Manager, expand the cluster created in the previous steps.
Right-click Roles; then select Virtual Machines > New Virtual Machine.
The New Virtual Machine window is displayed.
- Select the node you want to set the virtual machine up on. It doesn’t matter which node you choose, but we recommend you choose the node that you are currently working on.
The New Virtual Machine Wizard is displayed. This window explains the steps involved in setting up a virtual machine.
The Specify Name and Location window is displayed.
- In Name, enter a name for your virtual machine.
- In Location, enter the drive where the VM will be stored on the server.
For this VM to be able to move between the nodes, choose the ClusterStorage\Volume drive. This is usually located on C:\ClusterStorage\Volume1\ or similar.
The Specify Generation window is displayed.
- Select the generation of virtual machine you want to use:
- Generation 1—offers the best cross compatibility with versions. Generation 2 offers greater security, better performance, and supports the UEFI disk partitioning.
- Generation 2—supported on the current releases, but not the older versions of Hyper-V. Also Generation 2 virtual machines are not supported by Azure.
Once this selection is made, it most likely will not be able to be changed. There are some tools that allow you to switch generation (for example, Azure DR allows you to to convert to Gen1 or Gen2 based on failover location), but it is best to assume that it cannot be changed.
The Assign Memory window is displayed.
The Configure Networking window is displayed.
The Connect Virtual Hard Disk window is displayed.
The Installation Options window is displayed.
The Completing the New Virtual Machine Wizard window is displayed. This window displays the options that you have selected for the configuration of this Hyper-V machine.
Review your selections and if you are happy with them, click Finish.
When the Hyper-V machine has been configured, the Summary window displays a Success message.
- With the virtual machine turned off, in Hyper-V Manager, highlight the machine.
- From the Action pane select Settings > Processor.
Expand Processor; then click Compatibility.
- Select the Migrate to a physical computer with a different processor check box.
- Click OK.
Ensure that the new nodes are configured to accept the virtual machine in the event of a failover:
In the Failover Cluster Manager, select Roles.
Double-click your virtual machine.
On the General tab, under Preferred owners, select the nodes that you want to manage your virtual machine in the event of a failure.
- Click OK.
- From the Cluster Manager window start your VM and install Windows.
- Configure the machine on your network.
- For testing purposes, turn off Windows Firewall so you can RDP and Ping this machine.
- Configure your print queues.
- Test that printing works successfully:
- Check printing on the server.
- Map the print queueA print queue displays information about documents that are waiting to be printed, such as the printing status, document owner, and number of pages to print. You can use the print queue to view, pause, resume, restart, and cancel print jobs. from a client machine on the network and test printing.
- When printing is working successfully, test the failover.
- In Failover Cluster Manager, right-click the VM you have created.
- Click Move.
- Select Quick Migration > Select Node.
Select a node that the VM is not currently residing on.
The VM will now start up on the other node.
When the virtual machine status shows that it is running, remote desktop into this machine to ensure it is still operational. Any windows you had open previously should still be running.
- Install PaperCut NG onto this machine.
- Send several test print jobs to ensure that PaperCut NG is operating correctly.
- Install your shared print queues to a client machine.
- Print a document.
- Log in to the server node that the PaperCut NG VM is running on; then restart the server.
- While the server is restarting, from the client machine, send several test print documents to the shared printer.
The print jobs are processed and the VM fails over successfully to one of your other nodes.
Don’t forget to reapply your firewall services on your machine before deploying into production.