Quantcast
Channel: High Availability (Clustering) forum
Viewing all 6672 articles
Browse latest View live

Enable-ClusterS2D reports "No disks with supported bus types found to be used for S2D"

$
0
0

I'm working on a stretch cluster configuration as described athttps://technet.microsoft.com/en-us/windows-server-docs/storage/storage-replica/stretch-cluster-replication-using-shared-storage .

I got as far as a successful run ofTest-SRTopology on my 2 servers (1 node per site, more to come) with local SAN LUNs.

I have the 2 nodes joined to a cluster now and tried Enable-ClusterS2D (as seen inhttps://www.youtube.com/watch?v=k1ElKuyZPJg - but not on the TechNet page) and got the error mentioned in title.

Get-PhysicalDisk shows my LUNs with BusType=Fibre Channel and I still get this "no disks with supported bus types found" while the TechNet page clearly lists FC/SAN disks as a supported configuration.

I know by using SAN storage I'm giving up S2D benefits (especially on the cost side), but I'd love to use the replication possibilities for our DR solution.

I tried to skip Enable-ClusterS2D and continue anyway (as TechNet does not mention that step). But then "Get-ClusterAvailableDisk -All | Add-ClusterDisk" fails (for the same reason?). Of the 4 LUNs I have, only 2 get added by Add-ClusterDisk. The other 2 LUNs are no longer visible as "clusteravailabledisk". but why??

any tips or tricks how I can get this to work? what am I doing wrong? or is my FC/SAN configuration to blame?

thanks a lot!

SMB Access denied for Cluster Role Resource

$
0
0

Dear All,

   I have Window 2008 R2 File Server Fail over cluster which is having in Production. As part of DR fail-over test i have created another stand alone Windows 2008 R2 Server with File Server role enabled. 

currently File Server disk (Disk) replication to DR with 3rd party product and during fail-over productioncluster role offline and attaching production disk to DR stand alone machine

Once disk attached to the DR host then changing the "Cluster Role - DNS "A" record IP Address pointing to DR Server .

when the users are trying to access the user Home folder or shared folder user getting access denied error. tried the \\DNS and FQDN (the access denied error. )

when i login to any workstation or Server with local administrator try to access same SMB using \\DNS and FQDN name it's working fine.

Any idea? 

Unable to connect to cluster

$
0
0

Hi Team,

I am not able to connect to cluster in windows 2012 server getting error message as below.




Sivakumar Thayumanavan

Unable to set NodeAndFileShareMajority quorum setting

$
0
0

Set-ClusterQuorum : There was an error configuring the file share witness '\\server\SharedFolder

Unable to save property changes for 'File Share Witness'.
    The user name or password is incorrect
At line:1 char:1
+ Set-ClusterQuorum -NodeAndFileShareMajority "\\server\SharedFolder'
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
    + CategoryInfo          : NotSpecified: (:) [Set-ClusterQuorum], ClusterCmdletException
    + FullyQualifiedErrorId : Set-ClusterQuorum,Microsoft.FailoverClusters.PowerShell.SetClusterQuorumCommand

On this SharedFolder, I have given full permission to ClusterName and cluster nodes.

Let me know if you have any query.

Thanks for your help in advance!!!

Node Disk Management

$
0
0

We have a 3TB iSCSI Lun assigned to our Exchange server in our cluster and in Node Disk Management it shows as 2 drives. A 2048GB CSVFS Primary Partition and 1024GB as unallocated. Of which we cannot make any changes to the unallocated partition.

But on the Exchange server that is running as a VM on this LUN. We have 3 partitions. What I am concerned about is what will happen when we exceed 2TB on one of the partitions. Is the Node Disk Management display of 2 partition's something I can ignore. Or should I look at moving my VM off of this LUN an creating a new one and verifying that I shows all 3TB before. Moving it back?

 

unable to delete clustered virtual machine

$
0
0
I have a 7 node hyperv 2012 clustered environment. Today, we experienced an issue with connectivity back to the shared CSV environment. To make a long story short, I now have 4 Virtual machines that are showing in the failover cluster manager but not showing in the hyperv manager of the specific node it says it is on. I was able to successfully recreate the VM's but now I am trying to remove it as a clustered Vm and I am not having much luck. I have tried using the stop-clustergroup etc, I have tried to remove it from the failover cluster manager but all attempts to remove  these have failed.  Any help on how to forcibly remove these would be greatly appreciated.

Server Manager unable to show CSVs

$
0
0

server 2012r2
Failover Cluster
HP SAN
2x Nodes

node b suddenly started getting source 'disk' id '153' logged in eventvwr

node a is functioning ok and all VMs are migrated to it.

disk manager reported one of the CSVs as offline - unable to bring it online - no messages - reboot took 2 hrs

subsequent efforts to view disk manager or volumes in server manager causes a reboot

searches through google suggest DISKPART> SAN=OnlineAll but I cannot get DISKPART to do anything any then the server reboots. I have started to roll back last windows updates, removing software that has storage type drivers like arcserve, Sophos exlusions are ok.

node b cannot complete Cluster Validation in fact the last time I ran it the server unexpectedly rebooted....

any clues on how to resolve?

Performance AlwaysOn SQL 2014 in synchronous mode

$
0
0

Hello,

I have found a great article about AlwaysOn performance in synchronous mode:

_https://blogs.msdn.microsoft.com/sambetts/2014/08/21/alwayson-asynchronous-vs-synchronous-commit-performance-for-sharepoint/

The short version of this artice is: updates are about x2 slower than standalone for a x2 node AlwaysOn cluster; reading data is about the same performance (which would make sense).

As we currently create an AlwaysOn Cluster SQL 2014 with one primary synchronous replica if have also done some performance tests. For this I have used a scripts which inserts 20000 rows. Please find the script at the end.

My results: (Script see below)
Single Database (standalone, not in Availabilty Group): 9seconds
Synchronized Database (AlwaysOn, synchronous mode): 19seconds

It would be very interesting for me if you get similar results and what you think about this: Is this high latency really a normal behavior, means by design?

CREATE TABLE dbo.TestTableSize
(
 MyKeyField VARCHAR(10) NOT NULL,
 MyDate1 DATETIME NOT NULL,
 MyDate2 DATETIME NOT NULL,
 MyDate3 DATETIME NOT NULL,
 MyDate4 DATETIME NOT NULL,
 MyDate5 DATETIME NOT NULL
)

DECLARE @RowCount INT
DECLARE @RowString VARCHAR(10)
DECLARE @Random INT
DECLARE @Upper INT
DECLARE @Lower INT
DECLARE @InsertDate DATETIME

SET @Lower = -730
SET @Upper = -1
SET @RowCount = 0

WHILE @RowCount < 20000
BEGIN
 SET @RowString = CAST(@RowCount AS VARCHAR(10))
 SELECT @Random = ROUND(((@Upper - @Lower -1) * RAND() + @Lower), 0)
 SET @InsertDate = DATEADD(dd, @Random, GETDATE())
 
 INSERT INTO TestTableSize
  (MyKeyField
  ,MyDate1
  ,MyDate2
  ,MyDate3
  ,MyDate4
  ,MyDate5)
 VALUES
  (REPLICATE('0', 10 - DATALENGTH(@RowString)) + @RowString
  , @InsertDate
  ,DATEADD(dd, 1, @InsertDate)
  ,DATEADD(dd, 2, @InsertDate)
  ,DATEADD(dd, 3, @InsertDate)
  ,DATEADD(dd, 4, @InsertDate))

 SET @RowCount = @RowCount + 1
END



Adding new node to cluster can be done from any node?

$
0
0

Hi Team,

I have 2 node windows failover cluster (running SQL Failover instances). I wanted to add another node (3rd node) from DR site to existing Windows Cluster. If i try from Node 1 and Node 2 in primary datacenter, while adding node 3, 'Add node Wizard' error with 'Failed to access remote registry on <Node 3 name> . Ensure that the remote registry service is running and have remote administrator enabled'.

If i open failover cluster Manager on node 3 in DR site, connects to the existing cluster, try to add node 3 , it connects and start with the next step of validation.

I am wondering if these is any mandatory rule to add any nodes only through primary node i.e. first node from where Node 2 was added or new nodes can be added from any node's Faiolver Cluster Manager participating in cluster ?

Regards,

Cluster Service terminated by GUM Task

$
0
0

I've had an issue where one of my Windows 2012 R2 Hyper-V hosts just decided to keel over and die on me.  The event which I'm seeing is as follows:

Log Name:      System
Source:        Microsoft-Windows-FailoverClustering
Date:          20.07.2016 20:39:19
Event ID:      5377
Task Category: Global Update Mgr
Level:         Error
Keywords:
User:          SYSTEM
Computer:      mgmt45.mgmt.local
Description:
An internal Cluster service operation exceeded the defined threshold of '110' seconds. The Cluster service has been terminated to recover. Service Control Manager will restart the Cluster service and the node will rejoin the cluster.
Event Xml:<Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event"><System><Provider Name="Microsoft-Windows-FailoverClustering" Guid="{BAF908EA-3421-4CA9-9B84-6689B8C6F85F}" /><EventID>5377</EventID><Version>0</Version><Level>2</Level><Task>6</Task><Opcode>0</Opcode><Keywords>0x8000000000000000</Keywords><TimeCreated SystemTime="2016-07-20T18:39:19.244464800Z" /><EventRecordID>347017</EventRecordID><Correlation /><Execution ProcessID="4596" ThreadID="9184" /><Channel>System</Channel><Computer>mgmt45.mgmt.local</Computer><Security UserID="S-1-5-18" /></System><EventData><Data Name="OperationName">SynchronizeState</Data><Data Name="ThresholdTimeInSec">110</Data></EventData></Event>
I'm finding extremely little information regarding event 5377 on the Internet.  Apart from doing the standard checking for latest windows updates, and rebooting - how can I prevent this from happening again in the future?  This crash took down 64 virtual machines.

Replica Broker Failed state

$
0
0

hi guys,

First off, some info:

I'm helping set-up a DR site for a client and they'll be using Hyper-V Replica. The DR servers are up and running in a cluster and all roles have been configured including the Hyper-V Replica Broker (on the DR servers).

However, I ran into a problem bringing the Broker role on the existing production server online (it's on a failed state).

Now I've already searched the web a bit and found solutions similar to my problem:

https://blogs.technet.microsoft.com/askpfeplat/2012/12/09/why-adding-hyper-v-replica-connection-broker-fails-in-failover-cluster-manager/

Unfortunately, the methods described in the article and other webpages I've found (which offers pretty much the same solution) still do not work.

So my next step would be to try and do the steps described in this post:

https://www.experts-exchange.com/questions/28081209/Windows-2012-Hyper-V-Replica-Broker-Fails-to-Start.html

The post describes repairing the cluster but I'm not sure if there is any impact to the production environment as, I believe, the cluster name resource needs to be in a failed state or offline to attempt a repair.

So tl;dr, if I bring the cluster name resource offline, will it affect the virtual machines running on the cluster? Are there any other effects or steps I need to do to make sure that I can successfully bring the resource online again?

Heartbeat requirements from Node that is remote

$
0
0

Hi Team ,

We have existing 2 Node cluster, we are planning to add node 3 to the existing Windows Failover cluster in Primary Datacenter. Since 3rd node is actually from different location (and has subnet) in DR, i am wondering how heartbeat IPs need to be configured between 3node and other nodes? As of now both cluster nodes are from same location and has private virtual switch configured between them for heartbeat. Any guidance if we need to create some kind of VLAN for heartbeat ? Pls. node this 3rd node will not participate as owner and will not have vote in the cluster. Thank

Regards

Adding a new Cluster Node with different memory speed?

$
0
0

I have a requirement to add additional nodes to an existing Hyper-V Failover Cluster.

Dell have quoted for the new hardware and everything is identical to the existing kit, except for the memory speed, which they are unable to match.

I've spent some time looking into this, but I can't find any clear statements from Microsoft or other third parties.

Could anyone tell me if this would cause the Cluster Validation Wizard to fail and/or any other problems with the operation of the cluster? 


 

Exchange 2013 (On 2008 Server R2) - Cluster issues when quorum lost

$
0
0

9 Exchange 2013 mailbox servers all running on Windows 2008 Server R2. All part of the same DAG (so just one cluster).

Inadvertently, powerer off 5 of these mailbox servers - bad idea. According to my reckoning, due to odd number of members, the quorum should be relying on a majority count of members. Since it no lonfer has a majority then the databases should be dismounted on the remaining servers.

This didn't happen. The exchange databases stayed up - albeit the CAS servers went to 100%.

We then powered back up the 5 servers but the cluster did not recover on its own. On each of the 4 servers that did not power down the cluster status still showed down. A restart of the windows cluster service on each of the 4 servers sorted this out.

Question is :-

1. Any idea why exchange did not dismount DBs when quorum was lost?

2. Why did cluster not recover when the servers were back?

Storage Spaces Direct Windows Server 2016 Lab Testing physical disks issue

$
0
0

alright folks , first of all thanks for the help if i get a answer.

we are always on the look to test out some new stuff and this time we want to test Storage Spaces Direct,

in case it proves to be good , we can implement this in production.

for testing purpose i use 2 HP DL380 Gen 9 servers , i had some problems getting the disks to be recognized as SAS in Windows Server , but after a bit of google time i found out that after deleting all of the RAID config's on the HP P440ar and putting it into HBA Mode , it is possible.

then , after upgrading the server with the latest SPP i am now able to see the drives come up as SAS, so finally i thought i was able to start using S2D , but no , thats not the case.

at this point the 2 servers are identical , and they both have 2 SAS HDD 10K 146GB drives and 4 SAS SSD 460GB drives in there.

i installed windows server 2016 in UEFI mode on one of the 460GB drives , you could ask me why but i had some issue's installing and i just wanted a working system to test out S2D , i didn't cared about the storage at that point because after the Lab i'm erasing the configuration.

at this point , all i am trying to do is execute the command Enable-ClusterStorageSpacesDirect but i get a error straight away , saying S2D is not supported on my system , i verified the drives are shown as SAS in server manager , but when i run the cluster validation test , and specifically the storage spaces direct tests , i get the following error on one of the drives :

Disk is a boot volume. Disk is a system volume. Disk is used for paging files. Disk partition style is GPT. Disk has a System Partition. Cannot cluster a disk with a System Partition. Disk has a Microsoft Reserved Partition. Disk has a Basic Data Partition. Cannot cluster a disk with a Basic Data Partition. Disk has a Microsoft Recovery Partition. Disk type is DYNAMIC.

this error makes sense , since i had to install my OS on once of the drives , but what does that mean at point of configuration , did i made a mistake or can i exclude one of the disks ... i could use some help understanding this a bit better.


Microsoft Cluster failover policy recommendation - Maximum failures in specific period. once in 6 hours

$
0
0

Can someone give a clear explanation as to why in 2008 R2 server the recommended setting for "Maximum failures in specific period" is 1 in 6 hours.

We encountered a problem in a two node cluster.

A resource was in failed state and didn't attempt to restart the resource.

If the failures are configured as 1 in 6 hours then does the cluster attempt to start the resource in the same node or it will failover to other node and try to start.

Immediate response would be appreciated.

access is denied

$
0
0

Hello,

I have a win2012 r2 machine.

When I try to create a Failover cluster , i see 'access is denied'

the validate configuration wizard passed .

I am admin via a JIT Elevated access.

Do i require domain admin access ?

Thanks,Peter


The service principal name (SPN) was not found in Active Directory. Help me FIX Event 1207

$
0
0

I am having trouble with getting my 2012 Hyper-V HA Cluster at 100%. I have 3 nodes and it complains about the MS KB patches not matching across the 3, somethings that get installed have no "uninstall" - I can accept that.

It complains about 2 VMs that the tools don't match, well 1 is Ubuntu Server, 1 is a sandbox - I can accept that, but what keeps me at odds is how to FIX Event 1207, I am exhausted, looking at all the pages, fixes, help and I have tried everything I can think of, followed steps to no avail. I am missing something so simple - uggg. I have logged into AD (set it to advance) add permissions to the OU, CNO Etc... I understand you have to create it with an account that has proper creds, the CNO was created with my Creds (Domain Admin) 2 of the 3 nodes never complain, only the 3rd node, so I am thinking it is a AD Vs. this server node.

I am at a loss, any help would be appreciated.
Results by category (cluster manager validation report)

1 warning - Cluster Config - > Hyper V integration service version (Tools)

1 warning - Hyper-V Config -> Validate Software Update levels (Patches)

2 warning - cluster Config - > Validate Service Principal Name & Validate Resource Status (Sandbox is offline)

I am afraid I do not know enough about the SPN and how to fix it although I have tried everything and read articles, checked comm ports, checked creds to read status, create computer, update passwords etc..

Thank You


Enable-ClusterS2D hangs at "Waiting until all physical disks are reported by clustered storage subsystem"

$
0
0

The target cluster are running the final RTM MSDN bits of Nano server datacenter, with latest *.321.* windows update packages, and the member nodes have a mixture of NVMe and SATA storage. The -verbose output of enable-clusters2d shows the text at the bottom of this posting, up to the hang...

The log files on the cluster members show no useful information. The "cannotpoolreason" for the NVMe devices shows text similar to "waiting for verification", after the hang..

I've removed the cluster several times, cleaned the systems, and reinitiated the cluster create and enable s2d with same effect.

How may I further diagnose my problem?

---

VERBOSE: vms-c: 2016/10/16-16:25:56.110 Setting default fault domain awareness on clustered storage subsystem
VERBOSE: vms-c: 2016/10/16-16:25:56.751 Waiting until physical disks are claimed
VERBOSE: vms-c: 2016/10/16-16:25:59.767 Number of claimed disks on node 'VMS-1': 6/2
VERBOSE: vms-c: 2016/10/16-16:25:59.783 Number of claimed disks on node 'VMS-2': 6/2
VERBOSE: vms-c: 2016/10/16-16:25:59.798 Node 'VMS-1': Waiting until cache reaches desired state (HDD:'ReadWrite'
SSD:'WriteOnly')
VERBOSE: vms-c: 2016/10/16-16:25:59.798 SBL disks initialized in cache on node 'VMS-1': 6 (6 on all nodes)
VERBOSE: vms-c: 2016/10/16-16:25:59.814 SBL disks initialized in cache on node 'VMS-2': 6 (12 on all nodes)
VERBOSE: vms-c: 2016/10/16-16:25:59.814 Cache reached desired state on VMS-1
VERBOSE: vms-c: 2016/10/16-16:25:59.829 Node 'VMS-2': Waiting until cache reaches desired state (HDD:'ReadWrite'
SSD:'WriteOnly')
VERBOSE: vms-c: 2016/10/16-16:25:59.845 Cache reached desired state on VMS-2
VERBOSE: vms-c: 2016/10/16-16:25:59.845 Waiting until SBL disks are surfaced
VERBOSE: vms-c: 2016/10/16-16:26:03.267 Disks surfaced on node 'VMS-1': 12/12
VERBOSE: vms-c: 2016/10/16-16:26:03.298 Disks surfaced on node 'VMS-2': 12/12
VERBOSE: vms-c: 2016/10/16-16:26:06.945 Waiting until all physical disks are reported by clustered storage subsystem
VERBOSE: vms-c: 2016/10/16-16:26:10.188 Physical disks in clustered storage subsystem: 0


2012 R2 hyper V cluster - VMs not able to move to another node

$
0
0

hello friends...i recently created a 2 node cluster and whenever i am running the config / validation wizard i am not getting any error ... but every few hours .. my VMs on this go down..my switches are up .. i use link bonding... and one of my machine is specifically giving issues.. when i am trying to move VMs to this node .. it fails .. with below error codes.. when i look at the error there is nothing much but saying .. resource failure .. or policy failure for the node.. can you guys please guide me to the right path.. now also i am not able to move my VMs .. all Vms are on one node.. running validation shows all good .. but VMs not moving. below is screen shot




Thanks, Happiness Always
Jatin



Viewing all 6672 articles
Browse latest View live


<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>