I just built a new environment and was greeted by this error. This fix will likely work on other Dell servers, and the settings may apply to other vendors.
High level is you need to set TPM2 Algorithm Selection to SHA256 in the BIOS. You MIGHT have to turn on Intel TXT, and then enable Secure Boot. This SHOULD NOT impact the ESXi installation, but there is a chance it might. Enabling Secure Boot on a machine with modified or unsigned files carries with it the risk of rendering your machine unbootable with the current ESXi installation.
In This Post, I created certificates for my SRM & vCenter servers where I used a separate signing authority. What if you don’t have one, but still want to use your own certs? You create your own Root Certificate Authority (root CA) via OpenSSL. Here’s how…
Some of the documentation around creating certificates for vCenter or SRM seems to be lacking, so I documented a few steps for each and outlined the differences, also created a video :)
This can be done from any machine, as long as openssl is installed. If you’re creating/requesting multiple certs, create folders for each request and work from within there so you don’t mix them up. I use d:\cert\vcenter and d:\cert\srm. I added “D:\OpenSSL-Win32\bin\” to may path variable so it’ll work in any folder I’m in.
I upgraded one of my lab hosts to ESXi 4.1 yesterday and was plagued with this error:
A lot of people are getting this error without a lot of direction. As it turns out, this is because I upgraded one of my hosts to 4.1 without upgrading vCenter to 4.1. Silly me! Who would have thought that vSphere vCenter couldn’t manage a vSphere host because it’s rev is 0.1 higher?
ESXi 4.0 Update 1 brought with it one major update (as I pointed out here). Now that 4.1 was released on July 13th, I wanted to take a look and see if anything else major has been changed.
Biggest change was they lifted the 160 VMs per host in an 8-node HA cluster. Now it’s the maximum of 320 VMs per host, and a maximum of 32 nodes per HA cluster. Problem is, they imposed a maximum of 3000 VMs per cluster (standard, HA, or DRS, they no longer differentiate them), so you’d just have to find your sweet spot to maximize how you want your cluster set up. Not that 3000 VMs per cluster is a problem, but if you ran 320 VMs on 75% of a 32-node cluster (leaving 25% for failover), that’s 7680. That’s a difference of 4680 VMs. At any rate, I’m glad they lifted the 40 VMs per host in a 9+ configuration.
The Configuration Maximums for 4.1 can be found here.
Here are some of the key features that have changed:
When building a new cluster, your storage team (or you) may need to add several hosts into the shared storage zone. It’s a pain to go to each host, configuration, storage adapters, then copy out the WWN.
With this script, you can supply a vCenter server and Cluster/Folder/Datacenter (any logical container) and it will list all the WWNs for Fibre Channel devices. But what if you don’t have vCenter stood up yet? No problem, you can also supply a list of ESX/ESXi hosts to scan.
Shawn & I built this because we have 20 hosts we need the WWNs from to provide to our storage team, and vCenter isn’t alive yet.