In this post i will go over upgrading my 8.2 vRA appliance to 8.3. As a pre requirement we do need to have vRSLCM (vRealize Lifecycle Manager) upgraded to 8.3. Instructions can be found here
To get started we can go to vRealize Lifecycle Manager -> Lifecycle Operations -> Settings -> Binary Mapping. (If you havent added your My VMware credentials you will need to do that first by going to vRealize Lifecycle Manager -> Lifecycle Operations -> Settings -> My VMware)
Click on Add Binaries under Product Binaries
Select My VMware and click on Discover
We can see a list of binaries that have been discovered. We can select what we need and click on Add
This will create a request and start downloading the package. To view the progress we can click on the Click Here hyperlink
Click on the in Progress button to view the details
We now have to wait for the download to complete
After the download is complete we can go to Environments -> View Details on the environment that includes vRA
Click on Upgrade
An Inventory sync is recommended if the environment has changed since LCM performed the last sync. We trigger the sync from the UI or click on Proceed to continue
Select product Version 8.3.0 and click Next. We can also review the compatibility matrix to make sure the environment is compatible.
Run the Precheck to make sure there are no errors
Once the check is complete, click on Next. Review the upgrade details and click on Next. We are taken to the progress screen where we can follow the progress.
The system will get rebooted and once its back up we will be on 8.3
While doing maintenance on my vSAN cluster recently i had the need to disable the vCLS in order to fully shut down the cluster. Doing some reasearch i found kb article 80472 that talk about temporarily disabling the service in order to perform maintenance. The steps are fairly easy
First we need to get the Cluster id from vSphere. To do so all we need to do is select the cluster and look at the url. For example
With the release of of VCF 4.2 i wanted to get my lab upgraded. The release blog can be found here and the release notes are here
In order to get to 22.214.171.124 we have to upgrade to 126.96.36.199. We can do so by going to Repository -> Bundle management -> Download now
The next step is to upgrade VCF by going to Inventory -> Workload Domains -> Select the workload domain -> Update/Patches -> Update Now for the VMware Cloud Foundation Update 188.8.131.52. The release notes can be found here
Next we are taken to the Upgrade page where we can follow the upgrade for each one of the components
Once the upgrade is complete we can click Finish to be returned back to the main screen
Because we are changing the SDDC-Manager versions i would strongly recommend to clear cache and log back in before going forward.
Next is the 184.108.40.206 update. Repository -> Bundle management -> Download now. In my case i already had it downloaded so the next step is to apply the upgrade by going to Inventory -> Workload Domains -> Select the workload domain -> Update/Patches -> Update Now for the VMware Cloud Foundation Update 220.127.116.11. The release notes can be found here
Once the upgrade starts we can follow its progress
Once the upgrade is completed we can click finish and go to the next step
Again i would recommend clearing the cache since we changed sddc-manager versions.
Once the upgrade is complete we are taken back to the previous page where we can see that the ESXi servers are next. The release notes can be found here. Click on Download Now
Once the download is complete we can click on Update now
If we have multiple clusters we can enable Cluster-level selection and select the specific cluster(s) we want to upgrade.
We can also enable sequential cluster upgrade as well as quick boot
We get to review the options once again before we click finish to to submit the task
Once submitted we can view the status by clicking on View Status
And with that we are finished with the workload domain. We can get back to the Update/Patches page
The next update is the configuration drift bundle. We can go to inventory -> Workload Domains -> Select the workload domain -> Update/Patches -> Download now. You will notice a new drop down that allows us to pick the Cloud Foundation version.
Once the download is complete click on update now
Once the upgrade started i got redirected to the Update status page.
Considering the update is only 219 MB the upgrade went through pretty quickly. Once its completed we can click finish to get back to the main sddc manager page
Next step is to upgrade NSX-T installation to NSX-T 3.1.0. The release notes can be found here. We can go to inventory -> Workload Domains -> Select the workload domain -> Update/Patches -> Download now.
Once the download is complete click on Update Now
We can view the status and the steps by clicking on View Status.
Once the upgrade is complete we are redirected back to the available updates page showing that the vCenter server is next
Click on Download now and wait for the download to complete. Once the download is complete click on update now
We can view the task by clicking on View Status
Next are the the additional domains that we might have where we can follow the same instructions as above. The process will be allot quicker because the upgrades are already downloaded
If no additional upgrades are needed we can clean up the downloads by following the instructions on my other post here
I wanted to go through the upgrade of vIDM 3.3.4 outside of lifecycle manager. I wanted to do it using the online upgrade via the console. Official documentation can be found here. First step is to log on to the console via the root credentials and upgrade the upgrade manager to 3.3.4
Because we are switching operating systems to Photon OS 3 we need to download the dual boot update package from here. Look for VMware Identity Manager Service Virtual Appliance Dual Boot Update and transfer it to the vIDM server
And finally run the update
Specify the path to the dualboot file. It would look something like this:
If everything went through properly you would be greeted with something like this
All we need to do now is to reboot.
If you are greeted by the following screen follow my other post here
Assuming that you already downloaded the SaltStack Config binaries in this guide we will go over the deployment of SaltStack Config in an existing VRA 8.3 environment via Lifecycle Manager 8.3
Log on to Lifecycle Manager -> Go to environments, on the vRA environment click on the Menu and go to Add Product
Select the checkbox next to vRealize Automation SaltStack Config and click Next
Review the EULA, click on the agree box on the bottom of the page and click on next
Click on Select and select a vrealize or vra advanced\enterprise. Clieck on the validate button and click next
Select an existing certificate or click on the + sign and generate a new certificate. In my case i didnt have a certificate created so im going to create a new one. If you already have a certificate this step can be skipped.
Complete the fields and click on generate
Now the list is populated with the certificate we just created. Click Next
Select the infrastructure details and click on Next
Verify the network details and click next
If the tenant list is empty a sync of vra is required
To do this we can go to Environments. Dont worry the progress will be saved
Click on submit
Once the sync is complete we can go back to requests
Under requests click on Pending Action
Complete the details and click next
Click on Run precheck in order to make sure everything is in a working order
If everything looks good click on Next
Verify the summary and click on Submit
LCM will go through everything needed to. As we can see theres allot of steps that otherwise would need to perform manually
Get a cup of coffee and wait. In my case it took about 13 minutes for the installation to finish
In this guide i will go over the deployment of vRA 8.3 using the Easy Installer. The reason i like the Easy Installer is because it deploys everything that i need. vIDM, Lifecycle Manager as well as vRA.
To get started we need to download the installer from here.
Once downloaded mount the iso as a drive using your favorite tool or extract it as a archive using something like 7-zip
Once mounted go to the CDROM in the mac directory if on mac or win32 if you are on windows. In here run the installer file
We will be presented with the following window:
The migration option allows us to migrate from from vRSLCM 2.1 to 8.3, it can migrate Datacenter and vCenters, all existing environments, DNS, SNMP, NTP, My VMware, proxy details, migration of vIDM installation as well as Blackstone Content endpoints. For the purpose of this guide we will process with the Install option
Once we click Install we will be presented with an Introduction of what components will be included part on the installation. In my case i can see vRealize Automation 8.3, Lifecycle Manager 8.3, Identity Manager 3.3.4.
Click Next, Review, accept the agreement and click next
Provide the vCenter details and click next. In my case i used the administrator account. A detailed list of permissions needed for deployment can be found here
Sections 4, 5, 6 and 7 are straight foward so i will skip them
In section 8 select a password that will be used across all products for the local usernames. ex for LCM root and local admin, vIDM admin, sshuser, root etc
Select a VM name for the Lifecycle Manager appliance and its ip credentials. If you are expecting a large repository in lifecycle manager we can add additional disk under the Increase Disk Size in GB section.
The FIPS Mode compliance option enforces FIPS compliance, however keep in mind that with the FIPS mode enabled there are limitation on what the product can do. This can be disabled later and re enabled as needed.
Complete the fields and click next
The next portion is vIDM. In case there is an existing vIDM appliance in the environment we can also import and existing vIDM appliance. In my case i will deploy a new one.
One important option under vIDM is the Sync Group Members to the Directory When Adding Group. When enabled, members of the groups are synced when groups are added from Active Directory. When this is disabled, group names are synced to the directory, but members of the group are not synced until the group is entitled to an application or the group name is added to an access policy. Note: Post deployment this value cannot be changed from vRealize Suite Lifecycle Manager. To update this field post deployment, navigate to VMware Identity Manager
My configuration page looks like this:
The next section is the vRA Configuration. In here we have a couple of options. We can perform a standard 1 node deployment or a cluster deployment which includes 3 appliances. The FIPS Compliance mode enables FIPS compliance. Unlike LCM this mode cannot be disabled after the deployment. This disables a number of options in vRA from an LCM perspective. Please make sure that its only enabled if required.
The advanced configuration at the bottom of the page includes the option to change the internal Kubernetes cluster and Service ip range. This is useful if the default range is already in use on the internal network. We want to make sure we pick an ip range thats not used somewhere else in order to avoid conflicts. Once complete click next
The next page gives us a summary of our deployment and we can click submit to start the deployment.
Next we are presented with the installation process. We can follow it along, in my environment the full deployment took about 1.5 hrs
After the install is complete we are presented to links for the different services
Next i would recommend LCM certificate management found here