Translate

Total Pageviews

My YouTube Channel

Monday, 19 February 2018

Monday, 12 February 2018

New User Roles in VMware vRA 7.2 and vRA 7.3

I have covered vRA 7.x User Roles in my previous blog post that you can check it from here http://virtualization24x7.blogspot.in/2016/05/vmware-vrealize-automation-7x-user-roles.html
Now in this blog post we will only discuss about the new roles those are introduced in vRA 7.2 and vRA 7.3, this means to use some roles you need vRA 7.2 or later release and to use some other roles you need vRA 7.3 or later.


Sunday, 4 February 2018

VMware vSAN - Synchronizes all Storage Providers with the current State of the Environment

Whenever ESXi Hosts are added in the vSAN Cluster, these ESXi Hosts offers Storage Providers (One Per ESXi). These Storage Providers Provides Out of band information about the vSAN. If these are not communicating with SMS (Storage Monitoring Service) on vCenter, then vCenter will not be able to display any of the capabilities of the vSAN datastore. For this you need to run the re-synchronization operation. This can be triggered by:-
Select vCenter > Configure Tab > Storage Providers > Click on "Synchronizes all Storage Providers with the current State of the Environment"

 

What's New in VMware Horizon View 7.4

VMware Horizon 7 version 7.4 provides the following new features and enhancements. This information is grouped by installable component.

Horizon Connection Server

  • Cloud Pod Architecture
    • The Cloud Pod Architecture feature now supports one-way Active Directory trust domains. Horizon administrators can implement one-way trust topologies in a Cloud Pod Architecture environment, and end-users can select and launch remote desktops and published applications across the trust.
    • Users can now launch remote desktops and applications on a later version pod when connected to an earlier version pod in a Cloud Pod Architecture environment.
    • You can enable the Session Collaboration feature for global desktop entitlements.   
  • Published Desktops and Application Pools
    • You can use NVIDIA GRID vGPU for 3-D rendering when you create an instant-clone farm.
    • The "Empty session timeout (applications only)" option provides a new “immediate” timeout value in the farm settings to log off the session when the last application window is closed.

Horizon Agent for Linux

  • Instant-Clone Floating Desktop Pool
    You can create a Linux-based instant-clone floating desktop pool with Horizon 7 Agent for Linux. This feature is supported only on Ubuntu 14.04 and 16.04 systems.
  • Active Directory Integration with PBISO 
    Linux-based desktops are allowed to authenticate with Active Directory using PowerBroker Identity Services Open (PBISO). This feature is supported only on Ubuntu 14.04 and 16.04 systems.
  • 3Dconnexion Mouse
    The 3Dconnexion mouse support is available in Horizon 7 Agent for Linux  using USB redirection.
  • Additional platform
    SUSE Linux Enterprise Server 12 SP3 and SUSE Linux Enterprise Desktop 12 SP3 are supported.

Horizon Agent

  • Session Collaboration
    With the Session Collaboration feature, users can invite other users to join an existing remote desktop session.
  • Fingerprint Scanner Redirection by Using the Device Bridge BAS Plugin
    You can redirect biometric devices, specifically fingerprint scanners, that are plugged into a serial port on a Windows client system to virtual desktops, published desktops, and published applications by using the device bridge feature. You can install the Device Bridge BAS Plugin by selecting the Device Bridge BAS Plugin custom setup option in the Horizon Agent installer.
  • Secure boot enabled machines supported with URL Content Redirection
    URL content can now be redirected from machines that have secure boot enabled.
  • URL Content Redirection with Chrome 
    URL content can now be redirected in the Chrome browser. Both agent-to-client and client-to-agent redirection are supported. To use this feature, the VMware Horizon URL Content Redirection Helper extension must be installed and enabled in Chrome.

Horizon GPO Bundle

  • Configure Session Collaboration group policy setting
    Use the Collaboration group policy setting in the VMware View Agent Configuration ADMX template file (vdm_agent.admx) to configure Session Collaboration on remote desktops.
  • Device Bridge BAS Plugin group policy setting
    Use the Disable Device Bridge BAS Plugin group policy setting in the VMware View Agent Configuration ADMX template file (vdm_agent.admx) to disable the Device Bridge BAS Plugin.
  • Server clipboard memory size group policy settings for VMware Blast and PCoIP
    Use the Clipboard memory size on server group policy setting for VMware Blast sessions, and the Configure clipboard memory size on server for PCoIP sessions, to specify the server's memory size value for clipboard redirection in bytes or kilobytes. The VMware Blast setting is in the VMware Blast ADMX template file (vdm_blast.admx). The PCoIP setting is in the PCoIP ADMX template file (pcoip.admx).
  • H264 High Color Accuracy
    The VMware Blast setting in the VMware Blast ADMX template file (vdm_blast.admx) increases color accuracy when using H.264 encoding by using the YUV 4:4:4 colorspace instead of 4:2:0.

Horizon Client

For information about new features in Horizon Client 4.7, including HTML Access, see the Horizon Clients Documentation page.
Reference VMware Release Notes 

Tuesday, 23 January 2018

Supported KMS Servers with vSAN 6.6

To use the vSAN Encryption feature KMS Server is needed. One should know what are KMS Servers supported by vSAN 6.6. To Know this, it is always recommended to check VMware Compatibility Guide. Link for the Compatibility Guide is <<<<<here>>>>> 
Remember to get the latest info always refer this Compatibility Guide

vSAN Ready Nodes - AF Series and HY Series


vSAN Ready Node provides the most flexible hardware options to build Hyper-Converged Infrastructure based on VMware Hyper-Converged Infrastructure Software.

AF = All Flash
HY = Hybrid

https://vsanreadynode.vmware.com/RN/RN

Download vSAN Ready Nodes Wallpaper from here:-
https://www.vmware.com/resources/compatibility/pdf/vi_vsan_rn_guide.pdf

Get vSAN Ready Nodes Info from here:-
https://vsanreadynode.vmware.com/RN/RN

Sunday, 21 January 2018

Vembu BDR Suite v3.9.0 is now Generally Available and includes Tape Support & Flexible Restores


It is highly important that data needs to be backed up and there should be an effective Disaster Recovery plan in case of data threat or a catastrophe. While data continues to grow and there are number of technology providers who offer better and comprehensive storage techniques to businesses, there has not been an alternative to the concept of backup. While costs are a major factor for businesses, having a steady backup plan to counter data threats and compliant to strict regulatory standards(including the upcoming EU’s GDPR) is necessary. Be it virtual environment backup like VMware Backup, Microsoft Hyper-V Backup or legacy environment backup like Windows Server Backup, Workstation backup, Vembu BDR Suite has been offering Backup & Recovery with their own file-system, VembuHIVE thereby easing the backup process, storage management at an extremely affordable pricing.

Last week, they did announce the release of Vembu BDR Suite v3.9.0 which offers manifold features and enhancements to meet the different needs of Diverse IT environments. According to them, the overall goal of the new version v3.9.0 is to provide advancements in terms of Storage, Security, and Data Restoration.

Vembu BDR Suite v3.9.0 release is distinct because a number of critical features are incorporated for maintaining business continuity and to function effectively for high availability. Here are some of the key highlights of this release:

Tape Backup Support
Vembu now provides the popular 3-2-1 backup strategy( copies of backup in 2 medias(Disk and Tape) and 1 backup copy at offsite) to businesses by announcing the support for Native Tape Backup for  Image-based Backups (VMware, Hyper-V, and Physical Windows Servers & Workstations) providing an option for Long-Term Archival and Offsite storage.  Also, Vembu Tape Backup Support makes the DR possible on any physical or virtual environment. Thus Vembu Tape backup is designed keeping in mind the future needs of the ever evolving IT demands.


Quick VM Recovery on ESXi host for Hyper-V and Windows Image Backups
As we speak of data backup, recovery of data is equally important, if not more. While data recovery is crucial, the amount of time taken to restore data decides the business continuity of any organization. In the previous versions, Vembu has provided instant recovery capabilities only for VMware Backups from the GUI. From this release v3.9.0, Vembu makes the instant recovery process much simpler and quicker than before by making the Quick VM Recovery possible on VMware ESXi from Vembu BDR backup server console for all image-based backups(VMware, Hyper-V and Microsoft Windows). Thus, Vembu lowers the Recovery Time Objectives of the organizations and provides quicker data regain and access.

Backup-level Encryption
With the newest release, Vembu provides the ability to encrypt the data while creating a backup job. Each backup job that is configured from the distributed agents or through the Vembu BDR backup server is now highly secured through Backup-level Encryption. By using customized passwords, users will now be able to enable additional security for their backup jobs. And the backup data can be restored/accessed only by providing the password. Thus, the data is encrypted and can only be accessed by authorized users. This step is to fight data threats and also to ensure data compliances.
   
Auto Authorization at Vembu OffsiteDR Server
Offsite data protection is critical in terms of business continuity and is primarily done to keep a backup instance of key business data. To increase data security, Vembu BDR Suite v3.9.0 has Auto Authorization feature at Vembu OffsiteDR server that lets only the registered BDR backup servers to connect to the OffsiteDR server. Vembu BDR servers are authorized through unique registration key generated at the OffsiteDR server. Thus, safeguard all your backup data even if they are transferred to offsite through Auto-Authorization at OffsiteDR server
 
Many businesses are in a need to execute certain business logic before or after a backup job. But running these logics manually through scripts is difficult and is not feasible for organizations having multiple backup jobs running in their IT infrastructure. To make this process simpler, Vembu BDR Suite v3.9.0 provides a separate wizard in NetworkBackup, OnlineBackup and ImageBackup clients, where one can add a number of pre and post executable commands/scripts. This helps in automatically executing the added commands/scripts at specific stages based on the configuration and provides the ability to run the custom actions before/after the backup schedules.
 
Besides all the listed features, Vembu BDR Suite v3.9.0 has few interesting features like Windows Event Viewer Integration along with some Enhancements.

Interested in trying Vembu BDR suite?, Try Now on a 30-days free trial

Saturday, 20 January 2018

vROPS 6.6 Sizing Guidelines Worksheet

Access KB 2150421 from <<<<<here>>>>>. Scroll Down till the end of KB article and there you will find link for downloading the attachment under attachment category.

https://kb.vmware.com/s/article/2150421
 

vRealize Operations Manager 6.6 and 6.6.1 Sizing Guidelines (2150421)

By default, VMware offers Extra Small, Small, Medium, Large, and Extra Large configurations during installation. You can size the environment according to the existing infrastructure to be monitored. After the vRealize Operations Manager instance outgrows the existing size, you must expand the cluster to add nodes of the same size.
Characteristics/
Node Size
Extra
Small
SmallMediumLargeExtra Large
Standard Size Remote
Collectors
Large Size Remote Collectors
vCPU248162424
Memory (GB)8163248128416
Maximum Memory Configuration (GB)N/A326496N/A832
Datastore latencyConsistently lower than 10 ms with possible occasional peaks up to 15 ms
Network latency for data nodes< 5 ms
Network latency for remote collectors< 200 ms
Network latency between agents and vRealize Operations Manager nodes and remote collectors< 20 ms
vCPU: Physical core ratio for data nodes (*)1 vCPU to 1 physical core at scale maximums
IOPSSee the attached Sizing Guidelines worksheet for details.
Disk Space
Single-Node Maximum Objects2502,4008,50015,00035,0001,500 (****)15,000 (****)
Single-Node Maximum Collected Metrics (**)70,000800,0002,500,0004,000,00010,000,000600,0004,375,000
Multi-Node Maximum Objects Per Node (***)NA2,0006,25012,50030,000NANA
Multi-Node Maximum Collected Metrics Per Node (***)NA700,0001,875,0003,000,0007,500,000NANA
Maximum number of nodes in a cluster12161665050
Maximum number of End Point Operations Management agents per node1003001,2002,5002,5002502,000
Maximum Objects for the configuration with the maximum supported number of nodes (***)250400075,000150,000180,000NANA
Maximum Metrics for the configuration with the maximum supported number of nodes(***)70,0001,400,00019,000,00037,500,00045,000,000NANA
 
(*) It is critical to allocate enough CPU resources for environments running at scale maximums to avoid performance degradation. Refer to the vRealize Operations Manager Cluster Node Best Practices in the vRealize Operations Manager 6.6 Help for more guidelines regarding CPU allocation.
(**) Metric numbers reflect the total number of metrics that are collected from all adapter instances in vRealize Operations Manager. To get this number, you can go to the Cluster Management page in vRealize Operations Manager, and view the adapter instances of each node at the bottom of the page. You can get the number of metrics collected by each adapter instance. The sum of these metrics is what is estimated in this sheet. Note: The number shown in the overall metrics on the Cluster Management page reflects the metrics that are collected from different data sources and the metrics that vRealize Operations Manager creates.
(***) In large, 16-node configurations, note the reduction in maximum metrics to permit some head room. This adjustment is accounted for in the calculations.
(****) The object limit for the remote collector is based on the VMware vCenter adapter.

What's new with vRealize Operations 6.6.x sizing:
Monitor larger environment with improved scale: The vRealize Operations 6.6 cluster can handle 6 Extra Large Nodes in a cluster which can support up to 180,000 Objects and 45,000,000 metrics.
Monitor more vCenter Servers: A single instance of vRealize Operations can now monitor up to 60 vCenter Servers.
Deploy larger sized nodes: A large Remote Collector can support up to 15000 objects.
Notes:
  • Maximum number of remote collectors (RC) certified: 50.
  • Maximum number of VMware vCenter adapter instances certified: 60.
  • Maximum number of VMware vCenter adapter instances that were tested on a single collector: 40.
  • Maximum number of certified concurrent users: 200.
  • This maximum number of concurrent users is achieved on a system configured with the objects and metrics at 50% of supported maximums (For example: 4 large nodes with 20K objects or 7 nodes medium nodes with 17.5K objects). When the cluster is running with the nodes filled with objects or metrics at maximum levels, then the maximum number of concurrent users is 5 users per node (For example: 16 nodes with 150K objects can support 80 concurrent users).
  • Maximum number of the End Point Operations Management agents per cluster certified - 10,000 on 4 large nodes cluster.
  • When High Availability (HA) is enabled, each object is replicated in one of the nodes of a cluster. The limit of objects for HA based environment is half compared to a non-HA configuration. vRealize Operations HA supports only one node failure and you can avoid Single-point-of-failure by adding vRealize Operations cluster nodes into different hosts in the cluster.
  • An object in this table represents a basic entity in vRealize Operations Manager that is characterized by properties and metrics that are collected from adapter data sources. Example of objects include a virtual machine, a host, a datastore for a VMware vCenter adapter, a storage switch port for a storage devices adapter, an Exchange server, a Microsoft SQL Server, a Hyper-V server, or Hyper-V virtual machine for a Hyperic adapter, and an AWS instance for a AWS adapter.
  • The limitation of a collector per node: The object or metric limit of a collector is the same as the scaling limit of objects per node. The collector process on a node will support adapter instances where the total number of resources is not more than 2,400, 8,500, and 15,000 respectively, on a small, medium, and large multi-node vRealize Operations Manager cluster. For example, a 4-node system of medium nodes will support a total of 25,000 objects. However, if an adapter instance needs to collect 8,000 objects, a collector that runs on a medium node cannot support that as a medium node can handle only 6,250 objects. In this situation, you can add a large remote collector or use a configuration that uses large nodes instead of small nodes.
  • A large node can collect more than 20,000 vRealize Operations for Horizon objects when a dedicated remote collector is used.
  • A large node can collect more than 20,000 vRealize Operations for Published Apps objects when a dedicated remote collector is used.
  • If the number of objects is close to the high-end limit, dependent on the monitored environment, increase the memory on the nodes. Contact Product Support for more details.
  • The performance of vRealize Operations Manager can be impacted by the usage of snapshots. Presence of a snapshot on the disk causes slow IO performance and high CPU costop values which degrade the performance of vRealize Operations Manager. Use snapshots minimally in the production setups.
  • The sizing guides are version specific, please use the sizing guide based on the vRealize Operations version you are planning to deploy.
    Extra small and small node are designed for test environment and POC, we do not recommend to scale-out small node more than two nodes and we do not recommend to scale out extra small node.
  • Simply scale up as you scale - Increase memory instead of configuring more nodes to monitor larger environments. We recommend to scale up to the possible maximum (default memory x 2) and then do scale out based on underlying hardware that will support the scale requirements. Example: Large node default memory requirements are 48GB and now, if needed, can be configured up to 96GB. All nodes must be scaled equally.