Enter the required information as shown in the following example.
Do you want to configure static IPv4 address?(y/N) y Enter Static IPv4 Address (e.g. 10.136.97.98) 10.136.7.1 Enter Netmask (e.g. 255.255.255.0) 255.255.255.0 Enter Gateway IP Address (e.g. 10.136.97.1) 10.136.72.1 Enter DNS Server 1 IP Address (e.g. 10.136.74.189) 10.136.72.189 Enter DNS Server 2 IP Address (e.g. 10.136.74.190) 10.136.72.190 Enter Domain (e.g. blr.ste.lab) bltr.ste.lab
Cross hypervisor disaster recovery (CHDR) provides the ability to migrate the VMs from one hypervisor to another (ESXi to AHV or AHV to ESXi) by using the protection domain semantics of protecting VMs, generating snapshots, replicating the snapshots, and then recovering the VMs from the snapshots
install and configure NGT on all the VMs
When you recover VMs from the AHV hypervisor on the ESXi hypervisor, the VMs come up with E1000 NIC attached by default. However, if you want to bring the VMs up withVMXNET NIC is attached, you must install VMware Tools
Static IP address preservation is not supported for CHDR workflows. If a VM has an astatic IP address configured, you must manually reconfigure the VM IP address after its recovery on the target site.
Hypervisor-specific properties like multi-writer flags, independent persistent and independent non-persistent disks, or changed block tracking (CBT) is not preserved in the CHDR operation
Nutanix supports only VMs with flat files. Nutanix does not support vSphere snapshots or delta disk files. If you have delta disks attached to the VM and you proceed with the migration, the VMs are lost
Nutanix does not support VMs with attached volume groups or shared virtual disks
Nutanix supports IDE/SCSI and SATA disks only. PCI/NVMe disks are not supported
Nutanix supports CHDR on both the sites with all the AOS versions that are not EOL
Set the SAN policy to OnlineAll for all the Windows VMs for all the non-boot SCSI disksso that they can automatically be brought online. For more information about setting SANpolicy, see Bringing Multiple SCSI Disks Online on page 208
(For migration from AHV to ESXi) Automatic reattachment of iSCSI-based volume groups(VGs) fails after VMs are migrated to ESXi because vCenter and, in turn, the Nutanix cluster do not obtain the IP addresses of the VMs. Therefore, after migration, manually reattach theVGs to the VMs
Do not enable VMware vSphere Fault Tolerance (FT) on VMs that are protected with CHDR.If already enabled, disable VMware FT. VMware FT does not allow registration of VMs. enabled with FT on any ESXi node after migration. This also results in a failure or crash of Uhura service
Limitations of CHDR
Nutanix cluster nodes to vSphere cluster mapping: The Nutanix data distribution architecture does not support the mapping of the nodes of a Nutanix cluster to multiple vSphere clusters
Ensure that all the nodes in the Nutanix cluster are in the same vSphere (ESXi) host cluster and that the network is available on all the nodes in the Nutanix cluster at the primary or remote site. Any configuration using Nutanix DR where there is more than one ESXi cluster on top of a single Nutanix cluster is not supported
Recommends that each application that constitutes a set of entities is protected by a unique protection domain
Supports one snapshot per hour as the shortest possible snapshot frequency for asynchronous replication
Does not support snapshots of the entire file systems or storage containers
Supports a VM to be a part of a protection domain only if the VM is entirely on a Nutanixdatastore and not on an external storage
Does not support running multiple third-party backups in Parallel on the same set of VMs.The system becomes unstable
Does not support VMX editing when you restore a VM. As a result, the characteristics of the restored VM such as MAC addresses may be in conflict with other VMs in the cluster
The replication speed reduces when deduplication is enabled on storage containers containing protected VM
Nutanix Cloud Infrastructure (NCI) is a complete software stack to unify your hybrid cloud infrastructure including compute, storage and network, hypervisors, and containers, in public or enterprise clouds; all with built-in resilience, self-healing, disaster recovery capabilities, and security. It includes enterprise data services and consolidated storage, data protection and disaster recovery, native virtualization and container management, networking, and security.
How it looks like in BOQ
NCI Software Editions
Nutanix Cloud Manager (NCM) offers our customers simplicity and ease of use to build and grow their cloud deployments faster and realize rapid ROI, by providing intelligent operations, self service and orchestration, visibility and governance of spend, security and teams, all through a unified Multi-cloud management solution.
NCM licenses can be purchased and applied on the number of physical CPU cores capacity in your deployment. Licenses are portable across hardware platforms and are available in 1 through 5-year term options.
By default, NCM provides coverage for all Nutanix and on-prem VMWare environments, metered per core. For supporting public cloud environments using the same NCM deployment, customers should purchase appropriate NCM Cloud SKUs as add-ons. Three public cloud-focused add-ons are available as SKUs- NCM Self-Service add-on for Public Cloud SKU, NCM Cost Governance SaaS SKU, and NCM Security Central SaaS SKU. These add-ons are metered by the number of Virtual Machines (VM) managed in the public cloud. Note: For on-prem environments, Cost Governance is available for AHV and ESXi on AOS, and Security Central is known for AHV.
NCM is also available as a fully managed Software as Service Option. Customers can experience multi-cloud self-service, app automation, governance, and security compliance capabilities, without requiring to run any on-prem Nutanix software. The NCM SaaS offering is available to purchase as à la carte SaaS licenses for these four NCM SaaS modules:
NCM SaaS – Operations (in development)
NCM SaaS – Self-Service
NCM SaaS – Cost Governance
NCM SaaS – Security Central
NCM SaaS licenses are metered by the number of Virtual Machines (VM) managed in the public cloud.
Nutanix Cloud Infrastructure (NCI) and Nutanix Cloud Manager (NCM) can be purchased together in 3 ‘better together’ Nutanix Cloud Platform (NCP) bundles:
ICAP stands for internet content adaption protocol is an open standard being adopted to connect devices to enterprise-level virus scan engine . same way with the nutanix files is to enable communication with external servers hosting third-party anti-virus software to scan inbound data (files) in transit via Secure Proxy before sending it to the backend destination server.
The ICAP service runs on each Nutanix Files file server and can interact with more than one ICAP server in parallel to support horizontal scale-out of the antivirus server. The scale-out nature of Files and one-click optimization greatly mitigate any antivirus scanning performance overhead. If the scanning affects Nutanix Files FSVM performance, one-click optimization recommends either increasing the virtual CPU resources or scaling out the FSVMs. This feature also helps both the ICAP server and Files scale out, ensuring fast responses from the customer’s antivirus vendor
WHY Nutanix files integration with AV server is important
Ransomware is a persistent concern that requires multiple security controls and software layers to mitigate integration is important to protect users from malware and viruses,
WHAT all third-party vendor are support with Nutanix files
HOW to configure integration
In the Files Console, go to Configuration > Antivirus.
Connect the ICAP server.
Click + Connect ICAP Server.A new row appears for new ICAP server details.
Enter the following information in the corresponding fields:
IP address or hostname
Port (the default port number is 1344)
To save the configuration, click the check mark icon.For a detected antivirus server, the software tests the validity of the configured server and updates the status to OK.
Ensure the connection status automatically updates to OK.
AOS integrates storage, compute and networking resources into a scalable, secure, resilient and easy-to-use software-defined infrastructure.
AOS 6.0 has various new features and notably enables advanced disaster recovery capabilities for organizations of all sizes and expands workload support for leading big data and analytics applications. Additionally, has other workload acceleration and enterprise grade features, such as resiliency improvements to gain insights and control over the self-healing process and ability to reserve capacity.
Please review the release notes before deploying AOS 6.0
Here is a sample of the latest features:
Near-Sync Autonomous Schedules: Native near-zero always-on data loss solution for applications running on Nutanix HCI. Available with Protection Domain
DR Dashboard: Provides global visibility into DR landscape with LEAP. Gives team one place to monitor and manage their ability to keep business operations functional in the face of disaster. Brings observability to DR for meeting/maintaining RPO SLAs, Recovery Readiness and DR Reports
DR Traffic Encryption: Encrypt DR traffic across WAN links without the need for dedicated encryption hardware or VPN.Native SSL based encryption is now integrated with Nutanix AOS, enabling secure replication even where VPN tunnels are unavailable. Available with Protection Domain
DR Replication Performance: DR Replication Performance: Next gen architecture for data mobility across clouds with improved performance and tight SLAs
Instant Restore for AHV: Instantly recover the workload even though the disk data for the workload are not locally available in the underlying AOS cluster. Migrate the disk data back to (seed) the AOS cluster without interruption to the recovered workload
Core Data Path
Option to reserve rebuild(spare) capacity for a cluster: The Reserve Rebuild Capacity feature simplifies cluster resiliency management by allowing you to reserve storage capacity required to recover and rebuild from failures. AOS calculates, reserves, and dynamically manages capacity required for self-healing after a failure.
Rebuild progress indicator and ETA: The new Rebuild progress indicator in Prism enables administrators to better manage cluster resiliency by providing additional insight into the self-healing process such as the rate of progress, expected time of completion and other details for rebuild operations
Brownfield AES: Provides ability to convert existing pre-AES containers to AES
Support mix storage capacity for HW Swap: This feature adds support mixing storage drive capacity for drive RMA and node capacity increase use-cases.
Improve performance (e.g.: metadata re-warming) after failures/CVM restart/change of vdisk host: This feature helps mitigate potential application performance impact during certain CVM restarts or live migrations by proactively pre-filling the metadata cache
P2P intra cluster VM live migration: While VM live migration is in progress if the migrate task restarts, VMs can be left in a paused state or killed in certain scenarios causing downtime, this feature overcomes these issues
ADS vGPU Support: We now have Acropolis Dynamic Scheduling (ADS) support for VMs with virtual GPUs (vGPUs)
Nvidia vGPU Console Support: Provide GPU support for our VM console (Also part of 5.20)
Scale out PE Lazan: With this feature, cluster state capture phase is distributed among every Lazan slave to make it faster and more scalable. Each slave then sends the captured state to Lazan master making anomaly detection faster and the CPU usage to compute the cluster state is also distributed among the nodes.
PC UI using v3 APIs: The goal was to upgrade the APIs used for VM operations for our Prism UI, while preserving the existing feature set and workflows
By default, a virtual NIC (VM NIC) on a guest VM operates in access mode. In this mode, the VM NIC can send and receive traffic only over its own specified VLAN. If restricted to using access mode interfaces, a VM running an application on multiple VLANs (such as a firewall application) must use multiple VM NICs—one for each VLAN. Instead of configuring multiple VM NICs in access mode, you can configure a single VM NIC to operate in trunked mode. A VM NIC in trunked mode can send and receive traffic over any number of VLANs in addition to its own VLAN. You can trunk specific VLANs or all VLANs. A VM NIC in trunk mode can convert back to access mode, meaning the VM NIC reverts to sending and receiving traffic only over its own VLAN.
acli vm.nic get <vm_name>
acli vm.nic_create <VM_name> network=<net_name> vlan_mode=kTrunked trunked_networks=<comma separated VLAN ID list>
To update an existing virtual NIC from access mode to trunked mode.
Jumbo Frame is basically a ability to increase packet payload size which can be beneficial when processing a large amount of data through the network itself.
Nutanix AHV uses the standard Ethernet max transmission unit (MTU) of 1,500 bytes for all external communication by default. Jumbo frames are optional, because the standard 1,500 byte MTU still delivers excellent performance, but increasing the MTU can be beneficial in some scenarios
Default MTU vaule is 1500
Jubmo Frame MTU Vault is 9000
Nutanix recommanded to use Jubmo frame to use for Nutanix Volume to represent high Performance ISCSI Storage External Server which can help to increase Effciency for application such as oracle database connected to the Nutanix volumes .
Be sure to plan carefully before enabling jumbo frames. Configure the entire network infrastructure from end to end with jumbo frames to avoid packet fragmentation, poor performance, or disrupted connectivity.
Check the current configuratin :-
root@ahv# ovs-appctl bond/show
root@ahv# ip link | grep mtu
Reconfigure CVM to make changes :-
nutanix@cvm$ sudo ip link set eth0 mtu 9000
nutanix@cvm$ sudo ip link set eth2 mtu 9000
To permanently modify the MTU so it persists through a restart, append the string ‘MTU=”9000“’ to the end of the following configuration files on the CVM: