System Administration > Configuration > Fabric > Nodes > Transport Nodes

Patch a Host Transport Node

Transport nodes are hypervisor hosts that will participate
in an NSX-T overlay. For a hypervisor host, this means that it hosts
VMs that will communicate over NSX-T logical switches.

This API creates transport node for a host node (hypervisor) in the transport network.

When you run this command for a host, NSX Manager attempts to install the
NSX kernel modules, which are packaged as VIB, RPM, or DEB files. For the
installation to succeed, you must provide the host login credentials and the
host thumbprint.

To get the ESXi host thumbprint, SSH to the host and run the
openssl x509 -in /etc/vmware/ssl/rui.crt -fingerprint -sha256 -noout
command.

To generate host key thumbprint using SHA-256 algorithm please follow the
steps below.

Log into the host, making sure that the connection is not vulnerable to a
man in the middle attack. Check whether a public key already exists.
Host public key is generally located at '/etc/ssh/ssh_host_rsa_key.pub'.
If the key is not present then generate a new key by running the following
command and follow the instructions.

ssh-keygen -t rsa

Now generate a SHA256 hash of the key using the following command. Please
make sure to pass the appropriate file name if the public key is stored with
a different file name other than the default 'id_rsa.pub'.

awk '{print $2}' id_rsa.pub | base64 -d | sha256sum -b | sed 's/ .*$//' | xxd -r -p | base64

Additional documentation on creating a transport node can be found
in the NSX-T Installation Guide.

In order for the transport node to forward packets,
the host_switch_spec property must be specified.

Host switches (called bridges in OVS on KVM hypervisors) are the
individual switches within the host virtual switch. Virtual machines
are connected to the host switches.

When creating a transport node, you need to specify if the host switches
are already manually preconfigured on the node, or if NSX should create
and manage the host switches. You specify this choice by the type
of host switches you pass in the host_switch_spec property of the
TransportNode request payload.

For a KVM host, you can preconfigure the host switch, or you can have
NSX Manager perform the configuration. For an ESXi host NSX Manager always
configures the host switch.

To preconfigure the host switches on a KVM host, pass an array
of PreconfiguredHostSwitchSpec objects that describes those host
switches. In the current NSX-T release, only one prefonfigured host
switch can be specified. See the PreconfiguredHostSwitchSpec schema
definition for documentation on the properties that must be provided.
Preconfigured host switches are only supported on KVM hosts, not on
ESXi hosts.

To allow NSX to manage the host switch configuration on KVM hosts,
ESXi hosts, pass an array of StandardHostSwitchSpec objects in the
host_switch_spec property, and NSX will automatically
create host switches with the properties you provide. In the current
NSX-T release, up to 16 host switches can be automatically managed.
See the StandardHostSwitchSpec schema definition for documentation on
the properties that must be provided.

The request should provide node_deployement_info.
This API is available when using VMware Cloud (AWS, Dell-EMC, Outpost, Hyperscalers) or VMware NSX-T.

Request:

Method:
PATCH
URI Path(s):
/policy/api/v1/infra/sites/<site-id>/enforcement-points/<enforcementpoint-id>/host-transport-nodes/<host-transport-node-id>
Request Headers:
n/a
Query Parameters:
TransportNodeUpdateParameters+
Request Body:
HostTransportNode+

Example Request:

PATCH https://<nsx-mgr>/policy/api/v1/infra/sites//enforcement-points//host-transport-nodes/ { "display_name": "VDS-TransportNode" "description": "NSX configured VDS Transport Node", "resource_type": "HostTransportNode", "host_switch_spec": { "host_switches": [ { "host_switch_id": "50 24 3c 5b a4 42 56 9e-c1 7e 5d 1d bd 52 61 14", "host_switch_type": "VDS", "host_switch_mode": "STANDARD", "host_switch_profile_ids": [ { "key": "UplinkHostSwitchProfile", "value": "0a26d126-7116-11e5-9d70-feff819cdc9f" } ], "uplinks": [ { "vds_uplink_name": "Uplink 1", "uplink_name": "uplink-1" } ], "ip_assignment_spec": { "resource_type": "AssignedByDhcp" }, "transport_zone_endpoints": [ { "transport_zone_id": "3c930f4e-1817-418d-aed6-d73f6b309e98", "transport_zone_profile_ids": [ { "resource_type": "BfdHealthMonitoringProfile", "profile_id": "52035bb3-ab02-4a08-9884-18631312e50a" } ] } ] } ], "resource_type": "StandardHostSwitchSpec" }, "maintenance_mode": "DISABLED", "node_deployment_info": { "os_type": "ESXI", "os_version": "7.0.0", "managed_by_server": "10.184.158.114", "discovered_node_id":"c59b9674-ce9b-4e52-9e8c-6bd1c6301ad3:host-13", "ip_addresses": [ "10.193.58.122" ] } }

Successful Response:

Response Code:
200 OK
Response Headers:
n/a
Response Body:
n/a

Required Permissions:

crud

Feature:

nodes_transport_nodes

Additional Errors: