304 lines
15 KiB
ReStructuredText
304 lines
15 KiB
ReStructuredText
|
.. SPDX-License-Identifier: GPL-2.0
|
||
|
|
||
|
VMbus
|
||
|
=====
|
||
|
VMbus is a software construct provided by Hyper-V to guest VMs. It
|
||
|
consists of a control path and common facilities used by synthetic
|
||
|
devices that Hyper-V presents to guest VMs. The control path is
|
||
|
used to offer synthetic devices to the guest VM and, in some cases,
|
||
|
to rescind those devices. The common facilities include software
|
||
|
channels for communicating between the device driver in the guest VM
|
||
|
and the synthetic device implementation that is part of Hyper-V, and
|
||
|
signaling primitives to allow Hyper-V and the guest to interrupt
|
||
|
each other.
|
||
|
|
||
|
VMbus is modeled in Linux as a bus, with the expected /sys/bus/vmbus
|
||
|
entry in a running Linux guest. The VMbus driver (drivers/hv/vmbus_drv.c)
|
||
|
establishes the VMbus control path with the Hyper-V host, then
|
||
|
registers itself as a Linux bus driver. It implements the standard
|
||
|
bus functions for adding and removing devices to/from the bus.
|
||
|
|
||
|
Most synthetic devices offered by Hyper-V have a corresponding Linux
|
||
|
device driver. These devices include:
|
||
|
|
||
|
* SCSI controller
|
||
|
* NIC
|
||
|
* Graphics frame buffer
|
||
|
* Keyboard
|
||
|
* Mouse
|
||
|
* PCI device pass-thru
|
||
|
* Heartbeat
|
||
|
* Time Sync
|
||
|
* Shutdown
|
||
|
* Memory balloon
|
||
|
* Key/Value Pair (KVP) exchange with Hyper-V
|
||
|
* Hyper-V online backup (a.k.a. VSS)
|
||
|
|
||
|
Guest VMs may have multiple instances of the synthetic SCSI
|
||
|
controller, synthetic NIC, and PCI pass-thru devices. Other
|
||
|
synthetic devices are limited to a single instance per VM. Not
|
||
|
listed above are a small number of synthetic devices offered by
|
||
|
Hyper-V that are used only by Windows guests and for which Linux
|
||
|
does not have a driver.
|
||
|
|
||
|
Hyper-V uses the terms "VSP" and "VSC" in describing synthetic
|
||
|
devices. "VSP" refers to the Hyper-V code that implements a
|
||
|
particular synthetic device, while "VSC" refers to the driver for
|
||
|
the device in the guest VM. For example, the Linux driver for the
|
||
|
synthetic NIC is referred to as "netvsc" and the Linux driver for
|
||
|
the synthetic SCSI controller is "storvsc". These drivers contain
|
||
|
functions with names like "storvsc_connect_to_vsp".
|
||
|
|
||
|
VMbus channels
|
||
|
--------------
|
||
|
An instance of a synthetic device uses VMbus channels to communicate
|
||
|
between the VSP and the VSC. Channels are bi-directional and used
|
||
|
for passing messages. Most synthetic devices use a single channel,
|
||
|
but the synthetic SCSI controller and synthetic NIC may use multiple
|
||
|
channels to achieve higher performance and greater parallelism.
|
||
|
|
||
|
Each channel consists of two ring buffers. These are classic ring
|
||
|
buffers from a university data structures textbook. If the read
|
||
|
and writes pointers are equal, the ring buffer is considered to be
|
||
|
empty, so a full ring buffer always has at least one byte unused.
|
||
|
The "in" ring buffer is for messages from the Hyper-V host to the
|
||
|
guest, and the "out" ring buffer is for messages from the guest to
|
||
|
the Hyper-V host. In Linux, the "in" and "out" designations are as
|
||
|
viewed by the guest side. The ring buffers are memory that is
|
||
|
shared between the guest and the host, and they follow the standard
|
||
|
paradigm where the memory is allocated by the guest, with the list
|
||
|
of GPAs that make up the ring buffer communicated to the host. Each
|
||
|
ring buffer consists of a header page (4 Kbytes) with the read and
|
||
|
write indices and some control flags, followed by the memory for the
|
||
|
actual ring. The size of the ring is determined by the VSC in the
|
||
|
guest and is specific to each synthetic device. The list of GPAs
|
||
|
making up the ring is communicated to the Hyper-V host over the
|
||
|
VMbus control path as a GPA Descriptor List (GPADL). See function
|
||
|
vmbus_establish_gpadl().
|
||
|
|
||
|
Each ring buffer is mapped into contiguous Linux kernel virtual
|
||
|
space in three parts: 1) the 4 Kbyte header page, 2) the memory
|
||
|
that makes up the ring itself, and 3) a second mapping of the memory
|
||
|
that makes up the ring itself. Because (2) and (3) are contiguous
|
||
|
in kernel virtual space, the code that copies data to and from the
|
||
|
ring buffer need not be concerned with ring buffer wrap-around.
|
||
|
Once a copy operation has completed, the read or write index may
|
||
|
need to be reset to point back into the first mapping, but the
|
||
|
actual data copy does not need to be broken into two parts. This
|
||
|
approach also allows complex data structures to be easily accessed
|
||
|
directly in the ring without handling wrap-around.
|
||
|
|
||
|
On arm64 with page sizes > 4 Kbytes, the header page must still be
|
||
|
passed to Hyper-V as a 4 Kbyte area. But the memory for the actual
|
||
|
ring must be aligned to PAGE_SIZE and have a size that is a multiple
|
||
|
of PAGE_SIZE so that the duplicate mapping trick can be done. Hence
|
||
|
a portion of the header page is unused and not communicated to
|
||
|
Hyper-V. This case is handled by vmbus_establish_gpadl().
|
||
|
|
||
|
Hyper-V enforces a limit on the aggregate amount of guest memory
|
||
|
that can be shared with the host via GPADLs. This limit ensures
|
||
|
that a rogue guest can't force the consumption of excessive host
|
||
|
resources. For Windows Server 2019 and later, this limit is
|
||
|
approximately 1280 Mbytes. For versions prior to Windows Server
|
||
|
2019, the limit is approximately 384 Mbytes.
|
||
|
|
||
|
VMbus messages
|
||
|
--------------
|
||
|
All VMbus messages have a standard header that includes the message
|
||
|
length, the offset of the message payload, some flags, and a
|
||
|
transactionID. The portion of the message after the header is
|
||
|
unique to each VSP/VSC pair.
|
||
|
|
||
|
Messages follow one of two patterns:
|
||
|
|
||
|
* Unidirectional: Either side sends a message and does not
|
||
|
expect a response message
|
||
|
* Request/response: One side (usually the guest) sends a message
|
||
|
and expects a response
|
||
|
|
||
|
The transactionID (a.k.a. "requestID") is for matching requests &
|
||
|
responses. Some synthetic devices allow multiple requests to be in-
|
||
|
flight simultaneously, so the guest specifies a transactionID when
|
||
|
sending a request. Hyper-V sends back the same transactionID in the
|
||
|
matching response.
|
||
|
|
||
|
Messages passed between the VSP and VSC are control messages. For
|
||
|
example, a message sent from the storvsc driver might be "execute
|
||
|
this SCSI command". If a message also implies some data transfer
|
||
|
between the guest and the Hyper-V host, the actual data to be
|
||
|
transferred may be embedded with the control message, or it may be
|
||
|
specified as a separate data buffer that the Hyper-V host will
|
||
|
access as a DMA operation. The former case is used when the size of
|
||
|
the data is small and the cost of copying the data to and from the
|
||
|
ring buffer is minimal. For example, time sync messages from the
|
||
|
Hyper-V host to the guest contain the actual time value. When the
|
||
|
data is larger, a separate data buffer is used. In this case, the
|
||
|
control message contains a list of GPAs that describe the data
|
||
|
buffer. For example, the storvsc driver uses this approach to
|
||
|
specify the data buffers to/from which disk I/O is done.
|
||
|
|
||
|
Three functions exist to send VMbus messages:
|
||
|
|
||
|
1. vmbus_sendpacket(): Control-only messages and messages with
|
||
|
embedded data -- no GPAs
|
||
|
2. vmbus_sendpacket_pagebuffer(): Message with list of GPAs
|
||
|
identifying data to transfer. An offset and length is
|
||
|
associated with each GPA so that multiple discontinuous areas
|
||
|
of guest memory can be targeted.
|
||
|
3. vmbus_sendpacket_mpb_desc(): Message with list of GPAs
|
||
|
identifying data to transfer. A single offset and length is
|
||
|
associated with a list of GPAs. The GPAs must describe a
|
||
|
single logical area of guest memory to be targeted.
|
||
|
|
||
|
Historically, Linux guests have trusted Hyper-V to send well-formed
|
||
|
and valid messages, and Linux drivers for synthetic devices did not
|
||
|
fully validate messages. With the introduction of processor
|
||
|
technologies that fully encrypt guest memory and that allow the
|
||
|
guest to not trust the hypervisor (AMD SNP-SEV, Intel TDX), trusting
|
||
|
the Hyper-V host is no longer a valid assumption. The drivers for
|
||
|
VMbus synthetic devices are being updated to fully validate any
|
||
|
values read from memory that is shared with Hyper-V, which includes
|
||
|
messages from VMbus devices. To facilitate such validation,
|
||
|
messages read by the guest from the "in" ring buffer are copied to a
|
||
|
temporary buffer that is not shared with Hyper-V. Validation is
|
||
|
performed in this temporary buffer without the risk of Hyper-V
|
||
|
maliciously modifying the message after it is validated but before
|
||
|
it is used.
|
||
|
|
||
|
VMbus interrupts
|
||
|
----------------
|
||
|
VMbus provides a mechanism for the guest to interrupt the host when
|
||
|
the guest has queued new messages in a ring buffer. The host
|
||
|
expects that the guest will send an interrupt only when an "out"
|
||
|
ring buffer transitions from empty to non-empty. If the guest sends
|
||
|
interrupts at other times, the host deems such interrupts to be
|
||
|
unnecessary. If a guest sends an excessive number of unnecessary
|
||
|
interrupts, the host may throttle that guest by suspending its
|
||
|
execution for a few seconds to prevent a denial-of-service attack.
|
||
|
|
||
|
Similarly, the host will interrupt the guest when it sends a new
|
||
|
message on the VMbus control path, or when a VMbus channel "in" ring
|
||
|
buffer transitions from empty to non-empty. Each CPU in the guest
|
||
|
may receive VMbus interrupts, so they are best modeled as per-CPU
|
||
|
interrupts in Linux. This model works well on arm64 where a single
|
||
|
per-CPU IRQ is allocated for VMbus. Since x86/x64 lacks support for
|
||
|
per-CPU IRQs, an x86 interrupt vector is statically allocated (see
|
||
|
HYPERVISOR_CALLBACK_VECTOR) across all CPUs and explicitly coded to
|
||
|
call the VMbus interrupt service routine. These interrupts are
|
||
|
visible in /proc/interrupts on the "HYP" line.
|
||
|
|
||
|
The guest CPU that a VMbus channel will interrupt is selected by the
|
||
|
guest when the channel is created, and the host is informed of that
|
||
|
selection. VMbus devices are broadly grouped into two categories:
|
||
|
|
||
|
1. "Slow" devices that need only one VMbus channel. The devices
|
||
|
(such as keyboard, mouse, heartbeat, and timesync) generate
|
||
|
relatively few interrupts. Their VMbus channels are all
|
||
|
assigned to interrupt the VMBUS_CONNECT_CPU, which is always
|
||
|
CPU 0.
|
||
|
|
||
|
2. "High speed" devices that may use multiple VMbus channels for
|
||
|
higher parallelism and performance. These devices include the
|
||
|
synthetic SCSI controller and synthetic NIC. Their VMbus
|
||
|
channels interrupts are assigned to CPUs that are spread out
|
||
|
among the available CPUs in the VM so that interrupts on
|
||
|
multiple channels can be processed in parallel.
|
||
|
|
||
|
The assignment of VMbus channel interrupts to CPUs is done in the
|
||
|
function init_vp_index(). This assignment is done outside of the
|
||
|
normal Linux interrupt affinity mechanism, so the interrupts are
|
||
|
neither "unmanaged" nor "managed" interrupts.
|
||
|
|
||
|
The CPU that a VMbus channel will interrupt can be seen in
|
||
|
/sys/bus/vmbus/devices/<deviceGUID>/ channels/<channelRelID>/cpu.
|
||
|
When running on later versions of Hyper-V, the CPU can be changed
|
||
|
by writing a new value to this sysfs entry. Because the interrupt
|
||
|
assignment is done outside of the normal Linux affinity mechanism,
|
||
|
there are no entries in /proc/irq corresponding to individual
|
||
|
VMbus channel interrupts.
|
||
|
|
||
|
An online CPU in a Linux guest may not be taken offline if it has
|
||
|
VMbus channel interrupts assigned to it. Any such channel
|
||
|
interrupts must first be manually reassigned to another CPU as
|
||
|
described above. When no channel interrupts are assigned to the
|
||
|
CPU, it can be taken offline.
|
||
|
|
||
|
When a guest CPU receives a VMbus interrupt from the host, the
|
||
|
function vmbus_isr() handles the interrupt. It first checks for
|
||
|
channel interrupts by calling vmbus_chan_sched(), which looks at a
|
||
|
bitmap setup by the host to determine which channels have pending
|
||
|
interrupts on this CPU. If multiple channels have pending
|
||
|
interrupts for this CPU, they are processed sequentially. When all
|
||
|
channel interrupts have been processed, vmbus_isr() checks for and
|
||
|
processes any message received on the VMbus control path.
|
||
|
|
||
|
The VMbus channel interrupt handling code is designed to work
|
||
|
correctly even if an interrupt is received on a CPU other than the
|
||
|
CPU assigned to the channel. Specifically, the code does not use
|
||
|
CPU-based exclusion for correctness. In normal operation, Hyper-V
|
||
|
will interrupt the assigned CPU. But when the CPU assigned to a
|
||
|
channel is being changed via sysfs, the guest doesn't know exactly
|
||
|
when Hyper-V will make the transition. The code must work correctly
|
||
|
even if there is a time lag before Hyper-V starts interrupting the
|
||
|
new CPU. See comments in target_cpu_store().
|
||
|
|
||
|
VMbus device creation/deletion
|
||
|
------------------------------
|
||
|
Hyper-V and the Linux guest have a separate message-passing path
|
||
|
that is used for synthetic device creation and deletion. This
|
||
|
path does not use a VMbus channel. See vmbus_post_msg() and
|
||
|
vmbus_on_msg_dpc().
|
||
|
|
||
|
The first step is for the guest to connect to the generic
|
||
|
Hyper-V VMbus mechanism. As part of establishing this connection,
|
||
|
the guest and Hyper-V agree on a VMbus protocol version they will
|
||
|
use. This negotiation allows newer Linux kernels to run on older
|
||
|
Hyper-V versions, and vice versa.
|
||
|
|
||
|
The guest then tells Hyper-V to "send offers". Hyper-V sends an
|
||
|
offer message to the guest for each synthetic device that the VM
|
||
|
is configured to have. Each VMbus device type has a fixed GUID
|
||
|
known as the "class ID", and each VMbus device instance is also
|
||
|
identified by a GUID. The offer message from Hyper-V contains
|
||
|
both GUIDs to uniquely (within the VM) identify the device.
|
||
|
There is one offer message for each device instance, so a VM with
|
||
|
two synthetic NICs will get two offers messages with the NIC
|
||
|
class ID. The ordering of offer messages can vary from boot-to-boot
|
||
|
and must not be assumed to be consistent in Linux code. Offer
|
||
|
messages may also arrive long after Linux has initially booted
|
||
|
because Hyper-V supports adding devices, such as synthetic NICs,
|
||
|
to running VMs. A new offer message is processed by
|
||
|
vmbus_process_offer(), which indirectly invokes vmbus_add_channel_work().
|
||
|
|
||
|
Upon receipt of an offer message, the guest identifies the device
|
||
|
type based on the class ID, and invokes the correct driver to set up
|
||
|
the device. Driver/device matching is performed using the standard
|
||
|
Linux mechanism.
|
||
|
|
||
|
The device driver probe function opens the primary VMbus channel to
|
||
|
the corresponding VSP. It allocates guest memory for the channel
|
||
|
ring buffers and shares the ring buffer with the Hyper-V host by
|
||
|
giving the host a list of GPAs for the ring buffer memory. See
|
||
|
vmbus_establish_gpadl().
|
||
|
|
||
|
Once the ring buffer is set up, the device driver and VSP exchange
|
||
|
setup messages via the primary channel. These messages may include
|
||
|
negotiating the device protocol version to be used between the Linux
|
||
|
VSC and the VSP on the Hyper-V host. The setup messages may also
|
||
|
include creating additional VMbus channels, which are somewhat
|
||
|
mis-named as "sub-channels" since they are functionally
|
||
|
equivalent to the primary channel once they are created.
|
||
|
|
||
|
Finally, the device driver may create entries in /dev as with
|
||
|
any device driver.
|
||
|
|
||
|
The Hyper-V host can send a "rescind" message to the guest to
|
||
|
remove a device that was previously offered. Linux drivers must
|
||
|
handle such a rescind message at any time. Rescinding a device
|
||
|
invokes the device driver "remove" function to cleanly shut
|
||
|
down the device and remove it. Once a synthetic device is
|
||
|
rescinded, neither Hyper-V nor Linux retains any state about
|
||
|
its previous existence. Such a device might be re-added later,
|
||
|
in which case it is treated as an entirely new device. See
|
||
|
vmbus_onoffer_rescind().
|