[Hypervisor Live Update] Notes from November 3, 2025

David Rientjes rientjes at google.com
Sat Nov 8 15:48:32 PST 2025


Hi everybody,

Here are the notes from the last Hypervisor Live Update call that happened 
on Monday, November 3.  Thanks to everybody who was involved!

These notes are intended to bring people up to speed who could not attend 
the call as well as keep the conversation going in between meetings.

----->o-----
We chatted briefly about the status of the stateless KHO RFC patch series 
intended to simply LUO support.

Pasha started us off by updating that Jason Miu will be updating his patch 
series and is expected to send those patches to the mailing list after 
some more internal code review.  That would be expected to be posted 
either this week or next week.

Pasha also updated that is plan is to remove subsystems to simply the 
state machine and the UAPI; this would be replaced by the file lifecycle 
bound global data created and destroyed automatically based on reserved 
file state.  He was also simplifying the state machine to keep the minimum 
needed support in the initial version, but with extensibility for the 
future.  No exact timeline although it is currently ~90% ready.

----->o-----
Pratyush discussed the global file based state and how it may circumvent 
the LUO state machine; it's an implicit preserve of the subsystem 
completely independent of global state.  Pasha said the state is now bound 
to the state of the preserved files only based on sessions.  We're getting 
rid of global state for now.

Pratyush suggested to tie subsystems with the file handler but this would 
not be possible if subsystems are going away.  Pasha said the new global 
state is flexible and can share multiple file handlers; one global state 
can be bound to multiple file handlers.

----->o-----
Jork Loeser as if there is a design/API link for the memfd and whether 
this is something a driver could use to persistently holding data.  He was 
asking if a driver could associate arbitrary pages with a preserved memfd.

Pratyush said the memfd preservation was part of the LUO patch series at 
the end.  A driver can pass a memfd to LUO after creating an fd.  Pratyush 
suggested using KHO to preserve data; the data may moved at runtime but 
would need to be unmovable during preservation across kexec.

Jason Gunthorpe suggested using APIs to get a page from a memfd at a 
specific offset.

----->o-----
Vipin Sharma had posted a recent patch series for VFIO[1], David Matlack 
will be working on v2 of this will Vipin is on leave.  Feedback was 
received about not moving the PCI save state and making them public, so 
that's work in progress.  More feedback said there was missing bits and we 
need more PCI core changes that would be updated in v2 to be more complete 
(but also include more PCI changes).  No specific timeline yet on v2, but 
it will be based on LUO v5.

David said the VFIO patches are using an anonymous inode to recreate the 
file after live update and asked if we care about associating recreated 
fds for userspace after live update with a particular inode.  Jason said 
that VFIO would care because it uses the inode to get an address space 
which it uses with an unmapped mapping range and this must work correctly.

----->o-----
Sami summarized the discussion on IOMMU persistence.  He was working on 
updating the patch series to v2 based on the feedback from v1.  He talked 
about restoration of the HWPTs on the restore side.  Jason thought that we 
wouldn't have an IOAS for the restored domains and suggested it could be 
null instead.  Sami thought this may be slightly invasive including where 
we are taking locks; Jason suggested against a dummy IOAS.

----->o-----
We discussed briefly about deferred struct page initialization support 
with KHO.  Pasha said KHO isn't compatible with deferred struct pages 
although when using KHO we definitely want fast reboot performance.  We 
decided to discuss this more later after LPC where there will be some 
discussion about reboot performance.

----->o-----
Pratyush noted that he is working on the 1GB preservation but will take 
some more time to clean up and have it working properly.  He said 
guest_memfd would use hugetlb for 1GB pages so he's working on hugetlb 
preservation.  Pratyush was focusing on generic hugetlb support that could 
be ported for use with guest_memfd when it supports hugetlb.  He's aiming 
for an RFC to be ready by the time of LPC.

Ackerley updated that the hugetlb support for guest_memfd is currently in 
RFC patches posted upstream.

----->o-----
Next meeting will be on Monday, November 17 at 8am PST (UTC-8), everybody
is welcome: https://meet.google.com/rjn-dmzu-hgq

Topics for the next meeting:

 - update on the status of stateless KHO RFC patches that should simplify
   LUO support
 - update on the status of LUO v5 overall
 - follow up on the status of iommu persistence and its v2 patches based
   on v1 feedback
 - update on the v2 of the VFIO patch series based on LUO v5 and expected
   timelines
 - discuss status of hugetlb preservation, specifically 1GB support, with
   regular memfd, aiming for an RFC by the time of LPC
 - update on status of guest_memfd support for 1GB hugetlb pages
 - discuss any use cases for Confidential Computing where folios may need
   to be split after being marked as preserved during brown out
 - later: testing methodology to allow downstream consumers to qualify
   that live update works from one version to another
 - later: reducing blackout window during live update, including deferred
   struct page initialization

Please let me know if you'd like to propose additional topics for
discussion, thank you!

[1] https://marc.info/?l=linux-kernel&m=176074589311102&w=2 



More information about the kexec mailing list