From: Matthew Rosato <mjrosato@linux.ibm.com>
To: Douglas Freimuth <freimuth@linux.ibm.com>,
borntraeger@linux.ibm.com, imbrenda@linux.ibm.com,
frankja@linux.ibm.com, david@kernel.org, hca@linux.ibm.com,
gor@linux.ibm.com, agordeev@linux.ibm.com, svens@linux.ibm.com,
kvm@vger.kernel.org, linux-s390@vger.kernel.org,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH v3 1/3] KVM: s390: Add map/unmap ioctl and clean mappings post-guest
Date: Mon, 6 Apr 2026 09:39:35 -0400 [thread overview]
Message-ID: <e2c813a0-0ac0-4a1e-b980-fa2dce33eca7@linux.ibm.com> (raw)
In-Reply-To: <20260406064419.14384-2-freimuth@linux.ibm.com>
On 4/6/26 2:44 AM, Douglas Freimuth wrote:
> S390 needs map/unmap ioctls, which map the adapter set
> indicator pages, so the pages can be accessed when interrupts are
> disabled. The mappings are cleaned up when the guest is removed.
>
> Map/Unmap ioctls are fenced in order to avoid the longterm pinning
> in Secure Execution environments. In Secure Execution
> environments the path of execution available before this patch is followed.
>
> Statistical counters to count map/unmap functions for adapter indicator
> pages are added. The counters can be used to analyze
> map/unmap functions in non-Secure Execution environments and similarly
> can be used to analyze Secure Execution environments where the counters
> will not be incremented as the adapter indicator pages are not mapped.
>
> Signed-off-by: Douglas Freimuth <freimuth@linux.ibm.com>
[...]
> +static int kvm_s390_adapter_unmap(struct kvm *kvm, unsigned int id, __u64 addr)
> +{
> + struct s390_io_adapter *adapter = get_io_adapter(kvm, id);
> + struct s390_map_info *map, *tmp;
> + unsigned long flags;
> + int found = 0, idx;
> +
> + if (!adapter || !addr)
> + return -EINVAL;
> +
> + list_for_each_entry_safe(map, tmp, &adapter->maps, list) {
> + if (map->guest_addr == addr) {
> + spin_lock_irqsave(&adapter->maps_lock, flags);
This lock needs to be acquired before the list_for_each_entry_safe call,
that way it protects the list from changing until after you delete the
specified entry.
> + found = 1;
> + adapter->nr_maps--;
> + list_del(&map->list);
> + spin_unlock_irqrestore(&adapter->maps_lock, flags);
> + idx = srcu_read_lock(&kvm->srcu);
> + mark_page_dirty(kvm, map->addr >> PAGE_SHIFT);
This isn't the right address to mark dirty. Per
898885477e0f KVM: s390: Use guest address to mark guest page dirty
You need to keep track of the gaddr and use that to mark the page dirty.
> + set_page_dirty_lock(map->page);
> + srcu_read_unlock(&kvm->srcu, idx);
> + put_page(map->page);
> + kfree(map);
> + break;
> + }
> + }
> +
> + return found ? 0 : -ENOENT;
> +}
> +
> void kvm_s390_destroy_adapters(struct kvm *kvm)
> {
> int i;
> + struct s390_map_info *map, *tmp;
>
> - for (i = 0; i < MAX_S390_IO_ADAPTERS; i++)
> + for (i = 0; i < MAX_S390_IO_ADAPTERS; i++) {
> + if (!kvm->arch.adapters[i])
> + continue;
Should also be holding the maps_lock over this until the list is emptied
(e.g. right until the kfree(kvm->arch.adapters[i] call)
> + list_for_each_entry_safe(map, tmp,
> + &kvm->arch.adapters[i]->maps, list) {
> + list_del(&map->list);
> + put_page(map->page);
> + kfree(map);
> + }
> kfree(kvm->arch.adapters[i]);
> + }
> }
>
> static int modify_io_adapter(struct kvm_device *dev,
> @@ -2463,7 +2563,8 @@ static int modify_io_adapter(struct kvm_device *dev,
> {
> struct kvm_s390_io_adapter_req req;
> struct s390_io_adapter *adapter;
> - int ret;
> + __u64 host_addr;
> + int ret, idx;
>
> if (copy_from_user(&req, (void __user *)attr->addr, sizeof(req)))
> return -EFAULT;
> @@ -2477,14 +2578,30 @@ static int modify_io_adapter(struct kvm_device *dev,
> if (ret > 0)
> ret = 0;
> break;
> - /*
> - * The following operations are no longer needed and therefore no-ops.
> - * The gpa to hva translation is done when an IRQ route is set up. The
> - * set_irq code uses get_user_pages_remote() to do the actual write.
> - */
> case KVM_S390_IO_ADAPTER_MAP:
> case KVM_S390_IO_ADAPTER_UNMAP:
> - ret = 0;
> + /* If in Secure Execution mode do not long term pin. */
> + mutex_lock(&dev->kvm->lock);
> + if (kvm_s390_pv_is_protected(dev->kvm)) {
> + mutex_unlock(&dev->kvm->lock);
> + return 0;
> + }
> + idx = srcu_read_lock(&dev->kvm->srcu);
> + host_addr = gpa_to_hva(dev->kvm, req.addr);
> + if (kvm_is_error_hva(host_addr)) {
> + srcu_read_unlock(&dev->kvm->srcu, idx);
dev->kvm->lock also needs to be dropped here.
> + return -EFAULT;
> + }
> + srcu_read_unlock(&dev->kvm->srcu, idx);
> + if (req.type == KVM_S390_IO_ADAPTER_MAP) {
> + dev->kvm->stat.io_390_adapter_map++;
> + ret = kvm_s390_adapter_map(dev->kvm, req.id, host_addr);
> + mutex_unlock(&dev->kvm->lock);
This unlock...
> + } else {
> + dev->kvm->stat.io_390_adapter_unmap++;
> + ret = kvm_s390_adapter_unmap(dev->kvm, req.id, host_addr);
> + mutex_unlock(&dev->kvm->lock);
and this unlock...
> + }
Could be combined and moved here
> break;
> default:
> ret = -EINVAL;
> @@ -2730,24 +2847,6 @@ static unsigned long get_ind_bit(__u64 addr, unsigned long bit_nr, bool swap)
> return swap ? (bit ^ (BITS_PER_LONG - 1)) : bit;
> }
>
> -static struct page *get_map_page(struct kvm *kvm, u64 uaddr)
> -{
> - struct mm_struct *mm = kvm->mm;
> - struct page *page = NULL;
> - int locked = 1;
> -
> - if (mmget_not_zero(mm)) {
> - mmap_read_lock(mm);
> - get_user_pages_remote(mm, uaddr, 1, FOLL_WRITE,
> - &page, &locked);
> - if (locked)
> - mmap_read_unlock(mm);
> - mmput(mm);
> - }
> -
> - return page;
> -}
> -
> static int adapter_indicators_set(struct kvm *kvm,
> struct s390_io_adapter *adapter,
> struct kvm_s390_adapter_int *adapter_int)
> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
> index d7838334a338..4eada48c6e27 100644
> --- a/arch/s390/kvm/kvm-s390.c
> +++ b/arch/s390/kvm/kvm-s390.c
> @@ -68,6 +68,8 @@
> const struct kvm_stats_desc kvm_vm_stats_desc[] = {
> KVM_GENERIC_VM_STATS(),
> STATS_DESC_COUNTER(VM, inject_io),
> + STATS_DESC_COUNTER(VM, io_390_adapter_map),
> + STATS_DESC_COUNTER(VM, io_390_adapter_unmap),
> STATS_DESC_COUNTER(VM, inject_float_mchk),
> STATS_DESC_COUNTER(VM, inject_pfault_done),
> STATS_DESC_COUNTER(VM, inject_service_signal),
> @@ -2491,6 +2493,30 @@ static int kvm_s390_pv_dmp(struct kvm *kvm, struct kvm_pv_cmd *cmd,
> return r;
> }
>
> +static void kvm_s390_unmap_all_adapters_pv(struct kvm *kvm)
> +{
> + unsigned long flags;
> + struct s390_map_info *map, *tmp;
> + int i, idx;
> +
> + for (i = 0; i < MAX_S390_IO_ADAPTERS; i++) {
> + if (!kvm->arch.adapters[i])
> + continue;
> + list_for_each_entry_safe(map, tmp,
> + &kvm->arch.adapters[i]->maps, list) {
> + spin_lock_irqsave(&kvm->arch.adapters[i]->maps_lock, flags);
Same comment as kvm_s390_adapter_unmap, need to aquire this before the
list_for_each_entry_safe call.
Actually this one will need to be a bit more creative since you need to
drop the spinlock before each call to set_page_dirty_lock(), but you'll
need to re-acquire it again each time you go back to the list. That
makes list_for_each_entry_safe a bad choice.
Maybe using list_first_entry_or_null() each time you re-acquire the
spinlock, until you get a null (meaning the list is empty)?
> + list_del(&map->list);
You need to be decrementing nr_maps here before droppping the lock;
unlike kvm_s390_destroy_adapters we are not freeing the structure and if
we leave SE mode we could get more mappings later so the nr_maps value
has to be kept up-to-date.
> + spin_unlock_irqrestore(&kvm->arch.adapters[i]->maps_lock, flags);
> + idx = srcu_read_lock(&kvm->srcu);
> + mark_page_dirty(kvm, map->addr >> PAGE_SHIFT);
Same comment as above, need to use the gaddr
> + set_page_dirty_lock(map->page);
> + srcu_read_unlock(&kvm->srcu, idx);
> + put_page(map->page);
> + kfree(map);
> + }
> + }
> +}
> +
> static int kvm_s390_handle_pv(struct kvm *kvm, struct kvm_pv_cmd *cmd)
> {
> const bool need_lock = (cmd->cmd != KVM_PV_ASYNC_CLEANUP_PERFORM);
> @@ -2507,6 +2533,7 @@ static int kvm_s390_handle_pv(struct kvm *kvm, struct kvm_pv_cmd *cmd)
> if (kvm_s390_pv_is_protected(kvm))
> break;
>
> + kvm_s390_unmap_all_adapters_pv(kvm);
> mmap_write_lock(kvm->mm);
> /*
> * Disable creation of new THPs. Existing THPs can stay, they
next prev parent reply other threads:[~2026-04-06 13:39 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-06 6:44 [PATCH v3 0/3] KVM: s390: Introducing kvm_arch_set_irq_inatomic Fast Inject Douglas Freimuth
2026-04-06 6:44 ` [PATCH v3 1/3] KVM: s390: Add map/unmap ioctl and clean mappings post-guest Douglas Freimuth
2026-04-06 13:39 ` Matthew Rosato [this message]
2026-04-06 6:44 ` [PATCH v3 2/3] KVM: s390: Enable adapter_indicators_set to use mapped pages Douglas Freimuth
2026-04-06 15:33 ` Matthew Rosato
2026-04-06 6:44 ` [PATCH v3 3/3] KVM: s390: Introducing kvm_arch_set_irq_inatomic fast inject Douglas Freimuth
2026-04-06 16:15 ` Matthew Rosato
2026-04-06 17:59 ` Sean Christopherson
2026-04-06 18:24 ` Matthew Rosato
2026-04-06 16:25 ` Matthew Rosato
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=e2c813a0-0ac0-4a1e-b980-fa2dce33eca7@linux.ibm.com \
--to=mjrosato@linux.ibm.com \
--cc=agordeev@linux.ibm.com \
--cc=borntraeger@linux.ibm.com \
--cc=david@kernel.org \
--cc=frankja@linux.ibm.com \
--cc=freimuth@linux.ibm.com \
--cc=gor@linux.ibm.com \
--cc=hca@linux.ibm.com \
--cc=imbrenda@linux.ibm.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-s390@vger.kernel.org \
--cc=svens@linux.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox