1a28b0fc0SDavid Woodhouse /* 2a28b0fc0SDavid Woodhouse * QEMU Xen emulation: Grant table support 3a28b0fc0SDavid Woodhouse * 4a28b0fc0SDavid Woodhouse * Copyright © 2022 Amazon.com, Inc. or its affiliates. All Rights Reserved. 5a28b0fc0SDavid Woodhouse * 6a28b0fc0SDavid Woodhouse * Authors: David Woodhouse <dwmw2@infradead.org> 7a28b0fc0SDavid Woodhouse * 8a28b0fc0SDavid Woodhouse * This work is licensed under the terms of the GNU GPL, version 2 or later. 9a28b0fc0SDavid Woodhouse * See the COPYING file in the top-level directory. 10a28b0fc0SDavid Woodhouse */ 11a28b0fc0SDavid Woodhouse 12a28b0fc0SDavid Woodhouse #include "qemu/osdep.h" 13a28b0fc0SDavid Woodhouse #include "qemu/host-utils.h" 14a28b0fc0SDavid Woodhouse #include "qemu/module.h" 15a28b0fc0SDavid Woodhouse #include "qemu/lockable.h" 16a28b0fc0SDavid Woodhouse #include "qemu/main-loop.h" 17a28b0fc0SDavid Woodhouse #include "qapi/error.h" 18a28b0fc0SDavid Woodhouse #include "qom/object.h" 19a28b0fc0SDavid Woodhouse #include "exec/target_page.h" 20a28b0fc0SDavid Woodhouse #include "exec/address-spaces.h" 21a28b0fc0SDavid Woodhouse #include "migration/vmstate.h" 22a28b0fc0SDavid Woodhouse 23a28b0fc0SDavid Woodhouse #include "hw/sysbus.h" 24a28b0fc0SDavid Woodhouse #include "hw/xen/xen.h" 25*b08d88e3SDavid Woodhouse #include "hw/xen/xen_backend_ops.h" 26a28b0fc0SDavid Woodhouse #include "xen_overlay.h" 27a28b0fc0SDavid Woodhouse #include "xen_gnttab.h" 28a28b0fc0SDavid Woodhouse 29a28b0fc0SDavid Woodhouse #include "sysemu/kvm.h" 30a28b0fc0SDavid Woodhouse #include "sysemu/kvm_xen.h" 31a28b0fc0SDavid Woodhouse 32a28b0fc0SDavid Woodhouse #include "hw/xen/interface/memory.h" 33a28b0fc0SDavid Woodhouse #include "hw/xen/interface/grant_table.h" 34a28b0fc0SDavid Woodhouse 35a28b0fc0SDavid Woodhouse #define TYPE_XEN_GNTTAB "xen-gnttab" 36a28b0fc0SDavid Woodhouse OBJECT_DECLARE_SIMPLE_TYPE(XenGnttabState, XEN_GNTTAB) 37a28b0fc0SDavid Woodhouse 38e33cb789SDavid Woodhouse #define ENTRIES_PER_FRAME_V1 (XEN_PAGE_SIZE / sizeof(grant_entry_v1_t)) 39e33cb789SDavid Woodhouse 40*b08d88e3SDavid Woodhouse static struct gnttab_backend_ops emu_gnttab_backend_ops; 41*b08d88e3SDavid Woodhouse 42a28b0fc0SDavid Woodhouse struct XenGnttabState { 43a28b0fc0SDavid Woodhouse /*< private >*/ 44a28b0fc0SDavid Woodhouse SysBusDevice busdev; 45a28b0fc0SDavid Woodhouse /*< public >*/ 46a28b0fc0SDavid Woodhouse 47e33cb789SDavid Woodhouse QemuMutex gnt_lock; 48e33cb789SDavid Woodhouse 49a28b0fc0SDavid Woodhouse uint32_t nr_frames; 50a28b0fc0SDavid Woodhouse uint32_t max_frames; 51e33cb789SDavid Woodhouse 52e33cb789SDavid Woodhouse union { 53e33cb789SDavid Woodhouse grant_entry_v1_t *v1; 54e33cb789SDavid Woodhouse /* Theoretically, v2 support could be added here. */ 55e33cb789SDavid Woodhouse } entries; 56e33cb789SDavid Woodhouse 57e33cb789SDavid Woodhouse MemoryRegion gnt_frames; 58e33cb789SDavid Woodhouse MemoryRegion *gnt_aliases; 59e33cb789SDavid Woodhouse uint64_t *gnt_frame_gpas; 60*b08d88e3SDavid Woodhouse 61*b08d88e3SDavid Woodhouse uint8_t *map_track; 62a28b0fc0SDavid Woodhouse }; 63a28b0fc0SDavid Woodhouse 64a28b0fc0SDavid Woodhouse struct XenGnttabState *xen_gnttab_singleton; 65a28b0fc0SDavid Woodhouse 66a28b0fc0SDavid Woodhouse static void xen_gnttab_realize(DeviceState *dev, Error **errp) 67a28b0fc0SDavid Woodhouse { 68a28b0fc0SDavid Woodhouse XenGnttabState *s = XEN_GNTTAB(dev); 69e33cb789SDavid Woodhouse int i; 70a28b0fc0SDavid Woodhouse 71a28b0fc0SDavid Woodhouse if (xen_mode != XEN_EMULATE) { 72a28b0fc0SDavid Woodhouse error_setg(errp, "Xen grant table support is for Xen emulation"); 73a28b0fc0SDavid Woodhouse return; 74a28b0fc0SDavid Woodhouse } 75a28b0fc0SDavid Woodhouse s->nr_frames = 0; 76a28b0fc0SDavid Woodhouse s->max_frames = kvm_xen_get_gnttab_max_frames(); 77e33cb789SDavid Woodhouse memory_region_init_ram(&s->gnt_frames, OBJECT(dev), "xen:grant_table", 78e33cb789SDavid Woodhouse XEN_PAGE_SIZE * s->max_frames, &error_abort); 79e33cb789SDavid Woodhouse memory_region_set_enabled(&s->gnt_frames, true); 80e33cb789SDavid Woodhouse s->entries.v1 = memory_region_get_ram_ptr(&s->gnt_frames); 81e33cb789SDavid Woodhouse memset(s->entries.v1, 0, XEN_PAGE_SIZE * s->max_frames); 82e33cb789SDavid Woodhouse 83e33cb789SDavid Woodhouse /* Create individual page-sizes aliases for overlays */ 84e33cb789SDavid Woodhouse s->gnt_aliases = (void *)g_new0(MemoryRegion, s->max_frames); 85e33cb789SDavid Woodhouse s->gnt_frame_gpas = (void *)g_new(uint64_t, s->max_frames); 86e33cb789SDavid Woodhouse for (i = 0; i < s->max_frames; i++) { 87e33cb789SDavid Woodhouse memory_region_init_alias(&s->gnt_aliases[i], OBJECT(dev), 88e33cb789SDavid Woodhouse NULL, &s->gnt_frames, 89e33cb789SDavid Woodhouse i * XEN_PAGE_SIZE, XEN_PAGE_SIZE); 90e33cb789SDavid Woodhouse s->gnt_frame_gpas[i] = INVALID_GPA; 91e33cb789SDavid Woodhouse } 92e33cb789SDavid Woodhouse 93*b08d88e3SDavid Woodhouse s->entries.v1[GNTTAB_RESERVED_XENSTORE].flags = GTF_permit_access; 94*b08d88e3SDavid Woodhouse s->entries.v1[GNTTAB_RESERVED_XENSTORE].frame = XEN_SPECIAL_PFN(XENSTORE); 95e33cb789SDavid Woodhouse qemu_mutex_init(&s->gnt_lock); 96e33cb789SDavid Woodhouse 97e33cb789SDavid Woodhouse xen_gnttab_singleton = s; 98*b08d88e3SDavid Woodhouse 99*b08d88e3SDavid Woodhouse s->map_track = g_new0(uint8_t, s->max_frames * ENTRIES_PER_FRAME_V1); 100*b08d88e3SDavid Woodhouse 101*b08d88e3SDavid Woodhouse xen_gnttab_ops = &emu_gnttab_backend_ops; 102e33cb789SDavid Woodhouse } 103e33cb789SDavid Woodhouse 104e33cb789SDavid Woodhouse static int xen_gnttab_post_load(void *opaque, int version_id) 105e33cb789SDavid Woodhouse { 106e33cb789SDavid Woodhouse XenGnttabState *s = XEN_GNTTAB(opaque); 107e33cb789SDavid Woodhouse uint32_t i; 108e33cb789SDavid Woodhouse 109e33cb789SDavid Woodhouse for (i = 0; i < s->nr_frames; i++) { 110e33cb789SDavid Woodhouse if (s->gnt_frame_gpas[i] != INVALID_GPA) { 111e33cb789SDavid Woodhouse xen_overlay_do_map_page(&s->gnt_aliases[i], s->gnt_frame_gpas[i]); 112e33cb789SDavid Woodhouse } 113e33cb789SDavid Woodhouse } 114e33cb789SDavid Woodhouse return 0; 115a28b0fc0SDavid Woodhouse } 116a28b0fc0SDavid Woodhouse 117a28b0fc0SDavid Woodhouse static bool xen_gnttab_is_needed(void *opaque) 118a28b0fc0SDavid Woodhouse { 119a28b0fc0SDavid Woodhouse return xen_mode == XEN_EMULATE; 120a28b0fc0SDavid Woodhouse } 121a28b0fc0SDavid Woodhouse 122a28b0fc0SDavid Woodhouse static const VMStateDescription xen_gnttab_vmstate = { 123a28b0fc0SDavid Woodhouse .name = "xen_gnttab", 124a28b0fc0SDavid Woodhouse .version_id = 1, 125a28b0fc0SDavid Woodhouse .minimum_version_id = 1, 126a28b0fc0SDavid Woodhouse .needed = xen_gnttab_is_needed, 127e33cb789SDavid Woodhouse .post_load = xen_gnttab_post_load, 128a28b0fc0SDavid Woodhouse .fields = (VMStateField[]) { 129a28b0fc0SDavid Woodhouse VMSTATE_UINT32(nr_frames, XenGnttabState), 130e33cb789SDavid Woodhouse VMSTATE_VARRAY_UINT32(gnt_frame_gpas, XenGnttabState, nr_frames, 0, 131e33cb789SDavid Woodhouse vmstate_info_uint64, uint64_t), 132a28b0fc0SDavid Woodhouse VMSTATE_END_OF_LIST() 133a28b0fc0SDavid Woodhouse } 134a28b0fc0SDavid Woodhouse }; 135a28b0fc0SDavid Woodhouse 136a28b0fc0SDavid Woodhouse static void xen_gnttab_class_init(ObjectClass *klass, void *data) 137a28b0fc0SDavid Woodhouse { 138a28b0fc0SDavid Woodhouse DeviceClass *dc = DEVICE_CLASS(klass); 139a28b0fc0SDavid Woodhouse 140a28b0fc0SDavid Woodhouse dc->realize = xen_gnttab_realize; 141a28b0fc0SDavid Woodhouse dc->vmsd = &xen_gnttab_vmstate; 142a28b0fc0SDavid Woodhouse } 143a28b0fc0SDavid Woodhouse 144a28b0fc0SDavid Woodhouse static const TypeInfo xen_gnttab_info = { 145a28b0fc0SDavid Woodhouse .name = TYPE_XEN_GNTTAB, 146a28b0fc0SDavid Woodhouse .parent = TYPE_SYS_BUS_DEVICE, 147a28b0fc0SDavid Woodhouse .instance_size = sizeof(XenGnttabState), 148a28b0fc0SDavid Woodhouse .class_init = xen_gnttab_class_init, 149a28b0fc0SDavid Woodhouse }; 150a28b0fc0SDavid Woodhouse 151a28b0fc0SDavid Woodhouse void xen_gnttab_create(void) 152a28b0fc0SDavid Woodhouse { 153a28b0fc0SDavid Woodhouse xen_gnttab_singleton = XEN_GNTTAB(sysbus_create_simple(TYPE_XEN_GNTTAB, 154a28b0fc0SDavid Woodhouse -1, NULL)); 155a28b0fc0SDavid Woodhouse } 156a28b0fc0SDavid Woodhouse 157a28b0fc0SDavid Woodhouse static void xen_gnttab_register_types(void) 158a28b0fc0SDavid Woodhouse { 159a28b0fc0SDavid Woodhouse type_register_static(&xen_gnttab_info); 160a28b0fc0SDavid Woodhouse } 161a28b0fc0SDavid Woodhouse 162a28b0fc0SDavid Woodhouse type_init(xen_gnttab_register_types) 163a28b0fc0SDavid Woodhouse 164a28b0fc0SDavid Woodhouse int xen_gnttab_map_page(uint64_t idx, uint64_t gfn) 165a28b0fc0SDavid Woodhouse { 166e33cb789SDavid Woodhouse XenGnttabState *s = xen_gnttab_singleton; 167e33cb789SDavid Woodhouse uint64_t gpa = gfn << XEN_PAGE_SHIFT; 168e33cb789SDavid Woodhouse 169e33cb789SDavid Woodhouse if (!s) { 170e33cb789SDavid Woodhouse return -ENOTSUP; 171e33cb789SDavid Woodhouse } 172e33cb789SDavid Woodhouse 173e33cb789SDavid Woodhouse if (idx >= s->max_frames) { 174e33cb789SDavid Woodhouse return -EINVAL; 175e33cb789SDavid Woodhouse } 176e33cb789SDavid Woodhouse 177e33cb789SDavid Woodhouse QEMU_IOTHREAD_LOCK_GUARD(); 178e33cb789SDavid Woodhouse QEMU_LOCK_GUARD(&s->gnt_lock); 179e33cb789SDavid Woodhouse 180e33cb789SDavid Woodhouse xen_overlay_do_map_page(&s->gnt_aliases[idx], gpa); 181e33cb789SDavid Woodhouse 182e33cb789SDavid Woodhouse s->gnt_frame_gpas[idx] = gpa; 183e33cb789SDavid Woodhouse 184e33cb789SDavid Woodhouse if (s->nr_frames <= idx) { 185e33cb789SDavid Woodhouse s->nr_frames = idx + 1; 186e33cb789SDavid Woodhouse } 187e33cb789SDavid Woodhouse 188e33cb789SDavid Woodhouse return 0; 189a28b0fc0SDavid Woodhouse } 190a28b0fc0SDavid Woodhouse 19128b7ae94SDavid Woodhouse int xen_gnttab_set_version_op(struct gnttab_set_version *set) 19228b7ae94SDavid Woodhouse { 19328b7ae94SDavid Woodhouse int ret; 19428b7ae94SDavid Woodhouse 19528b7ae94SDavid Woodhouse switch (set->version) { 19628b7ae94SDavid Woodhouse case 1: 19728b7ae94SDavid Woodhouse ret = 0; 19828b7ae94SDavid Woodhouse break; 19928b7ae94SDavid Woodhouse 20028b7ae94SDavid Woodhouse case 2: 20128b7ae94SDavid Woodhouse /* Behave as before set_version was introduced. */ 20228b7ae94SDavid Woodhouse ret = -ENOSYS; 20328b7ae94SDavid Woodhouse break; 20428b7ae94SDavid Woodhouse 20528b7ae94SDavid Woodhouse default: 20628b7ae94SDavid Woodhouse ret = -EINVAL; 20728b7ae94SDavid Woodhouse } 20828b7ae94SDavid Woodhouse 20928b7ae94SDavid Woodhouse set->version = 1; 21028b7ae94SDavid Woodhouse return ret; 21128b7ae94SDavid Woodhouse } 21228b7ae94SDavid Woodhouse 21328b7ae94SDavid Woodhouse int xen_gnttab_get_version_op(struct gnttab_get_version *get) 21428b7ae94SDavid Woodhouse { 21528b7ae94SDavid Woodhouse if (get->dom != DOMID_SELF && get->dom != xen_domid) { 21628b7ae94SDavid Woodhouse return -ESRCH; 21728b7ae94SDavid Woodhouse } 21828b7ae94SDavid Woodhouse 21928b7ae94SDavid Woodhouse get->version = 1; 22028b7ae94SDavid Woodhouse return 0; 22128b7ae94SDavid Woodhouse } 222b46f9745SDavid Woodhouse 223b46f9745SDavid Woodhouse int xen_gnttab_query_size_op(struct gnttab_query_size *size) 224b46f9745SDavid Woodhouse { 225b46f9745SDavid Woodhouse XenGnttabState *s = xen_gnttab_singleton; 226b46f9745SDavid Woodhouse 227b46f9745SDavid Woodhouse if (!s) { 228b46f9745SDavid Woodhouse return -ENOTSUP; 229b46f9745SDavid Woodhouse } 230b46f9745SDavid Woodhouse 231b46f9745SDavid Woodhouse if (size->dom != DOMID_SELF && size->dom != xen_domid) { 232b46f9745SDavid Woodhouse size->status = GNTST_bad_domain; 233b46f9745SDavid Woodhouse return 0; 234b46f9745SDavid Woodhouse } 235b46f9745SDavid Woodhouse 236b46f9745SDavid Woodhouse size->status = GNTST_okay; 237b46f9745SDavid Woodhouse size->nr_frames = s->nr_frames; 238b46f9745SDavid Woodhouse size->max_nr_frames = s->max_frames; 239b46f9745SDavid Woodhouse return 0; 240b46f9745SDavid Woodhouse } 241*b08d88e3SDavid Woodhouse 242*b08d88e3SDavid Woodhouse /* Track per-open refs, to allow close() to clean up. */ 243*b08d88e3SDavid Woodhouse struct active_ref { 244*b08d88e3SDavid Woodhouse MemoryRegionSection mrs; 245*b08d88e3SDavid Woodhouse void *virtaddr; 246*b08d88e3SDavid Woodhouse uint32_t refcnt; 247*b08d88e3SDavid Woodhouse int prot; 248*b08d88e3SDavid Woodhouse }; 249*b08d88e3SDavid Woodhouse 250*b08d88e3SDavid Woodhouse static void gnt_unref(XenGnttabState *s, grant_ref_t ref, 251*b08d88e3SDavid Woodhouse MemoryRegionSection *mrs, int prot) 252*b08d88e3SDavid Woodhouse { 253*b08d88e3SDavid Woodhouse if (mrs && mrs->mr) { 254*b08d88e3SDavid Woodhouse if (prot & PROT_WRITE) { 255*b08d88e3SDavid Woodhouse memory_region_set_dirty(mrs->mr, mrs->offset_within_region, 256*b08d88e3SDavid Woodhouse XEN_PAGE_SIZE); 257*b08d88e3SDavid Woodhouse } 258*b08d88e3SDavid Woodhouse memory_region_unref(mrs->mr); 259*b08d88e3SDavid Woodhouse mrs->mr = NULL; 260*b08d88e3SDavid Woodhouse } 261*b08d88e3SDavid Woodhouse assert(s->map_track[ref] != 0); 262*b08d88e3SDavid Woodhouse 263*b08d88e3SDavid Woodhouse if (--s->map_track[ref] == 0) { 264*b08d88e3SDavid Woodhouse grant_entry_v1_t *gnt_p = &s->entries.v1[ref]; 265*b08d88e3SDavid Woodhouse qatomic_and(&gnt_p->flags, (uint16_t)~(GTF_reading | GTF_writing)); 266*b08d88e3SDavid Woodhouse } 267*b08d88e3SDavid Woodhouse } 268*b08d88e3SDavid Woodhouse 269*b08d88e3SDavid Woodhouse static uint64_t gnt_ref(XenGnttabState *s, grant_ref_t ref, int prot) 270*b08d88e3SDavid Woodhouse { 271*b08d88e3SDavid Woodhouse uint16_t mask = GTF_type_mask | GTF_sub_page; 272*b08d88e3SDavid Woodhouse grant_entry_v1_t gnt, *gnt_p; 273*b08d88e3SDavid Woodhouse int retries = 0; 274*b08d88e3SDavid Woodhouse 275*b08d88e3SDavid Woodhouse if (ref >= s->max_frames * ENTRIES_PER_FRAME_V1 || 276*b08d88e3SDavid Woodhouse s->map_track[ref] == UINT8_MAX) { 277*b08d88e3SDavid Woodhouse return INVALID_GPA; 278*b08d88e3SDavid Woodhouse } 279*b08d88e3SDavid Woodhouse 280*b08d88e3SDavid Woodhouse if (prot & PROT_WRITE) { 281*b08d88e3SDavid Woodhouse mask |= GTF_readonly; 282*b08d88e3SDavid Woodhouse } 283*b08d88e3SDavid Woodhouse 284*b08d88e3SDavid Woodhouse gnt_p = &s->entries.v1[ref]; 285*b08d88e3SDavid Woodhouse 286*b08d88e3SDavid Woodhouse /* 287*b08d88e3SDavid Woodhouse * The guest can legitimately be changing the GTF_readonly flag. Allow 288*b08d88e3SDavid Woodhouse * that, but don't let a malicious guest cause a livelock. 289*b08d88e3SDavid Woodhouse */ 290*b08d88e3SDavid Woodhouse for (retries = 0; retries < 5; retries++) { 291*b08d88e3SDavid Woodhouse uint16_t new_flags; 292*b08d88e3SDavid Woodhouse 293*b08d88e3SDavid Woodhouse /* Read the entry before an atomic operation on its flags */ 294*b08d88e3SDavid Woodhouse gnt = *(volatile grant_entry_v1_t *)gnt_p; 295*b08d88e3SDavid Woodhouse 296*b08d88e3SDavid Woodhouse if ((gnt.flags & mask) != GTF_permit_access || 297*b08d88e3SDavid Woodhouse gnt.domid != DOMID_QEMU) { 298*b08d88e3SDavid Woodhouse return INVALID_GPA; 299*b08d88e3SDavid Woodhouse } 300*b08d88e3SDavid Woodhouse 301*b08d88e3SDavid Woodhouse new_flags = gnt.flags | GTF_reading; 302*b08d88e3SDavid Woodhouse if (prot & PROT_WRITE) { 303*b08d88e3SDavid Woodhouse new_flags |= GTF_writing; 304*b08d88e3SDavid Woodhouse } 305*b08d88e3SDavid Woodhouse 306*b08d88e3SDavid Woodhouse if (qatomic_cmpxchg(&gnt_p->flags, gnt.flags, new_flags) == gnt.flags) { 307*b08d88e3SDavid Woodhouse return (uint64_t)gnt.frame << XEN_PAGE_SHIFT; 308*b08d88e3SDavid Woodhouse } 309*b08d88e3SDavid Woodhouse } 310*b08d88e3SDavid Woodhouse 311*b08d88e3SDavid Woodhouse return INVALID_GPA; 312*b08d88e3SDavid Woodhouse } 313*b08d88e3SDavid Woodhouse 314*b08d88e3SDavid Woodhouse struct xengntdev_handle { 315*b08d88e3SDavid Woodhouse GHashTable *active_maps; 316*b08d88e3SDavid Woodhouse }; 317*b08d88e3SDavid Woodhouse 318*b08d88e3SDavid Woodhouse static int xen_be_gnttab_set_max_grants(struct xengntdev_handle *xgt, 319*b08d88e3SDavid Woodhouse uint32_t nr_grants) 320*b08d88e3SDavid Woodhouse { 321*b08d88e3SDavid Woodhouse return 0; 322*b08d88e3SDavid Woodhouse } 323*b08d88e3SDavid Woodhouse 324*b08d88e3SDavid Woodhouse static void *xen_be_gnttab_map_refs(struct xengntdev_handle *xgt, 325*b08d88e3SDavid Woodhouse uint32_t count, uint32_t domid, 326*b08d88e3SDavid Woodhouse uint32_t *refs, int prot) 327*b08d88e3SDavid Woodhouse { 328*b08d88e3SDavid Woodhouse XenGnttabState *s = xen_gnttab_singleton; 329*b08d88e3SDavid Woodhouse struct active_ref *act; 330*b08d88e3SDavid Woodhouse 331*b08d88e3SDavid Woodhouse if (!s) { 332*b08d88e3SDavid Woodhouse errno = ENOTSUP; 333*b08d88e3SDavid Woodhouse return NULL; 334*b08d88e3SDavid Woodhouse } 335*b08d88e3SDavid Woodhouse 336*b08d88e3SDavid Woodhouse if (domid != xen_domid) { 337*b08d88e3SDavid Woodhouse errno = EINVAL; 338*b08d88e3SDavid Woodhouse return NULL; 339*b08d88e3SDavid Woodhouse } 340*b08d88e3SDavid Woodhouse 341*b08d88e3SDavid Woodhouse if (!count || count > 4096) { 342*b08d88e3SDavid Woodhouse errno = EINVAL; 343*b08d88e3SDavid Woodhouse return NULL; 344*b08d88e3SDavid Woodhouse } 345*b08d88e3SDavid Woodhouse 346*b08d88e3SDavid Woodhouse /* 347*b08d88e3SDavid Woodhouse * Making a contiguous mapping from potentially discontiguous grant 348*b08d88e3SDavid Woodhouse * references would be... distinctly non-trivial. We don't support it. 349*b08d88e3SDavid Woodhouse * Even changing the API to return an array of pointers, one per page, 350*b08d88e3SDavid Woodhouse * wouldn't be simple to use in PV backends because some structures 351*b08d88e3SDavid Woodhouse * actually cross page boundaries (e.g. 32-bit blkif_response ring 352*b08d88e3SDavid Woodhouse * entries are 12 bytes). 353*b08d88e3SDavid Woodhouse */ 354*b08d88e3SDavid Woodhouse if (count != 1) { 355*b08d88e3SDavid Woodhouse errno = EINVAL; 356*b08d88e3SDavid Woodhouse return NULL; 357*b08d88e3SDavid Woodhouse } 358*b08d88e3SDavid Woodhouse 359*b08d88e3SDavid Woodhouse QEMU_LOCK_GUARD(&s->gnt_lock); 360*b08d88e3SDavid Woodhouse 361*b08d88e3SDavid Woodhouse act = g_hash_table_lookup(xgt->active_maps, GINT_TO_POINTER(refs[0])); 362*b08d88e3SDavid Woodhouse if (act) { 363*b08d88e3SDavid Woodhouse if ((prot & PROT_WRITE) && !(act->prot & PROT_WRITE)) { 364*b08d88e3SDavid Woodhouse if (gnt_ref(s, refs[0], prot) == INVALID_GPA) { 365*b08d88e3SDavid Woodhouse return NULL; 366*b08d88e3SDavid Woodhouse } 367*b08d88e3SDavid Woodhouse act->prot |= PROT_WRITE; 368*b08d88e3SDavid Woodhouse } 369*b08d88e3SDavid Woodhouse act->refcnt++; 370*b08d88e3SDavid Woodhouse } else { 371*b08d88e3SDavid Woodhouse uint64_t gpa = gnt_ref(s, refs[0], prot); 372*b08d88e3SDavid Woodhouse if (gpa == INVALID_GPA) { 373*b08d88e3SDavid Woodhouse errno = EINVAL; 374*b08d88e3SDavid Woodhouse return NULL; 375*b08d88e3SDavid Woodhouse } 376*b08d88e3SDavid Woodhouse 377*b08d88e3SDavid Woodhouse act = g_new0(struct active_ref, 1); 378*b08d88e3SDavid Woodhouse act->prot = prot; 379*b08d88e3SDavid Woodhouse act->refcnt = 1; 380*b08d88e3SDavid Woodhouse act->mrs = memory_region_find(get_system_memory(), gpa, XEN_PAGE_SIZE); 381*b08d88e3SDavid Woodhouse 382*b08d88e3SDavid Woodhouse if (act->mrs.mr && 383*b08d88e3SDavid Woodhouse !int128_lt(act->mrs.size, int128_make64(XEN_PAGE_SIZE)) && 384*b08d88e3SDavid Woodhouse memory_region_get_ram_addr(act->mrs.mr) != RAM_ADDR_INVALID) { 385*b08d88e3SDavid Woodhouse act->virtaddr = qemu_map_ram_ptr(act->mrs.mr->ram_block, 386*b08d88e3SDavid Woodhouse act->mrs.offset_within_region); 387*b08d88e3SDavid Woodhouse } 388*b08d88e3SDavid Woodhouse if (!act->virtaddr) { 389*b08d88e3SDavid Woodhouse gnt_unref(s, refs[0], &act->mrs, 0); 390*b08d88e3SDavid Woodhouse g_free(act); 391*b08d88e3SDavid Woodhouse errno = EINVAL; 392*b08d88e3SDavid Woodhouse return NULL; 393*b08d88e3SDavid Woodhouse } 394*b08d88e3SDavid Woodhouse 395*b08d88e3SDavid Woodhouse s->map_track[refs[0]]++; 396*b08d88e3SDavid Woodhouse g_hash_table_insert(xgt->active_maps, GINT_TO_POINTER(refs[0]), act); 397*b08d88e3SDavid Woodhouse } 398*b08d88e3SDavid Woodhouse 399*b08d88e3SDavid Woodhouse return act->virtaddr; 400*b08d88e3SDavid Woodhouse } 401*b08d88e3SDavid Woodhouse 402*b08d88e3SDavid Woodhouse static gboolean do_unmap(gpointer key, gpointer value, gpointer user_data) 403*b08d88e3SDavid Woodhouse { 404*b08d88e3SDavid Woodhouse XenGnttabState *s = user_data; 405*b08d88e3SDavid Woodhouse grant_ref_t gref = GPOINTER_TO_INT(key); 406*b08d88e3SDavid Woodhouse struct active_ref *act = value; 407*b08d88e3SDavid Woodhouse 408*b08d88e3SDavid Woodhouse gnt_unref(s, gref, &act->mrs, act->prot); 409*b08d88e3SDavid Woodhouse g_free(act); 410*b08d88e3SDavid Woodhouse return true; 411*b08d88e3SDavid Woodhouse } 412*b08d88e3SDavid Woodhouse 413*b08d88e3SDavid Woodhouse static int xen_be_gnttab_unmap(struct xengntdev_handle *xgt, 414*b08d88e3SDavid Woodhouse void *start_address, uint32_t *refs, 415*b08d88e3SDavid Woodhouse uint32_t count) 416*b08d88e3SDavid Woodhouse { 417*b08d88e3SDavid Woodhouse XenGnttabState *s = xen_gnttab_singleton; 418*b08d88e3SDavid Woodhouse struct active_ref *act; 419*b08d88e3SDavid Woodhouse 420*b08d88e3SDavid Woodhouse if (!s) { 421*b08d88e3SDavid Woodhouse return -ENOTSUP; 422*b08d88e3SDavid Woodhouse } 423*b08d88e3SDavid Woodhouse 424*b08d88e3SDavid Woodhouse if (count != 1) { 425*b08d88e3SDavid Woodhouse return -EINVAL; 426*b08d88e3SDavid Woodhouse } 427*b08d88e3SDavid Woodhouse 428*b08d88e3SDavid Woodhouse QEMU_LOCK_GUARD(&s->gnt_lock); 429*b08d88e3SDavid Woodhouse 430*b08d88e3SDavid Woodhouse act = g_hash_table_lookup(xgt->active_maps, GINT_TO_POINTER(refs[0])); 431*b08d88e3SDavid Woodhouse if (!act) { 432*b08d88e3SDavid Woodhouse return -ENOENT; 433*b08d88e3SDavid Woodhouse } 434*b08d88e3SDavid Woodhouse 435*b08d88e3SDavid Woodhouse if (act->virtaddr != start_address) { 436*b08d88e3SDavid Woodhouse return -EINVAL; 437*b08d88e3SDavid Woodhouse } 438*b08d88e3SDavid Woodhouse 439*b08d88e3SDavid Woodhouse if (!--act->refcnt) { 440*b08d88e3SDavid Woodhouse do_unmap(GINT_TO_POINTER(refs[0]), act, s); 441*b08d88e3SDavid Woodhouse g_hash_table_remove(xgt->active_maps, GINT_TO_POINTER(refs[0])); 442*b08d88e3SDavid Woodhouse } 443*b08d88e3SDavid Woodhouse 444*b08d88e3SDavid Woodhouse return 0; 445*b08d88e3SDavid Woodhouse } 446*b08d88e3SDavid Woodhouse 447*b08d88e3SDavid Woodhouse /* 448*b08d88e3SDavid Woodhouse * This looks a bit like the one for true Xen in xen-operations.c but 449*b08d88e3SDavid Woodhouse * in emulation we don't support multi-page mappings. And under Xen we 450*b08d88e3SDavid Woodhouse * *want* the multi-page mappings so we have fewer bounces through the 451*b08d88e3SDavid Woodhouse * kernel and the hypervisor. So the code paths end up being similar, 452*b08d88e3SDavid Woodhouse * but different. 453*b08d88e3SDavid Woodhouse */ 454*b08d88e3SDavid Woodhouse static int xen_be_gnttab_copy(struct xengntdev_handle *xgt, bool to_domain, 455*b08d88e3SDavid Woodhouse uint32_t domid, XenGrantCopySegment *segs, 456*b08d88e3SDavid Woodhouse uint32_t nr_segs, Error **errp) 457*b08d88e3SDavid Woodhouse { 458*b08d88e3SDavid Woodhouse int prot = to_domain ? PROT_WRITE : PROT_READ; 459*b08d88e3SDavid Woodhouse unsigned int i; 460*b08d88e3SDavid Woodhouse 461*b08d88e3SDavid Woodhouse for (i = 0; i < nr_segs; i++) { 462*b08d88e3SDavid Woodhouse XenGrantCopySegment *seg = &segs[i]; 463*b08d88e3SDavid Woodhouse void *page; 464*b08d88e3SDavid Woodhouse uint32_t ref = to_domain ? seg->dest.foreign.ref : 465*b08d88e3SDavid Woodhouse seg->source.foreign.ref; 466*b08d88e3SDavid Woodhouse 467*b08d88e3SDavid Woodhouse page = xen_be_gnttab_map_refs(xgt, 1, domid, &ref, prot); 468*b08d88e3SDavid Woodhouse if (!page) { 469*b08d88e3SDavid Woodhouse if (errp) { 470*b08d88e3SDavid Woodhouse error_setg_errno(errp, errno, 471*b08d88e3SDavid Woodhouse "xen_be_gnttab_map_refs failed"); 472*b08d88e3SDavid Woodhouse } 473*b08d88e3SDavid Woodhouse return -errno; 474*b08d88e3SDavid Woodhouse } 475*b08d88e3SDavid Woodhouse 476*b08d88e3SDavid Woodhouse if (to_domain) { 477*b08d88e3SDavid Woodhouse memcpy(page + seg->dest.foreign.offset, seg->source.virt, 478*b08d88e3SDavid Woodhouse seg->len); 479*b08d88e3SDavid Woodhouse } else { 480*b08d88e3SDavid Woodhouse memcpy(seg->dest.virt, page + seg->source.foreign.offset, 481*b08d88e3SDavid Woodhouse seg->len); 482*b08d88e3SDavid Woodhouse } 483*b08d88e3SDavid Woodhouse 484*b08d88e3SDavid Woodhouse if (xen_be_gnttab_unmap(xgt, page, &ref, 1)) { 485*b08d88e3SDavid Woodhouse if (errp) { 486*b08d88e3SDavid Woodhouse error_setg_errno(errp, errno, "xen_be_gnttab_unmap failed"); 487*b08d88e3SDavid Woodhouse } 488*b08d88e3SDavid Woodhouse return -errno; 489*b08d88e3SDavid Woodhouse } 490*b08d88e3SDavid Woodhouse } 491*b08d88e3SDavid Woodhouse 492*b08d88e3SDavid Woodhouse return 0; 493*b08d88e3SDavid Woodhouse } 494*b08d88e3SDavid Woodhouse 495*b08d88e3SDavid Woodhouse static struct xengntdev_handle *xen_be_gnttab_open(void) 496*b08d88e3SDavid Woodhouse { 497*b08d88e3SDavid Woodhouse struct xengntdev_handle *xgt = g_new0(struct xengntdev_handle, 1); 498*b08d88e3SDavid Woodhouse 499*b08d88e3SDavid Woodhouse xgt->active_maps = g_hash_table_new(g_direct_hash, g_direct_equal); 500*b08d88e3SDavid Woodhouse return xgt; 501*b08d88e3SDavid Woodhouse } 502*b08d88e3SDavid Woodhouse 503*b08d88e3SDavid Woodhouse static int xen_be_gnttab_close(struct xengntdev_handle *xgt) 504*b08d88e3SDavid Woodhouse { 505*b08d88e3SDavid Woodhouse XenGnttabState *s = xen_gnttab_singleton; 506*b08d88e3SDavid Woodhouse 507*b08d88e3SDavid Woodhouse if (!s) { 508*b08d88e3SDavid Woodhouse return -ENOTSUP; 509*b08d88e3SDavid Woodhouse } 510*b08d88e3SDavid Woodhouse 511*b08d88e3SDavid Woodhouse g_hash_table_foreach_remove(xgt->active_maps, do_unmap, s); 512*b08d88e3SDavid Woodhouse g_hash_table_destroy(xgt->active_maps); 513*b08d88e3SDavid Woodhouse g_free(xgt); 514*b08d88e3SDavid Woodhouse return 0; 515*b08d88e3SDavid Woodhouse } 516*b08d88e3SDavid Woodhouse 517*b08d88e3SDavid Woodhouse static struct gnttab_backend_ops emu_gnttab_backend_ops = { 518*b08d88e3SDavid Woodhouse .open = xen_be_gnttab_open, 519*b08d88e3SDavid Woodhouse .close = xen_be_gnttab_close, 520*b08d88e3SDavid Woodhouse .grant_copy = xen_be_gnttab_copy, 521*b08d88e3SDavid Woodhouse .set_max_grants = xen_be_gnttab_set_max_grants, 522*b08d88e3SDavid Woodhouse .map_refs = xen_be_gnttab_map_refs, 523*b08d88e3SDavid Woodhouse .unmap = xen_be_gnttab_unmap, 524*b08d88e3SDavid Woodhouse }; 525*b08d88e3SDavid Woodhouse 526