xref: /qemu/include/hw/vfio/vfio-container-base.h (revision 2a53c4f5c534a1ab825ba03e0d3ec45a7c2b90d8)
1 /*
2  * VFIO BASE CONTAINER
3  *
4  * Copyright (C) 2023 Intel Corporation.
5  * Copyright Red Hat, Inc. 2023
6  *
7  * Authors: Yi Liu <yi.l.liu@intel.com>
8  *          Eric Auger <eric.auger@redhat.com>
9  *
10  * SPDX-License-Identifier: GPL-2.0-or-later
11  */
12 
13 #ifndef HW_VFIO_VFIO_CONTAINER_BASE_H
14 #define HW_VFIO_VFIO_CONTAINER_BASE_H
15 
16 #include "system/memory.h"
17 
18 typedef struct VFIODevice VFIODevice;
19 typedef struct VFIOIOMMUClass VFIOIOMMUClass;
20 
21 typedef struct {
22     unsigned long *bitmap;
23     hwaddr size;
24     hwaddr pages;
25 } VFIOBitmap;
26 
27 typedef struct VFIOAddressSpace {
28     AddressSpace *as;
29     QLIST_HEAD(, VFIOContainerBase) containers;
30     QLIST_ENTRY(VFIOAddressSpace) list;
31 } VFIOAddressSpace;
32 
33 /*
34  * This is the base object for vfio container backends
35  */
36 typedef struct VFIOContainerBase {
37     Object parent;
38     VFIOAddressSpace *space;
39     MemoryListener listener;
40     Error *error;
41     bool initialized;
42     uint64_t dirty_pgsizes;
43     uint64_t max_dirty_bitmap_size;
44     unsigned long pgsizes;
45     unsigned int dma_max_mappings;
46     bool dirty_pages_supported;
47     bool dirty_pages_started; /* Protected by BQL */
48     QLIST_HEAD(, VFIOGuestIOMMU) giommu_list;
49     QLIST_HEAD(, VFIORamDiscardListener) vrdl_list;
50     QLIST_ENTRY(VFIOContainerBase) next;
51     QLIST_HEAD(, VFIODevice) device_list;
52     GList *iova_ranges;
53     NotifierWithReturn cpr_reboot_notifier;
54 } VFIOContainerBase;
55 
56 typedef struct VFIOGuestIOMMU {
57     VFIOContainerBase *bcontainer;
58     IOMMUMemoryRegion *iommu_mr;
59     hwaddr iommu_offset;
60     IOMMUNotifier n;
61     QLIST_ENTRY(VFIOGuestIOMMU) giommu_next;
62 } VFIOGuestIOMMU;
63 
64 typedef struct VFIORamDiscardListener {
65     VFIOContainerBase *bcontainer;
66     MemoryRegion *mr;
67     hwaddr offset_within_address_space;
68     hwaddr size;
69     uint64_t granularity;
70     RamDiscardListener listener;
71     QLIST_ENTRY(VFIORamDiscardListener) next;
72 } VFIORamDiscardListener;
73 
74 VFIOAddressSpace *vfio_address_space_get(AddressSpace *as);
75 void vfio_address_space_put(VFIOAddressSpace *space);
76 void vfio_address_space_insert(VFIOAddressSpace *space,
77                                VFIOContainerBase *bcontainer);
78 
79 int vfio_container_dma_map(VFIOContainerBase *bcontainer,
80                            hwaddr iova, ram_addr_t size,
81                            void *vaddr, bool readonly, MemoryRegion *mr);
82 int vfio_container_dma_unmap(VFIOContainerBase *bcontainer,
83                              hwaddr iova, ram_addr_t size,
84                              IOMMUTLBEntry *iotlb, bool unmap_all);
85 bool vfio_container_add_section_window(VFIOContainerBase *bcontainer,
86                                        MemoryRegionSection *section,
87                                        Error **errp);
88 void vfio_container_del_section_window(VFIOContainerBase *bcontainer,
89                                        MemoryRegionSection *section);
90 int vfio_container_set_dirty_page_tracking(VFIOContainerBase *bcontainer,
91                                            bool start, Error **errp);
92 bool vfio_container_dirty_tracking_is_started(
93     const VFIOContainerBase *bcontainer);
94 bool vfio_container_devices_dirty_tracking_is_supported(
95     const VFIOContainerBase *bcontainer);
96 int vfio_container_query_dirty_bitmap(const VFIOContainerBase *bcontainer,
97     uint64_t iova, uint64_t size, ram_addr_t ram_addr, Error **errp);
98 
99 GList *vfio_container_get_iova_ranges(const VFIOContainerBase *bcontainer);
100 
101 static inline uint64_t
vfio_container_get_page_size_mask(const VFIOContainerBase * bcontainer)102 vfio_container_get_page_size_mask(const VFIOContainerBase *bcontainer)
103 {
104     assert(bcontainer);
105     return bcontainer->pgsizes;
106 }
107 
108 #define TYPE_VFIO_IOMMU "vfio-iommu"
109 #define TYPE_VFIO_IOMMU_LEGACY TYPE_VFIO_IOMMU "-legacy"
110 #define TYPE_VFIO_IOMMU_SPAPR TYPE_VFIO_IOMMU "-spapr"
111 #define TYPE_VFIO_IOMMU_IOMMUFD TYPE_VFIO_IOMMU "-iommufd"
112 
113 OBJECT_DECLARE_TYPE(VFIOContainerBase, VFIOIOMMUClass, VFIO_IOMMU)
114 
115 struct VFIOIOMMUClass {
116     ObjectClass parent_class;
117 
118     /**
119      * @setup
120      *
121      * Perform basic setup of the container, including configuring IOMMU
122      * capabilities, IOVA ranges, supported page sizes, etc.
123      *
124      * @bcontainer: #VFIOContainerBase
125      * @errp: pointer to Error*, to store an error if it happens.
126      *
127      * Returns true to indicate success and false for error.
128      */
129     bool (*setup)(VFIOContainerBase *bcontainer, Error **errp);
130 
131     /**
132      * @listener_begin
133      *
134      * Called at the beginning of an address space update transaction.
135      * See #MemoryListener.
136      *
137      * @bcontainer: #VFIOContainerBase
138      */
139     void (*listener_begin)(VFIOContainerBase *bcontainer);
140 
141     /**
142      * @listener_commit
143      *
144      * Called at the end of an address space update transaction,
145      * See #MemoryListener.
146      *
147      * @bcontainer: #VFIOContainerBase
148      */
149     void (*listener_commit)(VFIOContainerBase *bcontainer);
150 
151     /**
152      * @dma_map
153      *
154      * Map an address range into the container. Note that the memory region is
155      * referenced within an RCU read lock region across this call.
156      *
157      * @bcontainer: #VFIOContainerBase to use
158      * @iova: start address to map
159      * @size: size of the range to map
160      * @vaddr: process virtual address of mapping
161      * @readonly: true if mapping should be readonly
162      * @mr: the memory region for this mapping
163      *
164      * Returns 0 to indicate success and -errno otherwise.
165      */
166     int (*dma_map)(const VFIOContainerBase *bcontainer,
167                    hwaddr iova, ram_addr_t size,
168                    void *vaddr, bool readonly, MemoryRegion *mr);
169     /**
170      * @dma_unmap
171      *
172      * Unmap an address range from the container.
173      *
174      * @bcontainer: #VFIOContainerBase to use for unmap
175      * @iova: start address to unmap
176      * @size: size of the range to unmap
177      * @iotlb: The IOMMU TLB mapping entry (or NULL)
178      * @unmap_all: if set, unmap the entire address space
179      *
180      * Returns 0 to indicate success and -errno otherwise.
181      */
182     int (*dma_unmap)(const VFIOContainerBase *bcontainer,
183                      hwaddr iova, ram_addr_t size,
184                      IOMMUTLBEntry *iotlb, bool unmap_all);
185 
186 
187     /**
188      * @attach_device
189      *
190      * Associate the given device with a container and do some related
191      * initialization of the device context.
192      *
193      * @name: name of the device
194      * @vbasedev: the device
195      * @as: address space to use
196      * @errp: pointer to Error*, to store an error if it happens.
197      *
198      * Returns true to indicate success and false for error.
199      */
200     bool (*attach_device)(const char *name, VFIODevice *vbasedev,
201                           AddressSpace *as, Error **errp);
202 
203     /*
204      * @detach_device
205      *
206      * Detach the given device from its container and clean up any necessary
207      * state.
208      *
209      * @vbasedev: the device to disassociate
210      */
211     void (*detach_device)(VFIODevice *vbasedev);
212 
213     /* migration feature */
214 
215     /**
216      * @set_dirty_page_tracking
217      *
218      * Start or stop dirty pages tracking on VFIO container
219      *
220      * @bcontainer: #VFIOContainerBase on which to de/activate dirty
221      *              page tracking
222      * @start: indicates whether to start or stop dirty pages tracking
223      * @errp: pointer to Error*, to store an error if it happens.
224      *
225      * Returns zero to indicate success and negative for error.
226      */
227     int (*set_dirty_page_tracking)(const VFIOContainerBase *bcontainer,
228                                    bool start, Error **errp);
229     /**
230      * @query_dirty_bitmap
231      *
232      * Get bitmap of dirty pages from container
233      *
234      * @bcontainer: #VFIOContainerBase from which to get dirty pages
235      * @vbmap: #VFIOBitmap internal bitmap structure
236      * @iova: iova base address
237      * @size: size of iova range
238      * @errp: pointer to Error*, to store an error if it happens.
239      *
240      * Returns zero to indicate success and negative for error.
241      */
242     int (*query_dirty_bitmap)(const VFIOContainerBase *bcontainer,
243                 VFIOBitmap *vbmap, hwaddr iova, hwaddr size, Error **errp);
244     /* PCI specific */
245     int (*pci_hot_reset)(VFIODevice *vbasedev, bool single);
246 
247     /* SPAPR specific */
248     bool (*add_window)(VFIOContainerBase *bcontainer,
249                        MemoryRegionSection *section,
250                        Error **errp);
251     void (*del_window)(VFIOContainerBase *bcontainer,
252                        MemoryRegionSection *section);
253     void (*release)(VFIOContainerBase *bcontainer);
254 };
255 
256 VFIORamDiscardListener *vfio_find_ram_discard_listener(
257     VFIOContainerBase *bcontainer, MemoryRegionSection *section);
258 
259 #endif /* HW_VFIO_VFIO_CONTAINER_BASE_H */
260