xref: /qemu/rust/qemu-api/src/memory.rs (revision 08b462dd9970a88d7f0e7c61ca48502463b0b78d)
1 // Copyright 2024 Red Hat, Inc.
2 // Author(s): Paolo Bonzini <pbonzini@redhat.com>
3 // SPDX-License-Identifier: GPL-2.0-or-later
4 
5 //! Bindings for `MemoryRegion`, `MemoryRegionOps` and `MemTxAttrs`
6 
7 use std::{
8     ffi::{CStr, CString},
9     marker::PhantomData,
10     os::raw::{c_uint, c_void},
11 };
12 
13 pub use bindings::{hwaddr, MemTxAttrs};
14 
15 use crate::{
16     bindings::{self, device_endian, memory_region_init_io},
17     callbacks::FnCall,
18     cell::Opaque,
19     prelude::*,
20     zeroable::Zeroable,
21 };
22 
23 pub struct MemoryRegionOps<T>(
24     bindings::MemoryRegionOps,
25     // Note: quite often you'll see PhantomData<fn(&T)> mentioned when discussing
26     // covariance and contravariance; you don't need any of those to understand
27     // this usage of PhantomData.  Quite simply, MemoryRegionOps<T> *logically*
28     // holds callbacks that take an argument of type &T, except the type is erased
29     // before the callback is stored in the bindings::MemoryRegionOps field.
30     // The argument of PhantomData is a function pointer in order to represent
31     // that relationship; while that will also provide desirable and safe variance
32     // for T, variance is not the point but just a consequence.
33     PhantomData<fn(&T)>,
34 );
35 
36 // SAFETY: When a *const T is passed to the callbacks, the call itself
37 // is done in a thread-safe manner.  The invocation is okay as long as
38 // T itself is `Sync`.
39 unsafe impl<T: Sync> Sync for MemoryRegionOps<T> {}
40 
41 #[derive(Clone)]
42 pub struct MemoryRegionOpsBuilder<T>(bindings::MemoryRegionOps, PhantomData<fn(&T)>);
43 
44 unsafe extern "C" fn memory_region_ops_read_cb<T, F: for<'a> FnCall<(&'a T, hwaddr, u32), u64>>(
45     opaque: *mut c_void,
46     addr: hwaddr,
47     size: c_uint,
48 ) -> u64 {
49     F::call((unsafe { &*(opaque.cast::<T>()) }, addr, size))
50 }
51 
52 unsafe extern "C" fn memory_region_ops_write_cb<T, F: for<'a> FnCall<(&'a T, hwaddr, u64, u32)>>(
53     opaque: *mut c_void,
54     addr: hwaddr,
55     data: u64,
56     size: c_uint,
57 ) {
58     F::call((unsafe { &*(opaque.cast::<T>()) }, addr, data, size))
59 }
60 
61 impl<T> MemoryRegionOpsBuilder<T> {
62     #[must_use]
63     pub const fn read<F: for<'a> FnCall<(&'a T, hwaddr, u32), u64>>(mut self, _f: &F) -> Self {
64         self.0.read = Some(memory_region_ops_read_cb::<T, F>);
65         self
66     }
67 
68     #[must_use]
69     pub const fn write<F: for<'a> FnCall<(&'a T, hwaddr, u64, u32)>>(mut self, _f: &F) -> Self {
70         self.0.write = Some(memory_region_ops_write_cb::<T, F>);
71         self
72     }
73 
74     #[must_use]
75     pub const fn big_endian(mut self) -> Self {
76         self.0.endianness = device_endian::DEVICE_BIG_ENDIAN;
77         self
78     }
79 
80     #[must_use]
81     pub const fn little_endian(mut self) -> Self {
82         self.0.endianness = device_endian::DEVICE_LITTLE_ENDIAN;
83         self
84     }
85 
86     #[must_use]
87     pub const fn native_endian(mut self) -> Self {
88         self.0.endianness = device_endian::DEVICE_NATIVE_ENDIAN;
89         self
90     }
91 
92     #[must_use]
93     pub const fn valid_sizes(mut self, min: u32, max: u32) -> Self {
94         self.0.valid.min_access_size = min;
95         self.0.valid.max_access_size = max;
96         self
97     }
98 
99     #[must_use]
100     pub const fn valid_unaligned(mut self) -> Self {
101         self.0.valid.unaligned = true;
102         self
103     }
104 
105     #[must_use]
106     pub const fn impl_sizes(mut self, min: u32, max: u32) -> Self {
107         self.0.impl_.min_access_size = min;
108         self.0.impl_.max_access_size = max;
109         self
110     }
111 
112     #[must_use]
113     pub const fn impl_unaligned(mut self) -> Self {
114         self.0.impl_.unaligned = true;
115         self
116     }
117 
118     #[must_use]
119     pub const fn build(self) -> MemoryRegionOps<T> {
120         MemoryRegionOps::<T>(self.0, PhantomData)
121     }
122 
123     #[must_use]
124     pub const fn new() -> Self {
125         Self(bindings::MemoryRegionOps::ZERO, PhantomData)
126     }
127 }
128 
129 impl<T> Default for MemoryRegionOpsBuilder<T> {
130     fn default() -> Self {
131         Self::new()
132     }
133 }
134 
135 /// A safe wrapper around [`bindings::MemoryRegion`].
136 #[repr(transparent)]
137 #[derive(qemu_api_macros::Wrapper)]
138 pub struct MemoryRegion(Opaque<bindings::MemoryRegion>);
139 
140 unsafe impl Send for MemoryRegion {}
141 unsafe impl Sync for MemoryRegion {}
142 
143 impl MemoryRegion {
144     // inline to ensure that it is not included in tests, which only
145     // link to hwcore and qom.  FIXME: inlining is actually the opposite
146     // of what we want, since this is the type-erased version of the
147     // init_io function below.  Look into splitting the qemu_api crate.
148     #[inline(always)]
149     unsafe fn do_init_io(
150         slot: *mut bindings::MemoryRegion,
151         owner: *mut Object,
152         ops: &'static bindings::MemoryRegionOps,
153         name: &'static str,
154         size: u64,
155     ) {
156         unsafe {
157             let cstr = CString::new(name).unwrap();
158             memory_region_init_io(
159                 slot,
160                 owner.cast::<bindings::Object>(),
161                 ops,
162                 owner.cast::<c_void>(),
163                 cstr.as_ptr(),
164                 size,
165             );
166         }
167     }
168 
169     pub fn init_io<T: IsA<Object>>(
170         &mut self,
171         owner: *mut T,
172         ops: &'static MemoryRegionOps<T>,
173         name: &'static str,
174         size: u64,
175     ) {
176         unsafe {
177             Self::do_init_io(
178                 self.0.as_mut_ptr(),
179                 owner.cast::<Object>(),
180                 &ops.0,
181                 name,
182                 size,
183             );
184         }
185     }
186 }
187 
188 unsafe impl ObjectType for MemoryRegion {
189     type Class = bindings::MemoryRegionClass;
190     const TYPE_NAME: &'static CStr =
191         unsafe { CStr::from_bytes_with_nul_unchecked(bindings::TYPE_MEMORY_REGION) };
192 }
193 qom_isa!(MemoryRegion: Object);
194 
195 /// A special `MemTxAttrs` constant, used to indicate that no memory
196 /// attributes are specified.
197 ///
198 /// Bus masters which don't specify any attributes will get this,
199 /// which has all attribute bits clear except the topmost one
200 /// (so that we can distinguish "all attributes deliberately clear"
201 /// from "didn't specify" if necessary).
202 pub const MEMTXATTRS_UNSPECIFIED: MemTxAttrs = MemTxAttrs {
203     unspecified: true,
204     ..Zeroable::ZERO
205 };
206