1*590faa03SPaolo Bonzini // Copyright 2024 Red Hat, Inc. 2*590faa03SPaolo Bonzini // Author(s): Paolo Bonzini <pbonzini@redhat.com> 3*590faa03SPaolo Bonzini // SPDX-License-Identifier: GPL-2.0-or-later 4*590faa03SPaolo Bonzini 5*590faa03SPaolo Bonzini //! Bindings for `MemoryRegion` and `MemoryRegionOps` 6*590faa03SPaolo Bonzini 7*590faa03SPaolo Bonzini use std::{ 8*590faa03SPaolo Bonzini ffi::{CStr, CString}, 9*590faa03SPaolo Bonzini marker::{PhantomData, PhantomPinned}, 10*590faa03SPaolo Bonzini os::raw::{c_uint, c_void}, 11*590faa03SPaolo Bonzini ptr::addr_of, 12*590faa03SPaolo Bonzini }; 13*590faa03SPaolo Bonzini 14*590faa03SPaolo Bonzini pub use bindings::hwaddr; 15*590faa03SPaolo Bonzini 16*590faa03SPaolo Bonzini use crate::{ 17*590faa03SPaolo Bonzini bindings::{self, device_endian, memory_region_init_io}, 18*590faa03SPaolo Bonzini callbacks::FnCall, 19*590faa03SPaolo Bonzini prelude::*, 20*590faa03SPaolo Bonzini zeroable::Zeroable, 21*590faa03SPaolo Bonzini }; 22*590faa03SPaolo Bonzini 23*590faa03SPaolo Bonzini pub struct MemoryRegionOps<T>( 24*590faa03SPaolo Bonzini bindings::MemoryRegionOps, 25*590faa03SPaolo Bonzini // Note: quite often you'll see PhantomData<fn(&T)> mentioned when discussing 26*590faa03SPaolo Bonzini // covariance and contravariance; you don't need any of those to understand 27*590faa03SPaolo Bonzini // this usage of PhantomData. Quite simply, MemoryRegionOps<T> *logically* 28*590faa03SPaolo Bonzini // holds callbacks that take an argument of type &T, except the type is erased 29*590faa03SPaolo Bonzini // before the callback is stored in the bindings::MemoryRegionOps field. 30*590faa03SPaolo Bonzini // The argument of PhantomData is a function pointer in order to represent 31*590faa03SPaolo Bonzini // that relationship; while that will also provide desirable and safe variance 32*590faa03SPaolo Bonzini // for T, variance is not the point but just a consequence. 33*590faa03SPaolo Bonzini PhantomData<fn(&T)>, 34*590faa03SPaolo Bonzini ); 35*590faa03SPaolo Bonzini 36*590faa03SPaolo Bonzini // SAFETY: When a *const T is passed to the callbacks, the call itself 37*590faa03SPaolo Bonzini // is done in a thread-safe manner. The invocation is okay as long as 38*590faa03SPaolo Bonzini // T itself is `Sync`. 39*590faa03SPaolo Bonzini unsafe impl<T: Sync> Sync for MemoryRegionOps<T> {} 40*590faa03SPaolo Bonzini 41*590faa03SPaolo Bonzini #[derive(Clone)] 42*590faa03SPaolo Bonzini pub struct MemoryRegionOpsBuilder<T>(bindings::MemoryRegionOps, PhantomData<fn(&T)>); 43*590faa03SPaolo Bonzini 44*590faa03SPaolo Bonzini unsafe extern "C" fn memory_region_ops_read_cb<T, F: for<'a> FnCall<(&'a T, hwaddr, u32), u64>>( 45*590faa03SPaolo Bonzini opaque: *mut c_void, 46*590faa03SPaolo Bonzini addr: hwaddr, 47*590faa03SPaolo Bonzini size: c_uint, 48*590faa03SPaolo Bonzini ) -> u64 { 49*590faa03SPaolo Bonzini F::call((unsafe { &*(opaque.cast::<T>()) }, addr, size)) 50*590faa03SPaolo Bonzini } 51*590faa03SPaolo Bonzini 52*590faa03SPaolo Bonzini unsafe extern "C" fn memory_region_ops_write_cb<T, F: for<'a> FnCall<(&'a T, hwaddr, u64, u32)>>( 53*590faa03SPaolo Bonzini opaque: *mut c_void, 54*590faa03SPaolo Bonzini addr: hwaddr, 55*590faa03SPaolo Bonzini data: u64, 56*590faa03SPaolo Bonzini size: c_uint, 57*590faa03SPaolo Bonzini ) { 58*590faa03SPaolo Bonzini F::call((unsafe { &*(opaque.cast::<T>()) }, addr, data, size)) 59*590faa03SPaolo Bonzini } 60*590faa03SPaolo Bonzini 61*590faa03SPaolo Bonzini impl<T> MemoryRegionOpsBuilder<T> { 62*590faa03SPaolo Bonzini #[must_use] 63*590faa03SPaolo Bonzini pub const fn read<F: for<'a> FnCall<(&'a T, hwaddr, u32), u64>>(mut self, _f: &F) -> Self { 64*590faa03SPaolo Bonzini self.0.read = Some(memory_region_ops_read_cb::<T, F>); 65*590faa03SPaolo Bonzini self 66*590faa03SPaolo Bonzini } 67*590faa03SPaolo Bonzini 68*590faa03SPaolo Bonzini #[must_use] 69*590faa03SPaolo Bonzini pub const fn write<F: for<'a> FnCall<(&'a T, hwaddr, u64, u32)>>(mut self, _f: &F) -> Self { 70*590faa03SPaolo Bonzini self.0.write = Some(memory_region_ops_write_cb::<T, F>); 71*590faa03SPaolo Bonzini self 72*590faa03SPaolo Bonzini } 73*590faa03SPaolo Bonzini 74*590faa03SPaolo Bonzini #[must_use] 75*590faa03SPaolo Bonzini pub const fn big_endian(mut self) -> Self { 76*590faa03SPaolo Bonzini self.0.endianness = device_endian::DEVICE_BIG_ENDIAN; 77*590faa03SPaolo Bonzini self 78*590faa03SPaolo Bonzini } 79*590faa03SPaolo Bonzini 80*590faa03SPaolo Bonzini #[must_use] 81*590faa03SPaolo Bonzini pub const fn little_endian(mut self) -> Self { 82*590faa03SPaolo Bonzini self.0.endianness = device_endian::DEVICE_LITTLE_ENDIAN; 83*590faa03SPaolo Bonzini self 84*590faa03SPaolo Bonzini } 85*590faa03SPaolo Bonzini 86*590faa03SPaolo Bonzini #[must_use] 87*590faa03SPaolo Bonzini pub const fn native_endian(mut self) -> Self { 88*590faa03SPaolo Bonzini self.0.endianness = device_endian::DEVICE_NATIVE_ENDIAN; 89*590faa03SPaolo Bonzini self 90*590faa03SPaolo Bonzini } 91*590faa03SPaolo Bonzini 92*590faa03SPaolo Bonzini #[must_use] 93*590faa03SPaolo Bonzini pub const fn valid_sizes(mut self, min: u32, max: u32) -> Self { 94*590faa03SPaolo Bonzini self.0.valid.min_access_size = min; 95*590faa03SPaolo Bonzini self.0.valid.max_access_size = max; 96*590faa03SPaolo Bonzini self 97*590faa03SPaolo Bonzini } 98*590faa03SPaolo Bonzini 99*590faa03SPaolo Bonzini #[must_use] 100*590faa03SPaolo Bonzini pub const fn valid_unaligned(mut self) -> Self { 101*590faa03SPaolo Bonzini self.0.valid.unaligned = true; 102*590faa03SPaolo Bonzini self 103*590faa03SPaolo Bonzini } 104*590faa03SPaolo Bonzini 105*590faa03SPaolo Bonzini #[must_use] 106*590faa03SPaolo Bonzini pub const fn impl_sizes(mut self, min: u32, max: u32) -> Self { 107*590faa03SPaolo Bonzini self.0.impl_.min_access_size = min; 108*590faa03SPaolo Bonzini self.0.impl_.max_access_size = max; 109*590faa03SPaolo Bonzini self 110*590faa03SPaolo Bonzini } 111*590faa03SPaolo Bonzini 112*590faa03SPaolo Bonzini #[must_use] 113*590faa03SPaolo Bonzini pub const fn impl_unaligned(mut self) -> Self { 114*590faa03SPaolo Bonzini self.0.impl_.unaligned = true; 115*590faa03SPaolo Bonzini self 116*590faa03SPaolo Bonzini } 117*590faa03SPaolo Bonzini 118*590faa03SPaolo Bonzini #[must_use] 119*590faa03SPaolo Bonzini pub const fn build(self) -> MemoryRegionOps<T> { 120*590faa03SPaolo Bonzini MemoryRegionOps::<T>(self.0, PhantomData) 121*590faa03SPaolo Bonzini } 122*590faa03SPaolo Bonzini 123*590faa03SPaolo Bonzini #[must_use] 124*590faa03SPaolo Bonzini pub const fn new() -> Self { 125*590faa03SPaolo Bonzini Self(bindings::MemoryRegionOps::ZERO, PhantomData) 126*590faa03SPaolo Bonzini } 127*590faa03SPaolo Bonzini } 128*590faa03SPaolo Bonzini 129*590faa03SPaolo Bonzini impl<T> Default for MemoryRegionOpsBuilder<T> { 130*590faa03SPaolo Bonzini fn default() -> Self { 131*590faa03SPaolo Bonzini Self::new() 132*590faa03SPaolo Bonzini } 133*590faa03SPaolo Bonzini } 134*590faa03SPaolo Bonzini 135*590faa03SPaolo Bonzini /// A safe wrapper around [`bindings::MemoryRegion`]. Compared to the 136*590faa03SPaolo Bonzini /// underlying C struct it is marked as pinned because the QOM tree 137*590faa03SPaolo Bonzini /// contains a pointer to it. 138*590faa03SPaolo Bonzini pub struct MemoryRegion { 139*590faa03SPaolo Bonzini inner: bindings::MemoryRegion, 140*590faa03SPaolo Bonzini _pin: PhantomPinned, 141*590faa03SPaolo Bonzini } 142*590faa03SPaolo Bonzini 143*590faa03SPaolo Bonzini impl MemoryRegion { 144*590faa03SPaolo Bonzini // inline to ensure that it is not included in tests, which only 145*590faa03SPaolo Bonzini // link to hwcore and qom. FIXME: inlining is actually the opposite 146*590faa03SPaolo Bonzini // of what we want, since this is the type-erased version of the 147*590faa03SPaolo Bonzini // init_io function below. Look into splitting the qemu_api crate. 148*590faa03SPaolo Bonzini #[inline(always)] 149*590faa03SPaolo Bonzini unsafe fn do_init_io( 150*590faa03SPaolo Bonzini slot: *mut bindings::MemoryRegion, 151*590faa03SPaolo Bonzini owner: *mut Object, 152*590faa03SPaolo Bonzini ops: &'static bindings::MemoryRegionOps, 153*590faa03SPaolo Bonzini name: &'static str, 154*590faa03SPaolo Bonzini size: u64, 155*590faa03SPaolo Bonzini ) { 156*590faa03SPaolo Bonzini unsafe { 157*590faa03SPaolo Bonzini let cstr = CString::new(name).unwrap(); 158*590faa03SPaolo Bonzini memory_region_init_io( 159*590faa03SPaolo Bonzini slot, 160*590faa03SPaolo Bonzini owner.cast::<Object>(), 161*590faa03SPaolo Bonzini ops, 162*590faa03SPaolo Bonzini owner.cast::<c_void>(), 163*590faa03SPaolo Bonzini cstr.as_ptr(), 164*590faa03SPaolo Bonzini size, 165*590faa03SPaolo Bonzini ); 166*590faa03SPaolo Bonzini } 167*590faa03SPaolo Bonzini } 168*590faa03SPaolo Bonzini 169*590faa03SPaolo Bonzini pub fn init_io<T: IsA<Object>>( 170*590faa03SPaolo Bonzini &mut self, 171*590faa03SPaolo Bonzini owner: *mut T, 172*590faa03SPaolo Bonzini ops: &'static MemoryRegionOps<T>, 173*590faa03SPaolo Bonzini name: &'static str, 174*590faa03SPaolo Bonzini size: u64, 175*590faa03SPaolo Bonzini ) { 176*590faa03SPaolo Bonzini unsafe { 177*590faa03SPaolo Bonzini Self::do_init_io(&mut self.inner, owner.cast::<Object>(), &ops.0, name, size); 178*590faa03SPaolo Bonzini } 179*590faa03SPaolo Bonzini } 180*590faa03SPaolo Bonzini 181*590faa03SPaolo Bonzini pub(crate) const fn as_mut_ptr(&self) -> *mut bindings::MemoryRegion { 182*590faa03SPaolo Bonzini addr_of!(self.inner) as *mut _ 183*590faa03SPaolo Bonzini } 184*590faa03SPaolo Bonzini } 185*590faa03SPaolo Bonzini 186*590faa03SPaolo Bonzini unsafe impl ObjectType for MemoryRegion { 187*590faa03SPaolo Bonzini type Class = bindings::MemoryRegionClass; 188*590faa03SPaolo Bonzini const TYPE_NAME: &'static CStr = 189*590faa03SPaolo Bonzini unsafe { CStr::from_bytes_with_nul_unchecked(bindings::TYPE_MEMORY_REGION) }; 190*590faa03SPaolo Bonzini } 191*590faa03SPaolo Bonzini qom_isa!(MemoryRegion: Object); 192