/qemu/hw/sd/ |
H A D | omap_mmc.c | 2 * OMAP on-chip MMC/SD host emulation. 80 static void omap_mmc_fifolevel_update(OMAPMMCState *host) in omap_mmc_fifolevel_update() argument 82 if (!host->transfer && !host->fifo_len) { in omap_mmc_fifolevel_update() 83 host->status &= 0xf3ff; in omap_mmc_fifolevel_update() 87 if (host->fifo_len > host->af_level && host->ddir) { in omap_mmc_fifolevel_update() 88 if (host->rx_dma) { in omap_mmc_fifolevel_update() 89 host->status &= 0xfbff; in omap_mmc_fifolevel_update() 90 qemu_irq_raise(host->dma_rx_gpio); in omap_mmc_fifolevel_update() 92 host->status |= 0x0400; in omap_mmc_fifolevel_update() 94 host->status &= 0xfbff; in omap_mmc_fifolevel_update() [all …]
|
/qemu/docs/system/ |
H A D | cpu-models-x86.rst.inc | 13 (1) **Host passthrough** 15 This passes the host CPU model features, model, stepping, exactly to 16 the guest. Note that KVM may filter out some host CPU model features 27 isolation from the host CPU, allowing greater flexibility in live 33 Libvirt supports a third way to configure CPU models known as "Host 35 a CPU model that is similar the host CPU, and then adding extra features 36 to approximate the host model as closely as possible. This does not 37 guarantee the CPU family, stepping, etc will precisely match the host 38 CPU, as they would with "Host passthrough", but gives much of the 69 matches the generation of the host CPUs in use. In a deployment with a [all …]
|
/qemu/include/system/ |
H A D | host_iommu_device.h | 2 * Host IOMMU device abstract declaration 26 * struct HostIOMMUDeviceCaps - Define host IOMMU device capabilities. 28 * @type: host platform IOMMU type. 30 * @hw_caps: host platform IOMMU capabilities (e.g. on IOMMUFD this represents 33 * @vendor_caps: host platform IOMMU vendor specific capabilities (e.g. on 35 * with host IOMMU @type specific hardware information data) 44 #define TYPE_HOST_IOMMU_DEVICE "host-iommu-device" 60 * struct HostIOMMUDeviceClass - The base class for all host IOMMU devices. 62 * Different types of host devices (e.g., VFIO or VDPA device) or devices 70 * @realize: initialize host IOMMU device instance further. [all …]
|
/qemu/tests/qemu-iotests/ |
H A D | 207.out | 3 … {"driver": "ssh", "location": {"path": "TEST_DIR/PID-t.img", "server": {"host": "127.0.0.1", "por… 17 === Test host-key-check options === 19 --- no host key checking -- 21 …s": {"driver": "ssh", "location": {"host-key-check": {"mode": "none"}, "path": "TEST_DIR/PID-t.img… 32 …{"driver": "ssh", "location": {"host-key-check": {"mode": "known_hosts"}, "path": "TEST_DIR/PID-t.… 43 …ssh", "location": {"host-key-check": {"hash": "wrong", "mode": "hash", "type": "md5"}, "path": "TE… 45 Job failed: remote host key fingerprint 'md5:HASH' does not match host_key_check 'md5:wrong' 49 …ssh", "location": {"host-key-check": {"hash": "HASH", "mode": "hash", "type": "md5"}, "path": "TES… 60 …sh", "location": {"host-key-check": {"hash": "wrong", "mode": "hash", "type": "sha1"}, "path": "TE… 62 Job failed: remote host key fingerprint 'sha1:HASH' does not match host_key_check 'sha1:wrong' [all …]
|
H A D | 207 | 64 'host': '127.0.0.1', 76 # Test host-key-check options 78 iotests.log("=== Test host-key-check options ===") 81 iotests.log("--- no host key checking --") 89 'host': '127.0.0.1', 92 'host-key-check': { 109 'host': '127.0.0.1', 112 'host-key-check': { 152 'host': '127.0.0.1', 176 'host': '127.0.0.1', [all …]
|
/qemu/target/mips/tcg/ |
H A D | lmmi_helper.c | 248 unsigned host = BYTE_ORDER_XOR(3); in helper_pshufh() local 255 vd.uh[i ^ host] = vs.uh[(ft & 3) ^ host]; in helper_pshufh() 334 unsigned host = BYTE_ORDER_XOR(3); in helper_punpcklhw() local 339 vd.uh[0 ^ host] = vs.uh[0 ^ host]; in helper_punpcklhw() 340 vd.uh[1 ^ host] = vt.uh[0 ^ host]; in helper_punpcklhw() 341 vd.uh[2 ^ host] = vs.uh[1 ^ host]; in helper_punpcklhw() 342 vd.uh[3 ^ host] = vt.uh[1 ^ host]; in helper_punpcklhw() 349 unsigned host = BYTE_ORDER_XOR(3); in helper_punpckhhw() local 354 vd.uh[0 ^ host] = vs.uh[2 ^ host]; in helper_punpckhhw() 355 vd.uh[1 ^ host] = vt.uh[2 ^ host]; in helper_punpckhhw() [all …]
|
/qemu/docs/ |
H A D | qdev-device-use.txt | 30 A QEMU block device (drive) has a host and a guest part. 35 and is connected to a host part. 39 controller is connected to up to two host drives. 41 The old ways to define block devices define host and guest part 45 The new way keeps the parts separate: you create the host part with 59 -drive if=none,id=DRIVE-ID,HOST-OPTS... 62 The old OPTS get split into HOST-OPTS and DEV-OPTS as follows: 65 HOST-OPTS. 67 * cyls, head, secs and trans go into HOST-OPTS. Future work: they 75 Additionally, readonly=on goes into HOST-OPTS. [all …]
|
H A D | bypass-iommu.txt | 13 PCI host bridges have a bypass_iommu property. This property is used to 14 determine whether the devices attached on the PCI host bridge will bypass 17 bypass vIOMMU. When bypass_iommu property is not set for a host bridge, 22 The bypass iommu feature support PXB host bridge and default main host 32 (2) Arm default host bridge 46 - a default host bridge which bypass SMMUv3 47 - a pxb host bridge which go through SMMUv3 48 - a pxb host bridge which bypass SMMUv3 60 - a default host bridge which bypass iommu 61 - a pxb host bridge which go through iommu [all …]
|
H A D | pci_expander_bridge.txt | 6 PXB is a "light-weight" host bridge in the same PCI domain 7 as the main host bridge whose purpose is to enable 8 the main host bridge to support multiple PCI root buses. 14 (different from the main host bridge) allowing the guest OS 24 -object memory-backend-ram,size=1024M,policy=bind,host-nodes=0,id=ram-node0 -numa node,nodeid=0,cpu… 25 -object memory-backend-ram,size=1024M,policy=bind,host-nodes=1,id=ram-node1 -numa node,nodeid=1,cpu… 31 …t, 0 and 1. (both mapped to the same NUMA node in host, but you can and should put it in different… 32 - a pxb host bridge attached to NUMA 1 with an e1000 behind it 33 - a pxb host bridge attached to NUMA 0 with an e1000 behind it 34 - a pxb host bridge not attached to any NUMA with a hard drive behind it. [all …]
|
H A D | igd-assign.txt | 24 | #5 Host's VGA ranges are mapped to IGD | | | * | | 28 use the VBT copied from host OpRegion to prevent incorrect configuration. 30 For #1, the "x-igd-opregion=on" option exposes a copy of host IGD OpRegion to 33 For #2, "x-igd-lpc=on" option copies the IDs of host LPC bridge and host bridge 40 For #4, the primary display must be set to IGD in host BIOS. 69 For either mode, depending on the host kernel, the i915 driver in the host 72 such driver binding unless the host driver is known to work well for this. 73 There are numerous ways to do this, i915 can be blacklisted on the host, 75 to the device on the host[2], virsh nodedev-detach can be used to bind the 78 graphics in the host and special options may be required beyond simply [all …]
|
/qemu/util/ |
H A D | vfio-helpers.c | 39 void *host; member 464 static void qemu_vfio_ram_block_added(RAMBlockNotifier *n, void *host, in qemu_vfio_ram_block_added() argument 471 trace_qemu_vfio_ram_block_added(s, host, max_size); in qemu_vfio_ram_block_added() 472 ret = qemu_vfio_dma_map(s, host, max_size, false, NULL, &local_err); in qemu_vfio_ram_block_added() 476 host, max_size); in qemu_vfio_ram_block_added() 480 static void qemu_vfio_ram_block_removed(RAMBlockNotifier *n, void *host, in qemu_vfio_ram_block_removed() argument 484 if (host) { in qemu_vfio_ram_block_removed() 485 trace_qemu_vfio_ram_block_removed(s, host, max_size); in qemu_vfio_ram_block_removed() 486 qemu_vfio_dma_unmap(s, host); in qemu_vfio_ram_block_removed() 532 trace_qemu_vfio_dump_mapping(s->mappings[i].host, in qemu_vfio_dump_mappings() [all …]
|
H A D | trace-events | 86 qemu_vfio_ram_block_added(void *s, void *p, size_t size) "s %p host %p size 0x%zx" 87 qemu_vfio_ram_block_removed(void *s, void *p, size_t size) "s %p host %p size 0x%zx" 88 qemu_vfio_dump_mapping(void *host, uint64_t iova, size_t size) "vfio mapping %p to iova 0x%08" PRIx… 89 qemu_vfio_find_mapping(void *s, void *p) "s %p host %p" 90 qemu_vfio_new_mapping(void *s, void *host, size_t size, int index, uint64_t iova) "s %p host %p siz… 91 qemu_vfio_do_mapping(void *s, void *host, uint64_t iova, size_t size) "s %p host %p <-> iova 0x%"PR… 92 qemu_vfio_dma_map(void *s, void *host, size_t size, bool temporary, uint64_t *iova) "s %p host %p s… 93 qemu_vfio_dma_mapped(void *s, void *host, uint64_t iova, size_t size) "s %p host %p <-> iova 0x%"PR… 94 qemu_vfio_dma_unmap(void *s, void *host) "s %p host %p" 98 …s, uint64_t region_size, int ofs, void *host) "map region bar#%d addr 0x%"PRIx64" size 0x%"PRIx64"…
|
/qemu/include/user/ |
H A D | safe-syscall.h | 31 * (Errnos are host errnos; we rely on QEMU_ERESTARTSYS not clashing 32 * with any of the host errno values.) 68 * Here it is important that the host syscall is made 69 * via this safe_syscall() function, and *not* via the host libc. 70 * If the host libc is used then the implementation will appear to work 72 * signal could arrive just before we make the host syscall inside libc, 79 * (If the implementation needs to make multiple host syscalls this is 82 * stay in the host kernel indefinitely) it's OK to use libc if necessary. 92 * The basic setup is that we make the host syscall via a known 93 * section of host native assembly. If a signal occurs, our signal [all …]
|
/qemu/include/hw/hyperv/ |
H A D | dynmem-proto.h | 163 * Version negotiation message. Sent from the guest to the host. 164 * The guest is free to try different versions until the host 180 * Version response message; Host to Guest and indicates 181 * if the host has accepted the version sent by the guest. 183 * is_accepted: If TRUE, host has accepted the version and the guest 198 * host. 209 * Response to the capabilities message. This is sent from the host to the 210 * guest. This message notifies if the host has accepted the guest's 211 * capabilities. If the host has not accepted, the guest must shutdown 214 * is_accepted: Indicates if the host has accepted guest's capabilities. [all …]
|
/qemu/docs/system/devices/ |
H A D | can.rst | 6 can be connected to host system CAN API (at this time only Linux 43 Add "can-host-socketcan" object to connect device to host system CAN bus:: 45 -object can-host-socketcan,id=canhost0,if=can0,canbus=canbus0 64 The tested setup was Linux 4.9 kernel on the host and guest side. 72 -object can-host-socketcan,id=canhost0,if=can0,canbus=canbus0 \ 85 -object can-host-socketcan,id=canhost0,if=can0,canbus=canbus0 \ 86 -device kvaser_pci,canbus=canbus0,host=can0 \ 88 The CAN interface of the host system has to be configured for proper 90 devices through bus to the physical host device. Example configuration 96 Virtual (host local only) can interface can be used on the host [all …]
|
H A D | usb.rst | 5 plug virtual USB devices or real host USB devices (only works with 6 certain host operating systems). QEMU will automatically create and 15 QEMU has XHCI host adapter support. The XHCI hardware design is much 160 ``usb-host,hostbus=bus,hostaddr=addr`` 161 Pass through the host device identified by bus and addr 163 ``usb-host,vendorid=vendor,productid=product`` 164 Pass through the host device identified by vendor and product ID 176 host character device id. 265 Using host USB devices on a Linux host 308 device_add usb-host,vendorid=0x1234,productid=0x5678 [all …]
|
H A D | cxl.rst | 3 From the view of a single host, CXL is an interconnect standard that 4 targets accelerators and memory devices attached to a CXL host. 6 software running on a QEMU emulated host or to the internals of 11 by considering only a single host and a static configuration. 14 with CXL Host Bridges, which have CXL Root Ports which may be directly 37 supported by the host for normal memory should also work for 49 **Type 1:** These support coherent caching of host memory. Example might 53 **Type 2:** These support coherent caching of host memory and host 55 by the host. This is a complex topic, so for more information on CXL 59 additional memory (HDM) to a CXL host including both volatile and [all …]
|
/qemu/tests/functional/ |
H A D | test_aarch64_device_passthrough.py | 41 # boot nested guest and execute /host/nested_guest.sh 47 -cpu host \ 50 -kernel /host/Image.gz \ 51 -drive format=raw,file=/host/guest.ext4,if=virtio \ 52 -append "root=/dev/vda init=/init -- bash /host/nested_guest.sh" \ 53 -virtfs local,path=/host,mount_tag=host,security_model=mapped,readonly=off \ 54 -device vfio-pci,host=$pci_vfio \ 56 -device vfio-pci,host=$pci_iommufd,iommufd=iommufd0 65 image_vfio=/host/disk_vfio 66 image_iommufd=/host/disk_iommufd [all …]
|
/qemu/tcg/tci/ |
H A D | README | 10 can be run on a host. 13 s390, sparc, x86_64). For others, unofficial host support was written. 17 support (almost) any host. 23 Like each TCG host frontend, TCI implements the code generator in 69 TCI needs special implementation for 32 and 64 bit host, 32 and 64 bit target, 70 host and target with same or different endianness. 72 | host (le) host (be) 103 * TCI is not widely tested. It was written and tested on a x86_64 host 105 A cross compiled QEMU for i386 host also works with the same basic tests. 106 A cross compiled QEMU for mipsel host works, too. It is terribly slow [all …]
|
/qemu/target/ppc/ |
H A D | mem_helper.c | 23 #include "qemu/host-utils.h" 78 /* If the two host pages are contiguous, optimize. */ in probe_contiguous() 89 void *host = probe_contiguous(env, addr, (32 - reg) * 4, in helper_lmw() local 92 if (likely(host)) { in helper_lmw() 93 /* Fast path -- the entire operation is in RAM at host. */ in helper_lmw() 95 env->gpr[reg] = (uint32_t)ldl_be_p(host); in helper_lmw() 96 host += 4; in helper_lmw() 111 void *host = probe_contiguous(env, addr, (32 - reg) * 4, in helper_stmw() local 114 if (likely(host)) { in helper_stmw() 115 /* Fast path -- the entire operation is in RAM at host. */ in helper_stmw() [all …]
|
/qemu/hw/pci-host/ |
H A D | remote.c | 2 * Remote PCI host device 4 * Unlike PCI host devices that model physical hardware, the purpose 5 * of this PCI host is to host multi-process QEMU devices. 7 * Multi-process QEMU extends the PCI host of a QEMU machine into a 12 * This PCI host is purely a container for PCI devices. It's fake in the 13 * sense that the guest never sees this PCI host and has no way of 30 #include "hw/pci-host/remote.h"
|
/qemu/docs/system/s390x/ |
H A D | vfio-ccw.rst | 6 host will not interact with those subchannels/devices any more. 14 Step 1: configure the host device 19 [root@host ~]# uuidgen 23 the host device. 30 [root@host ~]# driverctl -b css set-override 0.0.0313 vfio_ccw 31 [root@host ~]# mdevctl define -u 7e270a25-e163-4922-af60-757fc8ed48c6 \ 39 [root@host ~]# lscss | grep 0.0.2b09 | awk '{print $2}' 45 [root@host ~]# echo 0.0.0313 > /sys/bus/css/devices/0.0.0313/driver/unbind 46 [root@host ~]# echo 0.0.0313 > /sys/bus/css/drivers/vfio_ccw/bind 50 [root@host ~]# echo "7e270a25-e163-4922-af60-757fc8ed48c6" > \
|
/qemu/accel/tcg/ |
H A D | ldst_atomicity.c.inc | 12 #include "host/load-extract-al16-al8.h.inc" 13 #include "host/store-insert-al16.h.inc" 93 * host atomicity in order to avoid racing. This reduction 104 * @pv: host address 116 * @pv: host address 128 * @pv: host address 143 * @ra: host unwind address 144 * @pv: host address 178 * @ra: host unwind address 179 * @pv: host address [all …]
|
/qemu/tests/qtest/libqos/ |
H A D | ppc64_pseries-machine.c | 44 QSPAPR_pci_host *host = obj; in QSPAPR_host_get_device() local 46 return &host->pci.obj; in QSPAPR_host_get_device() 52 static void qos_create_QSPAPR_host(QSPAPR_pci_host *host, in qos_create_QSPAPR_host() argument 56 host->obj.get_device = QSPAPR_host_get_device; in qos_create_QSPAPR_host() 57 qpci_init_spapr(&host->pci, qts, alloc); in qos_create_QSPAPR_host() 82 if (!g_strcmp0(device, "spapr-pci-host-bridge")) { in spapr_get_device() 106 qos_node_create_driver("spapr-pci-host-bridge", NULL); in spapr_machine_register_nodes() 107 qos_node_contains("ppc64/pseries", "spapr-pci-host-bridge", NULL); in spapr_machine_register_nodes() 108 qos_node_contains("spapr-pci-host-bridge", "pci-bus-spapr", NULL); in spapr_machine_register_nodes()
|
/qemu/target/i386/ |
H A D | host-cpu.c | 2 * x86 host CPU functions, and "host" cpu type initialization 12 #include "host-cpu.h" 52 * host value. in host_cpu_adjust_phys_bits() 55 warn_report_once("Host physical bits (%u)" in host_cpu_adjust_phys_bits() 61 /* The user asked for us to use the host physical bits */ in host_cpu_adjust_phys_bits() 86 * Get CPUID model ID string from host CPU. 149 /* Use max host physical address bits if -cpu max option is applied */ in host_cpu_max_instance_init() 150 object_property_set_bool(OBJECT(cpu), "host-phys-bits", true, &error_abort); in host_cpu_max_instance_init() 171 g_strdup_printf("processor with all supported host features "); in host_cpu_class_init() 175 .name = X86_CPU_TYPE_NAME("host"),
|