1da9bf531SAlexander Bulekov /* 2da9bf531SAlexander Bulekov * Generic Virtual-Device Fuzzing Target 3da9bf531SAlexander Bulekov * 4da9bf531SAlexander Bulekov * Copyright Red Hat Inc., 2020 5da9bf531SAlexander Bulekov * 6da9bf531SAlexander Bulekov * Authors: 7da9bf531SAlexander Bulekov * Alexander Bulekov <alxndr@bu.edu> 8da9bf531SAlexander Bulekov * 9da9bf531SAlexander Bulekov * This work is licensed under the terms of the GNU GPL, version 2 or later. 10da9bf531SAlexander Bulekov * See the COPYING file in the top-level directory. 11da9bf531SAlexander Bulekov */ 12da9bf531SAlexander Bulekov 13da9bf531SAlexander Bulekov #include "qemu/osdep.h" 14da9bf531SAlexander Bulekov 15da9bf531SAlexander Bulekov #include <wordexp.h> 16da9bf531SAlexander Bulekov 17da9bf531SAlexander Bulekov #include "hw/core/cpu.h" 18da9bf531SAlexander Bulekov #include "tests/qtest/libqos/libqtest.h" 19da9bf531SAlexander Bulekov #include "fuzz.h" 20da9bf531SAlexander Bulekov #include "fork_fuzz.h" 21da9bf531SAlexander Bulekov #include "exec/address-spaces.h" 22da9bf531SAlexander Bulekov #include "string.h" 23da9bf531SAlexander Bulekov #include "exec/memory.h" 24da9bf531SAlexander Bulekov #include "exec/ramblock.h" 25da9bf531SAlexander Bulekov #include "exec/address-spaces.h" 26da9bf531SAlexander Bulekov #include "hw/qdev-core.h" 2705efbf24SAlexander Bulekov #include "hw/pci/pci.h" 2820f5a302SAlexander Bulekov #include "hw/boards.h" 29da9bf531SAlexander Bulekov 30da9bf531SAlexander Bulekov /* 31da9bf531SAlexander Bulekov * SEPARATOR is used to separate "operations" in the fuzz input 32da9bf531SAlexander Bulekov */ 33da9bf531SAlexander Bulekov #define SEPARATOR "FUZZ" 34da9bf531SAlexander Bulekov 35da9bf531SAlexander Bulekov enum cmds { 36da9bf531SAlexander Bulekov OP_IN, 37da9bf531SAlexander Bulekov OP_OUT, 38da9bf531SAlexander Bulekov OP_READ, 39da9bf531SAlexander Bulekov OP_WRITE, 4005efbf24SAlexander Bulekov OP_PCI_READ, 4105efbf24SAlexander Bulekov OP_PCI_WRITE, 42*ccbd4bc8SAlexander Bulekov OP_DISABLE_PCI, 4320f5a302SAlexander Bulekov OP_ADD_DMA_PATTERN, 4420f5a302SAlexander Bulekov OP_CLEAR_DMA_PATTERNS, 45da9bf531SAlexander Bulekov OP_CLOCK_STEP, 46da9bf531SAlexander Bulekov }; 47da9bf531SAlexander Bulekov 48da9bf531SAlexander Bulekov #define DEFAULT_TIMEOUT_US 100000 49da9bf531SAlexander Bulekov #define USEC_IN_SEC 1000000000 50da9bf531SAlexander Bulekov 5120f5a302SAlexander Bulekov #define MAX_DMA_FILL_SIZE 0x10000 5220f5a302SAlexander Bulekov 5305efbf24SAlexander Bulekov #define PCI_HOST_BRIDGE_CFG 0xcf8 5405efbf24SAlexander Bulekov #define PCI_HOST_BRIDGE_DATA 0xcfc 5505efbf24SAlexander Bulekov 56da9bf531SAlexander Bulekov typedef struct { 57da9bf531SAlexander Bulekov ram_addr_t addr; 58da9bf531SAlexander Bulekov ram_addr_t size; /* The number of bytes until the end of the I/O region */ 59da9bf531SAlexander Bulekov } address_range; 60da9bf531SAlexander Bulekov 61da9bf531SAlexander Bulekov static useconds_t timeout = DEFAULT_TIMEOUT_US; 62da9bf531SAlexander Bulekov 63da9bf531SAlexander Bulekov static bool qtest_log_enabled; 64da9bf531SAlexander Bulekov 65da9bf531SAlexander Bulekov /* 6620f5a302SAlexander Bulekov * A pattern used to populate a DMA region or perform a memwrite. This is 6720f5a302SAlexander Bulekov * useful for e.g. populating tables of unique addresses. 6820f5a302SAlexander Bulekov * Example {.index = 1; .stride = 2; .len = 3; .data = "\x00\x01\x02"} 6920f5a302SAlexander Bulekov * Renders as: 00 01 02 00 03 02 00 05 02 00 07 02 ... 7020f5a302SAlexander Bulekov */ 7120f5a302SAlexander Bulekov typedef struct { 7220f5a302SAlexander Bulekov uint8_t index; /* Index of a byte to increment by stride */ 7320f5a302SAlexander Bulekov uint8_t stride; /* Increment each index'th byte by this amount */ 7420f5a302SAlexander Bulekov size_t len; 7520f5a302SAlexander Bulekov const uint8_t *data; 7620f5a302SAlexander Bulekov } pattern; 7720f5a302SAlexander Bulekov 7820f5a302SAlexander Bulekov /* Avoid filling the same DMA region between MMIO/PIO commands ? */ 7920f5a302SAlexander Bulekov static bool avoid_double_fetches; 8020f5a302SAlexander Bulekov 8120f5a302SAlexander Bulekov static QTestState *qts_global; /* Need a global for the DMA callback */ 8220f5a302SAlexander Bulekov 8320f5a302SAlexander Bulekov /* 84da9bf531SAlexander Bulekov * List of memory regions that are children of QOM objects specified by the 85da9bf531SAlexander Bulekov * user for fuzzing. 86da9bf531SAlexander Bulekov */ 87da9bf531SAlexander Bulekov static GHashTable *fuzzable_memoryregions; 8805efbf24SAlexander Bulekov static GPtrArray *fuzzable_pci_devices; 89da9bf531SAlexander Bulekov 90da9bf531SAlexander Bulekov struct get_io_cb_info { 91da9bf531SAlexander Bulekov int index; 92da9bf531SAlexander Bulekov int found; 93da9bf531SAlexander Bulekov address_range result; 94da9bf531SAlexander Bulekov }; 95da9bf531SAlexander Bulekov 96da9bf531SAlexander Bulekov static int get_io_address_cb(Int128 start, Int128 size, 97da9bf531SAlexander Bulekov const MemoryRegion *mr, void *opaque) { 98da9bf531SAlexander Bulekov struct get_io_cb_info *info = opaque; 99da9bf531SAlexander Bulekov if (g_hash_table_lookup(fuzzable_memoryregions, mr)) { 100da9bf531SAlexander Bulekov if (info->index == 0) { 101da9bf531SAlexander Bulekov info->result.addr = (ram_addr_t)start; 102da9bf531SAlexander Bulekov info->result.size = (ram_addr_t)size; 103da9bf531SAlexander Bulekov info->found = 1; 104da9bf531SAlexander Bulekov return 1; 105da9bf531SAlexander Bulekov } 106da9bf531SAlexander Bulekov info->index--; 107da9bf531SAlexander Bulekov } 108da9bf531SAlexander Bulekov return 0; 109da9bf531SAlexander Bulekov } 110da9bf531SAlexander Bulekov 111da9bf531SAlexander Bulekov /* 11220f5a302SAlexander Bulekov * List of dma regions populated since the last fuzzing command. Used to ensure 11320f5a302SAlexander Bulekov * that we only write to each DMA address once, to avoid race conditions when 11420f5a302SAlexander Bulekov * building reproducers. 11520f5a302SAlexander Bulekov */ 11620f5a302SAlexander Bulekov static GArray *dma_regions; 11720f5a302SAlexander Bulekov 11820f5a302SAlexander Bulekov static GArray *dma_patterns; 11920f5a302SAlexander Bulekov static int dma_pattern_index; 120*ccbd4bc8SAlexander Bulekov static bool pci_disabled; 12120f5a302SAlexander Bulekov 12220f5a302SAlexander Bulekov /* 12320f5a302SAlexander Bulekov * Allocate a block of memory and populate it with a pattern. 12420f5a302SAlexander Bulekov */ 12520f5a302SAlexander Bulekov static void *pattern_alloc(pattern p, size_t len) 12620f5a302SAlexander Bulekov { 12720f5a302SAlexander Bulekov int i; 12820f5a302SAlexander Bulekov uint8_t *buf = g_malloc(len); 12920f5a302SAlexander Bulekov uint8_t sum = 0; 13020f5a302SAlexander Bulekov 13120f5a302SAlexander Bulekov for (i = 0; i < len; ++i) { 13220f5a302SAlexander Bulekov buf[i] = p.data[i % p.len]; 13320f5a302SAlexander Bulekov if ((i % p.len) == p.index) { 13420f5a302SAlexander Bulekov buf[i] += sum; 13520f5a302SAlexander Bulekov sum += p.stride; 13620f5a302SAlexander Bulekov } 13720f5a302SAlexander Bulekov } 13820f5a302SAlexander Bulekov return buf; 13920f5a302SAlexander Bulekov } 14020f5a302SAlexander Bulekov 14120f5a302SAlexander Bulekov static int memory_access_size(MemoryRegion *mr, unsigned l, hwaddr addr) 14220f5a302SAlexander Bulekov { 14320f5a302SAlexander Bulekov unsigned access_size_max = mr->ops->valid.max_access_size; 14420f5a302SAlexander Bulekov 14520f5a302SAlexander Bulekov /* 14620f5a302SAlexander Bulekov * Regions are assumed to support 1-4 byte accesses unless 14720f5a302SAlexander Bulekov * otherwise specified. 14820f5a302SAlexander Bulekov */ 14920f5a302SAlexander Bulekov if (access_size_max == 0) { 15020f5a302SAlexander Bulekov access_size_max = 4; 15120f5a302SAlexander Bulekov } 15220f5a302SAlexander Bulekov 15320f5a302SAlexander Bulekov /* Bound the maximum access by the alignment of the address. */ 15420f5a302SAlexander Bulekov if (!mr->ops->impl.unaligned) { 15520f5a302SAlexander Bulekov unsigned align_size_max = addr & -addr; 15620f5a302SAlexander Bulekov if (align_size_max != 0 && align_size_max < access_size_max) { 15720f5a302SAlexander Bulekov access_size_max = align_size_max; 15820f5a302SAlexander Bulekov } 15920f5a302SAlexander Bulekov } 16020f5a302SAlexander Bulekov 16120f5a302SAlexander Bulekov /* Don't attempt accesses larger than the maximum. */ 16220f5a302SAlexander Bulekov if (l > access_size_max) { 16320f5a302SAlexander Bulekov l = access_size_max; 16420f5a302SAlexander Bulekov } 16520f5a302SAlexander Bulekov l = pow2floor(l); 16620f5a302SAlexander Bulekov 16720f5a302SAlexander Bulekov return l; 16820f5a302SAlexander Bulekov } 16920f5a302SAlexander Bulekov 17020f5a302SAlexander Bulekov /* 17120f5a302SAlexander Bulekov * Call-back for functions that perform DMA reads from guest memory. Confirm 17220f5a302SAlexander Bulekov * that the region has not already been populated since the last loop in 17320f5a302SAlexander Bulekov * generic_fuzz(), avoiding potential race-conditions, which we don't have 17420f5a302SAlexander Bulekov * a good way for reproducing right now. 17520f5a302SAlexander Bulekov */ 17620f5a302SAlexander Bulekov void fuzz_dma_read_cb(size_t addr, size_t len, MemoryRegion *mr, bool is_write) 17720f5a302SAlexander Bulekov { 17820f5a302SAlexander Bulekov /* Are we in the generic-fuzzer or are we using another fuzz-target? */ 17920f5a302SAlexander Bulekov if (!qts_global) { 18020f5a302SAlexander Bulekov return; 18120f5a302SAlexander Bulekov } 18220f5a302SAlexander Bulekov 18320f5a302SAlexander Bulekov /* 18420f5a302SAlexander Bulekov * Return immediately if: 18520f5a302SAlexander Bulekov * - We have no DMA patterns defined 18620f5a302SAlexander Bulekov * - The length of the DMA read request is zero 18720f5a302SAlexander Bulekov * - The DMA read is hitting an MR other than the machine's main RAM 18820f5a302SAlexander Bulekov * - The DMA request is not a read (what happens for a address_space_map 18920f5a302SAlexander Bulekov * with is_write=True? Can the device use the same pointer to do reads?) 19020f5a302SAlexander Bulekov * - The DMA request hits past the bounds of our RAM 19120f5a302SAlexander Bulekov */ 19220f5a302SAlexander Bulekov if (dma_patterns->len == 0 19320f5a302SAlexander Bulekov || len == 0 19420f5a302SAlexander Bulekov /* || mr != MACHINE(qdev_get_machine())->ram */ 19520f5a302SAlexander Bulekov || is_write 19620f5a302SAlexander Bulekov || addr > current_machine->ram_size) { 19720f5a302SAlexander Bulekov return; 19820f5a302SAlexander Bulekov } 19920f5a302SAlexander Bulekov 20020f5a302SAlexander Bulekov /* 20120f5a302SAlexander Bulekov * If we overlap with any existing dma_regions, split the range and only 20220f5a302SAlexander Bulekov * populate the non-overlapping parts. 20320f5a302SAlexander Bulekov */ 20420f5a302SAlexander Bulekov address_range region; 20520f5a302SAlexander Bulekov bool double_fetch = false; 20620f5a302SAlexander Bulekov for (int i = 0; 20720f5a302SAlexander Bulekov i < dma_regions->len && (avoid_double_fetches || qtest_log_enabled); 20820f5a302SAlexander Bulekov ++i) { 20920f5a302SAlexander Bulekov region = g_array_index(dma_regions, address_range, i); 21020f5a302SAlexander Bulekov if (addr < region.addr + region.size && addr + len > region.addr) { 21120f5a302SAlexander Bulekov double_fetch = true; 21220f5a302SAlexander Bulekov if (addr < region.addr 21320f5a302SAlexander Bulekov && avoid_double_fetches) { 21420f5a302SAlexander Bulekov fuzz_dma_read_cb(addr, region.addr - addr, mr, is_write); 21520f5a302SAlexander Bulekov } 21620f5a302SAlexander Bulekov if (addr + len > region.addr + region.size 21720f5a302SAlexander Bulekov && avoid_double_fetches) { 21820f5a302SAlexander Bulekov fuzz_dma_read_cb(region.addr + region.size, 21920f5a302SAlexander Bulekov addr + len - (region.addr + region.size), mr, is_write); 22020f5a302SAlexander Bulekov } 22120f5a302SAlexander Bulekov return; 22220f5a302SAlexander Bulekov } 22320f5a302SAlexander Bulekov } 22420f5a302SAlexander Bulekov 22520f5a302SAlexander Bulekov /* Cap the length of the DMA access to something reasonable */ 22620f5a302SAlexander Bulekov len = MIN(len, MAX_DMA_FILL_SIZE); 22720f5a302SAlexander Bulekov 22820f5a302SAlexander Bulekov address_range ar = {addr, len}; 22920f5a302SAlexander Bulekov g_array_append_val(dma_regions, ar); 23020f5a302SAlexander Bulekov pattern p = g_array_index(dma_patterns, pattern, dma_pattern_index); 23120f5a302SAlexander Bulekov void *buf = pattern_alloc(p, ar.size); 23220f5a302SAlexander Bulekov hwaddr l, addr1; 23320f5a302SAlexander Bulekov MemoryRegion *mr1; 23420f5a302SAlexander Bulekov uint8_t *ram_ptr; 23520f5a302SAlexander Bulekov while (len > 0) { 23620f5a302SAlexander Bulekov l = len; 23720f5a302SAlexander Bulekov mr1 = address_space_translate(first_cpu->as, 23820f5a302SAlexander Bulekov addr, &addr1, &l, true, 23920f5a302SAlexander Bulekov MEMTXATTRS_UNSPECIFIED); 24020f5a302SAlexander Bulekov 24120f5a302SAlexander Bulekov if (!(memory_region_is_ram(mr1) || 24220f5a302SAlexander Bulekov memory_region_is_romd(mr1))) { 24320f5a302SAlexander Bulekov l = memory_access_size(mr1, l, addr1); 24420f5a302SAlexander Bulekov } else { 24520f5a302SAlexander Bulekov /* ROM/RAM case */ 24620f5a302SAlexander Bulekov ram_ptr = qemu_map_ram_ptr(mr1->ram_block, addr1); 24720f5a302SAlexander Bulekov memcpy(ram_ptr, buf, l); 24820f5a302SAlexander Bulekov break; 24920f5a302SAlexander Bulekov } 25020f5a302SAlexander Bulekov len -= l; 25120f5a302SAlexander Bulekov buf += l; 25220f5a302SAlexander Bulekov addr += l; 25320f5a302SAlexander Bulekov 25420f5a302SAlexander Bulekov } 25520f5a302SAlexander Bulekov if (qtest_log_enabled) { 25620f5a302SAlexander Bulekov /* 25720f5a302SAlexander Bulekov * With QTEST_LOG, use a normal, slow QTest memwrite. Prefix the log 25820f5a302SAlexander Bulekov * that will be written by qtest.c with a DMA tag, so we can reorder 25920f5a302SAlexander Bulekov * the resulting QTest trace so the DMA fills precede the last PIO/MMIO 26020f5a302SAlexander Bulekov * command. 26120f5a302SAlexander Bulekov */ 26220f5a302SAlexander Bulekov fprintf(stderr, "[DMA] "); 26320f5a302SAlexander Bulekov if (double_fetch) { 26420f5a302SAlexander Bulekov fprintf(stderr, "[DOUBLE-FETCH] "); 26520f5a302SAlexander Bulekov } 26620f5a302SAlexander Bulekov fflush(stderr); 26720f5a302SAlexander Bulekov } 26820f5a302SAlexander Bulekov qtest_memwrite(qts_global, ar.addr, buf, ar.size); 26920f5a302SAlexander Bulekov g_free(buf); 27020f5a302SAlexander Bulekov 27120f5a302SAlexander Bulekov /* Increment the index of the pattern for the next DMA access */ 27220f5a302SAlexander Bulekov dma_pattern_index = (dma_pattern_index + 1) % dma_patterns->len; 27320f5a302SAlexander Bulekov } 27420f5a302SAlexander Bulekov 27520f5a302SAlexander Bulekov /* 276da9bf531SAlexander Bulekov * Here we want to convert a fuzzer-provided [io-region-index, offset] to 277da9bf531SAlexander Bulekov * a physical address. To do this, we iterate over all of the matched 278da9bf531SAlexander Bulekov * MemoryRegions. Check whether each region exists within the particular io 279da9bf531SAlexander Bulekov * space. Return the absolute address of the offset within the index'th region 280da9bf531SAlexander Bulekov * that is a subregion of the io_space and the distance until the end of the 281da9bf531SAlexander Bulekov * memory region. 282da9bf531SAlexander Bulekov */ 283da9bf531SAlexander Bulekov static bool get_io_address(address_range *result, AddressSpace *as, 284da9bf531SAlexander Bulekov uint8_t index, 285da9bf531SAlexander Bulekov uint32_t offset) { 286da9bf531SAlexander Bulekov FlatView *view; 287da9bf531SAlexander Bulekov view = as->current_map; 288da9bf531SAlexander Bulekov g_assert(view); 289da9bf531SAlexander Bulekov struct get_io_cb_info cb_info = {}; 290da9bf531SAlexander Bulekov 291da9bf531SAlexander Bulekov cb_info.index = index; 292da9bf531SAlexander Bulekov 293da9bf531SAlexander Bulekov /* 294da9bf531SAlexander Bulekov * Loop around the FlatView until we match "index" number of 295da9bf531SAlexander Bulekov * fuzzable_memoryregions, or until we know that there are no matching 296da9bf531SAlexander Bulekov * memory_regions. 297da9bf531SAlexander Bulekov */ 298da9bf531SAlexander Bulekov do { 299da9bf531SAlexander Bulekov flatview_for_each_range(view, get_io_address_cb , &cb_info); 300da9bf531SAlexander Bulekov } while (cb_info.index != index && !cb_info.found); 301da9bf531SAlexander Bulekov 302da9bf531SAlexander Bulekov *result = cb_info.result; 303da9bf531SAlexander Bulekov return cb_info.found; 304da9bf531SAlexander Bulekov } 305da9bf531SAlexander Bulekov 306da9bf531SAlexander Bulekov static bool get_pio_address(address_range *result, 307da9bf531SAlexander Bulekov uint8_t index, uint16_t offset) 308da9bf531SAlexander Bulekov { 309da9bf531SAlexander Bulekov /* 310da9bf531SAlexander Bulekov * PIO BARs can be set past the maximum port address (0xFFFF). Thus, result 311da9bf531SAlexander Bulekov * can contain an addr that extends past the PIO space. When we pass this 312da9bf531SAlexander Bulekov * address to qtest_in/qtest_out, it is cast to a uint16_t, so we might end 313da9bf531SAlexander Bulekov * up fuzzing a completely different MemoryRegion/Device. Therefore, check 314da9bf531SAlexander Bulekov * that the address here is within the PIO space limits. 315da9bf531SAlexander Bulekov */ 316da9bf531SAlexander Bulekov bool found = get_io_address(result, &address_space_io, index, offset); 317da9bf531SAlexander Bulekov return result->addr <= 0xFFFF ? found : false; 318da9bf531SAlexander Bulekov } 319da9bf531SAlexander Bulekov 320da9bf531SAlexander Bulekov static bool get_mmio_address(address_range *result, 321da9bf531SAlexander Bulekov uint8_t index, uint32_t offset) 322da9bf531SAlexander Bulekov { 323da9bf531SAlexander Bulekov return get_io_address(result, &address_space_memory, index, offset); 324da9bf531SAlexander Bulekov } 325da9bf531SAlexander Bulekov 326da9bf531SAlexander Bulekov static void op_in(QTestState *s, const unsigned char * data, size_t len) 327da9bf531SAlexander Bulekov { 328da9bf531SAlexander Bulekov enum Sizes {Byte, Word, Long, end_sizes}; 329da9bf531SAlexander Bulekov struct { 330da9bf531SAlexander Bulekov uint8_t size; 331da9bf531SAlexander Bulekov uint8_t base; 332da9bf531SAlexander Bulekov uint16_t offset; 333da9bf531SAlexander Bulekov } a; 334da9bf531SAlexander Bulekov address_range abs; 335da9bf531SAlexander Bulekov 336da9bf531SAlexander Bulekov if (len < sizeof(a)) { 337da9bf531SAlexander Bulekov return; 338da9bf531SAlexander Bulekov } 339da9bf531SAlexander Bulekov memcpy(&a, data, sizeof(a)); 340da9bf531SAlexander Bulekov if (get_pio_address(&abs, a.base, a.offset) == 0) { 341da9bf531SAlexander Bulekov return; 342da9bf531SAlexander Bulekov } 343da9bf531SAlexander Bulekov 344da9bf531SAlexander Bulekov switch (a.size %= end_sizes) { 345da9bf531SAlexander Bulekov case Byte: 346da9bf531SAlexander Bulekov qtest_inb(s, abs.addr); 347da9bf531SAlexander Bulekov break; 348da9bf531SAlexander Bulekov case Word: 349da9bf531SAlexander Bulekov if (abs.size >= 2) { 350da9bf531SAlexander Bulekov qtest_inw(s, abs.addr); 351da9bf531SAlexander Bulekov } 352da9bf531SAlexander Bulekov break; 353da9bf531SAlexander Bulekov case Long: 354da9bf531SAlexander Bulekov if (abs.size >= 4) { 355da9bf531SAlexander Bulekov qtest_inl(s, abs.addr); 356da9bf531SAlexander Bulekov } 357da9bf531SAlexander Bulekov break; 358da9bf531SAlexander Bulekov } 359da9bf531SAlexander Bulekov } 360da9bf531SAlexander Bulekov 361da9bf531SAlexander Bulekov static void op_out(QTestState *s, const unsigned char * data, size_t len) 362da9bf531SAlexander Bulekov { 363da9bf531SAlexander Bulekov enum Sizes {Byte, Word, Long, end_sizes}; 364da9bf531SAlexander Bulekov struct { 365da9bf531SAlexander Bulekov uint8_t size; 366da9bf531SAlexander Bulekov uint8_t base; 367da9bf531SAlexander Bulekov uint16_t offset; 368da9bf531SAlexander Bulekov uint32_t value; 369da9bf531SAlexander Bulekov } a; 370da9bf531SAlexander Bulekov address_range abs; 371da9bf531SAlexander Bulekov 372da9bf531SAlexander Bulekov if (len < sizeof(a)) { 373da9bf531SAlexander Bulekov return; 374da9bf531SAlexander Bulekov } 375da9bf531SAlexander Bulekov memcpy(&a, data, sizeof(a)); 376da9bf531SAlexander Bulekov 377da9bf531SAlexander Bulekov if (get_pio_address(&abs, a.base, a.offset) == 0) { 378da9bf531SAlexander Bulekov return; 379da9bf531SAlexander Bulekov } 380da9bf531SAlexander Bulekov 381da9bf531SAlexander Bulekov switch (a.size %= end_sizes) { 382da9bf531SAlexander Bulekov case Byte: 383da9bf531SAlexander Bulekov qtest_outb(s, abs.addr, a.value & 0xFF); 384da9bf531SAlexander Bulekov break; 385da9bf531SAlexander Bulekov case Word: 386da9bf531SAlexander Bulekov if (abs.size >= 2) { 387da9bf531SAlexander Bulekov qtest_outw(s, abs.addr, a.value & 0xFFFF); 388da9bf531SAlexander Bulekov } 389da9bf531SAlexander Bulekov break; 390da9bf531SAlexander Bulekov case Long: 391da9bf531SAlexander Bulekov if (abs.size >= 4) { 392da9bf531SAlexander Bulekov qtest_outl(s, abs.addr, a.value); 393da9bf531SAlexander Bulekov } 394da9bf531SAlexander Bulekov break; 395da9bf531SAlexander Bulekov } 396da9bf531SAlexander Bulekov } 397da9bf531SAlexander Bulekov 398da9bf531SAlexander Bulekov static void op_read(QTestState *s, const unsigned char * data, size_t len) 399da9bf531SAlexander Bulekov { 400da9bf531SAlexander Bulekov enum Sizes {Byte, Word, Long, Quad, end_sizes}; 401da9bf531SAlexander Bulekov struct { 402da9bf531SAlexander Bulekov uint8_t size; 403da9bf531SAlexander Bulekov uint8_t base; 404da9bf531SAlexander Bulekov uint32_t offset; 405da9bf531SAlexander Bulekov } a; 406da9bf531SAlexander Bulekov address_range abs; 407da9bf531SAlexander Bulekov 408da9bf531SAlexander Bulekov if (len < sizeof(a)) { 409da9bf531SAlexander Bulekov return; 410da9bf531SAlexander Bulekov } 411da9bf531SAlexander Bulekov memcpy(&a, data, sizeof(a)); 412da9bf531SAlexander Bulekov 413da9bf531SAlexander Bulekov if (get_mmio_address(&abs, a.base, a.offset) == 0) { 414da9bf531SAlexander Bulekov return; 415da9bf531SAlexander Bulekov } 416da9bf531SAlexander Bulekov 417da9bf531SAlexander Bulekov switch (a.size %= end_sizes) { 418da9bf531SAlexander Bulekov case Byte: 419da9bf531SAlexander Bulekov qtest_readb(s, abs.addr); 420da9bf531SAlexander Bulekov break; 421da9bf531SAlexander Bulekov case Word: 422da9bf531SAlexander Bulekov if (abs.size >= 2) { 423da9bf531SAlexander Bulekov qtest_readw(s, abs.addr); 424da9bf531SAlexander Bulekov } 425da9bf531SAlexander Bulekov break; 426da9bf531SAlexander Bulekov case Long: 427da9bf531SAlexander Bulekov if (abs.size >= 4) { 428da9bf531SAlexander Bulekov qtest_readl(s, abs.addr); 429da9bf531SAlexander Bulekov } 430da9bf531SAlexander Bulekov break; 431da9bf531SAlexander Bulekov case Quad: 432da9bf531SAlexander Bulekov if (abs.size >= 8) { 433da9bf531SAlexander Bulekov qtest_readq(s, abs.addr); 434da9bf531SAlexander Bulekov } 435da9bf531SAlexander Bulekov break; 436da9bf531SAlexander Bulekov } 437da9bf531SAlexander Bulekov } 438da9bf531SAlexander Bulekov 439da9bf531SAlexander Bulekov static void op_write(QTestState *s, const unsigned char * data, size_t len) 440da9bf531SAlexander Bulekov { 441da9bf531SAlexander Bulekov enum Sizes {Byte, Word, Long, Quad, end_sizes}; 442da9bf531SAlexander Bulekov struct { 443da9bf531SAlexander Bulekov uint8_t size; 444da9bf531SAlexander Bulekov uint8_t base; 445da9bf531SAlexander Bulekov uint32_t offset; 446da9bf531SAlexander Bulekov uint64_t value; 447da9bf531SAlexander Bulekov } a; 448da9bf531SAlexander Bulekov address_range abs; 449da9bf531SAlexander Bulekov 450da9bf531SAlexander Bulekov if (len < sizeof(a)) { 451da9bf531SAlexander Bulekov return; 452da9bf531SAlexander Bulekov } 453da9bf531SAlexander Bulekov memcpy(&a, data, sizeof(a)); 454da9bf531SAlexander Bulekov 455da9bf531SAlexander Bulekov if (get_mmio_address(&abs, a.base, a.offset) == 0) { 456da9bf531SAlexander Bulekov return; 457da9bf531SAlexander Bulekov } 458da9bf531SAlexander Bulekov 459da9bf531SAlexander Bulekov switch (a.size %= end_sizes) { 460da9bf531SAlexander Bulekov case Byte: 461da9bf531SAlexander Bulekov qtest_writeb(s, abs.addr, a.value & 0xFF); 462da9bf531SAlexander Bulekov break; 463da9bf531SAlexander Bulekov case Word: 464da9bf531SAlexander Bulekov if (abs.size >= 2) { 465da9bf531SAlexander Bulekov qtest_writew(s, abs.addr, a.value & 0xFFFF); 466da9bf531SAlexander Bulekov } 467da9bf531SAlexander Bulekov break; 468da9bf531SAlexander Bulekov case Long: 469da9bf531SAlexander Bulekov if (abs.size >= 4) { 470da9bf531SAlexander Bulekov qtest_writel(s, abs.addr, a.value & 0xFFFFFFFF); 471da9bf531SAlexander Bulekov } 472da9bf531SAlexander Bulekov break; 473da9bf531SAlexander Bulekov case Quad: 474da9bf531SAlexander Bulekov if (abs.size >= 8) { 475da9bf531SAlexander Bulekov qtest_writeq(s, abs.addr, a.value); 476da9bf531SAlexander Bulekov } 477da9bf531SAlexander Bulekov break; 478da9bf531SAlexander Bulekov } 479da9bf531SAlexander Bulekov } 480da9bf531SAlexander Bulekov 48105efbf24SAlexander Bulekov static void op_pci_read(QTestState *s, const unsigned char * data, size_t len) 48205efbf24SAlexander Bulekov { 48305efbf24SAlexander Bulekov enum Sizes {Byte, Word, Long, end_sizes}; 48405efbf24SAlexander Bulekov struct { 48505efbf24SAlexander Bulekov uint8_t size; 48605efbf24SAlexander Bulekov uint8_t base; 48705efbf24SAlexander Bulekov uint8_t offset; 48805efbf24SAlexander Bulekov } a; 489*ccbd4bc8SAlexander Bulekov if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) { 49005efbf24SAlexander Bulekov return; 49105efbf24SAlexander Bulekov } 49205efbf24SAlexander Bulekov memcpy(&a, data, sizeof(a)); 49305efbf24SAlexander Bulekov PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices, 49405efbf24SAlexander Bulekov a.base % fuzzable_pci_devices->len); 49505efbf24SAlexander Bulekov int devfn = dev->devfn; 49605efbf24SAlexander Bulekov qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset); 49705efbf24SAlexander Bulekov switch (a.size %= end_sizes) { 49805efbf24SAlexander Bulekov case Byte: 49905efbf24SAlexander Bulekov qtest_inb(s, PCI_HOST_BRIDGE_DATA); 50005efbf24SAlexander Bulekov break; 50105efbf24SAlexander Bulekov case Word: 50205efbf24SAlexander Bulekov qtest_inw(s, PCI_HOST_BRIDGE_DATA); 50305efbf24SAlexander Bulekov break; 50405efbf24SAlexander Bulekov case Long: 50505efbf24SAlexander Bulekov qtest_inl(s, PCI_HOST_BRIDGE_DATA); 50605efbf24SAlexander Bulekov break; 50705efbf24SAlexander Bulekov } 50805efbf24SAlexander Bulekov } 50905efbf24SAlexander Bulekov 51005efbf24SAlexander Bulekov static void op_pci_write(QTestState *s, const unsigned char * data, size_t len) 51105efbf24SAlexander Bulekov { 51205efbf24SAlexander Bulekov enum Sizes {Byte, Word, Long, end_sizes}; 51305efbf24SAlexander Bulekov struct { 51405efbf24SAlexander Bulekov uint8_t size; 51505efbf24SAlexander Bulekov uint8_t base; 51605efbf24SAlexander Bulekov uint8_t offset; 51705efbf24SAlexander Bulekov uint32_t value; 51805efbf24SAlexander Bulekov } a; 519*ccbd4bc8SAlexander Bulekov if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) { 52005efbf24SAlexander Bulekov return; 52105efbf24SAlexander Bulekov } 52205efbf24SAlexander Bulekov memcpy(&a, data, sizeof(a)); 52305efbf24SAlexander Bulekov PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices, 52405efbf24SAlexander Bulekov a.base % fuzzable_pci_devices->len); 52505efbf24SAlexander Bulekov int devfn = dev->devfn; 52605efbf24SAlexander Bulekov qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset); 52705efbf24SAlexander Bulekov switch (a.size %= end_sizes) { 52805efbf24SAlexander Bulekov case Byte: 52905efbf24SAlexander Bulekov qtest_outb(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFF); 53005efbf24SAlexander Bulekov break; 53105efbf24SAlexander Bulekov case Word: 53205efbf24SAlexander Bulekov qtest_outw(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFF); 53305efbf24SAlexander Bulekov break; 53405efbf24SAlexander Bulekov case Long: 53505efbf24SAlexander Bulekov qtest_outl(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFFFFFF); 53605efbf24SAlexander Bulekov break; 53705efbf24SAlexander Bulekov } 53805efbf24SAlexander Bulekov } 53905efbf24SAlexander Bulekov 54020f5a302SAlexander Bulekov static void op_add_dma_pattern(QTestState *s, 54120f5a302SAlexander Bulekov const unsigned char *data, size_t len) 54220f5a302SAlexander Bulekov { 54320f5a302SAlexander Bulekov struct { 54420f5a302SAlexander Bulekov /* 54520f5a302SAlexander Bulekov * index and stride can be used to increment the index-th byte of the 54620f5a302SAlexander Bulekov * pattern by the value stride, for each loop of the pattern. 54720f5a302SAlexander Bulekov */ 54820f5a302SAlexander Bulekov uint8_t index; 54920f5a302SAlexander Bulekov uint8_t stride; 55020f5a302SAlexander Bulekov } a; 55120f5a302SAlexander Bulekov 55220f5a302SAlexander Bulekov if (len < sizeof(a) + 1) { 55320f5a302SAlexander Bulekov return; 55420f5a302SAlexander Bulekov } 55520f5a302SAlexander Bulekov memcpy(&a, data, sizeof(a)); 55620f5a302SAlexander Bulekov pattern p = {a.index, a.stride, len - sizeof(a), data + sizeof(a)}; 55720f5a302SAlexander Bulekov p.index = a.index % p.len; 55820f5a302SAlexander Bulekov g_array_append_val(dma_patterns, p); 55920f5a302SAlexander Bulekov return; 56020f5a302SAlexander Bulekov } 56120f5a302SAlexander Bulekov 56220f5a302SAlexander Bulekov static void op_clear_dma_patterns(QTestState *s, 56320f5a302SAlexander Bulekov const unsigned char *data, size_t len) 56420f5a302SAlexander Bulekov { 56520f5a302SAlexander Bulekov g_array_set_size(dma_patterns, 0); 56620f5a302SAlexander Bulekov dma_pattern_index = 0; 56720f5a302SAlexander Bulekov } 56820f5a302SAlexander Bulekov 569da9bf531SAlexander Bulekov static void op_clock_step(QTestState *s, const unsigned char *data, size_t len) 570da9bf531SAlexander Bulekov { 571da9bf531SAlexander Bulekov qtest_clock_step_next(s); 572da9bf531SAlexander Bulekov } 573da9bf531SAlexander Bulekov 574*ccbd4bc8SAlexander Bulekov static void op_disable_pci(QTestState *s, const unsigned char *data, size_t len) 575*ccbd4bc8SAlexander Bulekov { 576*ccbd4bc8SAlexander Bulekov pci_disabled = true; 577*ccbd4bc8SAlexander Bulekov } 578*ccbd4bc8SAlexander Bulekov 579da9bf531SAlexander Bulekov static void handle_timeout(int sig) 580da9bf531SAlexander Bulekov { 581da9bf531SAlexander Bulekov if (qtest_log_enabled) { 582da9bf531SAlexander Bulekov fprintf(stderr, "[Timeout]\n"); 583da9bf531SAlexander Bulekov fflush(stderr); 584da9bf531SAlexander Bulekov } 585da9bf531SAlexander Bulekov _Exit(0); 586da9bf531SAlexander Bulekov } 587da9bf531SAlexander Bulekov 588da9bf531SAlexander Bulekov /* 589da9bf531SAlexander Bulekov * Here, we interpret random bytes from the fuzzer, as a sequence of commands. 590da9bf531SAlexander Bulekov * Some commands can be variable-width, so we use a separator, SEPARATOR, to 591da9bf531SAlexander Bulekov * specify the boundaries between commands. SEPARATOR is used to separate 592da9bf531SAlexander Bulekov * "operations" in the fuzz input. Why use a separator, instead of just using 593da9bf531SAlexander Bulekov * the operations' length to identify operation boundaries? 594da9bf531SAlexander Bulekov * 1. This is a simple way to support variable-length operations 595da9bf531SAlexander Bulekov * 2. This adds "stability" to the input. 596da9bf531SAlexander Bulekov * For example take the input "AbBcgDefg", where there is no separator and 597da9bf531SAlexander Bulekov * Opcodes are capitalized. 598da9bf531SAlexander Bulekov * Simply, by removing the first byte, we end up with a very different 599da9bf531SAlexander Bulekov * sequence: 600da9bf531SAlexander Bulekov * BbcGdefg... 601da9bf531SAlexander Bulekov * By adding a separator, we avoid this problem: 602da9bf531SAlexander Bulekov * Ab SEP Bcg SEP Defg -> B SEP Bcg SEP Defg 603da9bf531SAlexander Bulekov * Since B uses two additional bytes as operands, the first "B" will be 604da9bf531SAlexander Bulekov * ignored. The fuzzer actively tries to reduce inputs, so such unused 605da9bf531SAlexander Bulekov * bytes are likely to be pruned, eventually. 606da9bf531SAlexander Bulekov * 607da9bf531SAlexander Bulekov * SEPARATOR is trivial for the fuzzer to discover when using ASan. Optionally, 608da9bf531SAlexander Bulekov * SEPARATOR can be manually specified as a dictionary value (see libfuzzer's 609da9bf531SAlexander Bulekov * -dict), though this should not be necessary. 610da9bf531SAlexander Bulekov * 611da9bf531SAlexander Bulekov * As a result, the stream of bytes is converted into a sequence of commands. 612da9bf531SAlexander Bulekov * In a simplified example where SEPARATOR is 0xFF: 613da9bf531SAlexander Bulekov * 00 01 02 FF 03 04 05 06 FF 01 FF ... 614da9bf531SAlexander Bulekov * becomes this sequence of commands: 615da9bf531SAlexander Bulekov * 00 01 02 -> op00 (0102) -> in (0102, 2) 616da9bf531SAlexander Bulekov * 03 04 05 06 -> op03 (040506) -> write (040506, 3) 617da9bf531SAlexander Bulekov * 01 -> op01 (-,0) -> out (-,0) 618da9bf531SAlexander Bulekov * ... 619da9bf531SAlexander Bulekov * 620da9bf531SAlexander Bulekov * Note here that it is the job of the individual opcode functions to check 621da9bf531SAlexander Bulekov * that enough data was provided. I.e. in the last command out (,0), out needs 622da9bf531SAlexander Bulekov * to check that there is not enough data provided to select an address/value 623da9bf531SAlexander Bulekov * for the operation. 624da9bf531SAlexander Bulekov */ 625da9bf531SAlexander Bulekov static void generic_fuzz(QTestState *s, const unsigned char *Data, size_t Size) 626da9bf531SAlexander Bulekov { 627da9bf531SAlexander Bulekov void (*ops[]) (QTestState *s, const unsigned char* , size_t) = { 628da9bf531SAlexander Bulekov [OP_IN] = op_in, 629da9bf531SAlexander Bulekov [OP_OUT] = op_out, 630da9bf531SAlexander Bulekov [OP_READ] = op_read, 631da9bf531SAlexander Bulekov [OP_WRITE] = op_write, 63205efbf24SAlexander Bulekov [OP_PCI_READ] = op_pci_read, 63305efbf24SAlexander Bulekov [OP_PCI_WRITE] = op_pci_write, 634*ccbd4bc8SAlexander Bulekov [OP_DISABLE_PCI] = op_disable_pci, 63520f5a302SAlexander Bulekov [OP_ADD_DMA_PATTERN] = op_add_dma_pattern, 63620f5a302SAlexander Bulekov [OP_CLEAR_DMA_PATTERNS] = op_clear_dma_patterns, 637da9bf531SAlexander Bulekov [OP_CLOCK_STEP] = op_clock_step, 638da9bf531SAlexander Bulekov }; 639da9bf531SAlexander Bulekov const unsigned char *cmd = Data; 640da9bf531SAlexander Bulekov const unsigned char *nextcmd; 641da9bf531SAlexander Bulekov size_t cmd_len; 642da9bf531SAlexander Bulekov uint8_t op; 643da9bf531SAlexander Bulekov 644da9bf531SAlexander Bulekov if (fork() == 0) { 645da9bf531SAlexander Bulekov /* 646da9bf531SAlexander Bulekov * Sometimes the fuzzer will find inputs that take quite a long time to 647da9bf531SAlexander Bulekov * process. Often times, these inputs do not result in new coverage. 648da9bf531SAlexander Bulekov * Even if these inputs might be interesting, they can slow down the 649da9bf531SAlexander Bulekov * fuzzer, overall. Set a timeout to avoid hurting performance, too much 650da9bf531SAlexander Bulekov */ 651da9bf531SAlexander Bulekov if (timeout) { 652da9bf531SAlexander Bulekov struct sigaction sact; 653da9bf531SAlexander Bulekov struct itimerval timer; 654da9bf531SAlexander Bulekov 655da9bf531SAlexander Bulekov sigemptyset(&sact.sa_mask); 656da9bf531SAlexander Bulekov sact.sa_flags = SA_NODEFER; 657da9bf531SAlexander Bulekov sact.sa_handler = handle_timeout; 658da9bf531SAlexander Bulekov sigaction(SIGALRM, &sact, NULL); 659da9bf531SAlexander Bulekov 660da9bf531SAlexander Bulekov memset(&timer, 0, sizeof(timer)); 661da9bf531SAlexander Bulekov timer.it_value.tv_sec = timeout / USEC_IN_SEC; 662da9bf531SAlexander Bulekov timer.it_value.tv_usec = timeout % USEC_IN_SEC; 663da9bf531SAlexander Bulekov setitimer(ITIMER_VIRTUAL, &timer, NULL); 664da9bf531SAlexander Bulekov } 665da9bf531SAlexander Bulekov 66620f5a302SAlexander Bulekov op_clear_dma_patterns(s, NULL, 0); 667*ccbd4bc8SAlexander Bulekov pci_disabled = false; 66820f5a302SAlexander Bulekov 669da9bf531SAlexander Bulekov while (cmd && Size) { 670da9bf531SAlexander Bulekov /* Get the length until the next command or end of input */ 671da9bf531SAlexander Bulekov nextcmd = memmem(cmd, Size, SEPARATOR, strlen(SEPARATOR)); 672da9bf531SAlexander Bulekov cmd_len = nextcmd ? nextcmd - cmd : Size; 673da9bf531SAlexander Bulekov 674da9bf531SAlexander Bulekov if (cmd_len > 0) { 675da9bf531SAlexander Bulekov /* Interpret the first byte of the command as an opcode */ 676da9bf531SAlexander Bulekov op = *cmd % (sizeof(ops) / sizeof((ops)[0])); 677da9bf531SAlexander Bulekov ops[op](s, cmd + 1, cmd_len - 1); 678da9bf531SAlexander Bulekov 679da9bf531SAlexander Bulekov /* Run the main loop */ 680da9bf531SAlexander Bulekov flush_events(s); 681da9bf531SAlexander Bulekov } 682da9bf531SAlexander Bulekov /* Advance to the next command */ 683da9bf531SAlexander Bulekov cmd = nextcmd ? nextcmd + sizeof(SEPARATOR) - 1 : nextcmd; 684da9bf531SAlexander Bulekov Size = Size - (cmd_len + sizeof(SEPARATOR) - 1); 68520f5a302SAlexander Bulekov g_array_set_size(dma_regions, 0); 686da9bf531SAlexander Bulekov } 687da9bf531SAlexander Bulekov _Exit(0); 688da9bf531SAlexander Bulekov } else { 689da9bf531SAlexander Bulekov flush_events(s); 690da9bf531SAlexander Bulekov wait(0); 691da9bf531SAlexander Bulekov } 692da9bf531SAlexander Bulekov } 693da9bf531SAlexander Bulekov 694da9bf531SAlexander Bulekov static void usage(void) 695da9bf531SAlexander Bulekov { 696da9bf531SAlexander Bulekov printf("Please specify the following environment variables:\n"); 697da9bf531SAlexander Bulekov printf("QEMU_FUZZ_ARGS= the command line arguments passed to qemu\n"); 698da9bf531SAlexander Bulekov printf("QEMU_FUZZ_OBJECTS= " 699da9bf531SAlexander Bulekov "a space separated list of QOM type names for objects to fuzz\n"); 70020f5a302SAlexander Bulekov printf("Optionally: QEMU_AVOID_DOUBLE_FETCH= " 70120f5a302SAlexander Bulekov "Try to avoid racy DMA double fetch bugs? %d by default\n", 70220f5a302SAlexander Bulekov avoid_double_fetches); 703da9bf531SAlexander Bulekov printf("Optionally: QEMU_FUZZ_TIMEOUT= Specify a custom timeout (us). " 704da9bf531SAlexander Bulekov "0 to disable. %d by default\n", timeout); 705da9bf531SAlexander Bulekov exit(0); 706da9bf531SAlexander Bulekov } 707da9bf531SAlexander Bulekov 708da9bf531SAlexander Bulekov static int locate_fuzz_memory_regions(Object *child, void *opaque) 709da9bf531SAlexander Bulekov { 710da9bf531SAlexander Bulekov const char *name; 711da9bf531SAlexander Bulekov MemoryRegion *mr; 712da9bf531SAlexander Bulekov if (object_dynamic_cast(child, TYPE_MEMORY_REGION)) { 713da9bf531SAlexander Bulekov mr = MEMORY_REGION(child); 714da9bf531SAlexander Bulekov if ((memory_region_is_ram(mr) || 715da9bf531SAlexander Bulekov memory_region_is_ram_device(mr) || 716da9bf531SAlexander Bulekov memory_region_is_rom(mr)) == false) { 717da9bf531SAlexander Bulekov name = object_get_canonical_path_component(child); 718da9bf531SAlexander Bulekov /* 719da9bf531SAlexander Bulekov * We don't want duplicate pointers to the same MemoryRegion, so 720da9bf531SAlexander Bulekov * try to remove copies of the pointer, before adding it. 721da9bf531SAlexander Bulekov */ 722da9bf531SAlexander Bulekov g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true); 723da9bf531SAlexander Bulekov } 724da9bf531SAlexander Bulekov } 725da9bf531SAlexander Bulekov return 0; 726da9bf531SAlexander Bulekov } 727da9bf531SAlexander Bulekov 728da9bf531SAlexander Bulekov static int locate_fuzz_objects(Object *child, void *opaque) 729da9bf531SAlexander Bulekov { 730da9bf531SAlexander Bulekov char *pattern = opaque; 731da9bf531SAlexander Bulekov if (g_pattern_match_simple(pattern, object_get_typename(child))) { 732da9bf531SAlexander Bulekov /* Find and save ptrs to any child MemoryRegions */ 733da9bf531SAlexander Bulekov object_child_foreach_recursive(child, locate_fuzz_memory_regions, NULL); 734da9bf531SAlexander Bulekov 73505efbf24SAlexander Bulekov /* 73605efbf24SAlexander Bulekov * We matched an object. If its a PCI device, store a pointer to it so 73705efbf24SAlexander Bulekov * we can map BARs and fuzz its config space. 73805efbf24SAlexander Bulekov */ 73905efbf24SAlexander Bulekov if (object_dynamic_cast(OBJECT(child), TYPE_PCI_DEVICE)) { 74005efbf24SAlexander Bulekov /* 74105efbf24SAlexander Bulekov * Don't want duplicate pointers to the same PCIDevice, so remove 74205efbf24SAlexander Bulekov * copies of the pointer, before adding it. 74305efbf24SAlexander Bulekov */ 74405efbf24SAlexander Bulekov g_ptr_array_remove_fast(fuzzable_pci_devices, PCI_DEVICE(child)); 74505efbf24SAlexander Bulekov g_ptr_array_add(fuzzable_pci_devices, PCI_DEVICE(child)); 74605efbf24SAlexander Bulekov } 747da9bf531SAlexander Bulekov } else if (object_dynamic_cast(OBJECT(child), TYPE_MEMORY_REGION)) { 748da9bf531SAlexander Bulekov if (g_pattern_match_simple(pattern, 749da9bf531SAlexander Bulekov object_get_canonical_path_component(child))) { 750da9bf531SAlexander Bulekov MemoryRegion *mr; 751da9bf531SAlexander Bulekov mr = MEMORY_REGION(child); 752da9bf531SAlexander Bulekov if ((memory_region_is_ram(mr) || 753da9bf531SAlexander Bulekov memory_region_is_ram_device(mr) || 754da9bf531SAlexander Bulekov memory_region_is_rom(mr)) == false) { 755da9bf531SAlexander Bulekov g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true); 756da9bf531SAlexander Bulekov } 757da9bf531SAlexander Bulekov } 758da9bf531SAlexander Bulekov } 759da9bf531SAlexander Bulekov return 0; 760da9bf531SAlexander Bulekov } 761da9bf531SAlexander Bulekov 762da9bf531SAlexander Bulekov static void generic_pre_fuzz(QTestState *s) 763da9bf531SAlexander Bulekov { 764da9bf531SAlexander Bulekov GHashTableIter iter; 765da9bf531SAlexander Bulekov MemoryRegion *mr; 766da9bf531SAlexander Bulekov char **result; 767da9bf531SAlexander Bulekov 768da9bf531SAlexander Bulekov if (!getenv("QEMU_FUZZ_OBJECTS")) { 769da9bf531SAlexander Bulekov usage(); 770da9bf531SAlexander Bulekov } 771da9bf531SAlexander Bulekov if (getenv("QTEST_LOG")) { 772da9bf531SAlexander Bulekov qtest_log_enabled = 1; 773da9bf531SAlexander Bulekov } 77420f5a302SAlexander Bulekov if (getenv("QEMU_AVOID_DOUBLE_FETCH")) { 77520f5a302SAlexander Bulekov avoid_double_fetches = 1; 77620f5a302SAlexander Bulekov } 777da9bf531SAlexander Bulekov if (getenv("QEMU_FUZZ_TIMEOUT")) { 778da9bf531SAlexander Bulekov timeout = g_ascii_strtoll(getenv("QEMU_FUZZ_TIMEOUT"), NULL, 0); 779da9bf531SAlexander Bulekov } 78020f5a302SAlexander Bulekov qts_global = s; 78120f5a302SAlexander Bulekov 78220f5a302SAlexander Bulekov dma_regions = g_array_new(false, false, sizeof(address_range)); 78320f5a302SAlexander Bulekov dma_patterns = g_array_new(false, false, sizeof(pattern)); 784da9bf531SAlexander Bulekov 785da9bf531SAlexander Bulekov fuzzable_memoryregions = g_hash_table_new(NULL, NULL); 78605efbf24SAlexander Bulekov fuzzable_pci_devices = g_ptr_array_new(); 787da9bf531SAlexander Bulekov 788da9bf531SAlexander Bulekov result = g_strsplit(getenv("QEMU_FUZZ_OBJECTS"), " ", -1); 789da9bf531SAlexander Bulekov for (int i = 0; result[i] != NULL; i++) { 790da9bf531SAlexander Bulekov printf("Matching objects by name %s\n", result[i]); 791da9bf531SAlexander Bulekov object_child_foreach_recursive(qdev_get_machine(), 792da9bf531SAlexander Bulekov locate_fuzz_objects, 793da9bf531SAlexander Bulekov result[i]); 794da9bf531SAlexander Bulekov } 795da9bf531SAlexander Bulekov g_strfreev(result); 796da9bf531SAlexander Bulekov printf("This process will try to fuzz the following MemoryRegions:\n"); 797da9bf531SAlexander Bulekov 798da9bf531SAlexander Bulekov g_hash_table_iter_init(&iter, fuzzable_memoryregions); 799da9bf531SAlexander Bulekov while (g_hash_table_iter_next(&iter, (gpointer)&mr, NULL)) { 800da9bf531SAlexander Bulekov printf(" * %s (size %lx)\n", 801da9bf531SAlexander Bulekov object_get_canonical_path_component(&(mr->parent_obj)), 802da9bf531SAlexander Bulekov (uint64_t)mr->size); 803da9bf531SAlexander Bulekov } 804da9bf531SAlexander Bulekov 805da9bf531SAlexander Bulekov if (!g_hash_table_size(fuzzable_memoryregions)) { 806da9bf531SAlexander Bulekov printf("No fuzzable memory regions found...\n"); 807da9bf531SAlexander Bulekov exit(1); 808da9bf531SAlexander Bulekov } 809da9bf531SAlexander Bulekov 810da9bf531SAlexander Bulekov counter_shm_init(); 811da9bf531SAlexander Bulekov } 812da9bf531SAlexander Bulekov 813da9bf531SAlexander Bulekov static GString *generic_fuzz_cmdline(FuzzTarget *t) 814da9bf531SAlexander Bulekov { 815da9bf531SAlexander Bulekov GString *cmd_line = g_string_new(TARGET_NAME); 816da9bf531SAlexander Bulekov if (!getenv("QEMU_FUZZ_ARGS")) { 817da9bf531SAlexander Bulekov usage(); 818da9bf531SAlexander Bulekov } 819da9bf531SAlexander Bulekov g_string_append_printf(cmd_line, " -display none \ 820da9bf531SAlexander Bulekov -machine accel=qtest, \ 821da9bf531SAlexander Bulekov -m 512M %s ", getenv("QEMU_FUZZ_ARGS")); 822da9bf531SAlexander Bulekov return cmd_line; 823da9bf531SAlexander Bulekov } 824da9bf531SAlexander Bulekov 825da9bf531SAlexander Bulekov static void register_generic_fuzz_targets(void) 826da9bf531SAlexander Bulekov { 827da9bf531SAlexander Bulekov fuzz_add_target(&(FuzzTarget){ 828da9bf531SAlexander Bulekov .name = "generic-fuzz", 829da9bf531SAlexander Bulekov .description = "Fuzz based on any qemu command-line args. ", 830da9bf531SAlexander Bulekov .get_init_cmdline = generic_fuzz_cmdline, 831da9bf531SAlexander Bulekov .pre_fuzz = generic_pre_fuzz, 832da9bf531SAlexander Bulekov .fuzz = generic_fuzz, 833da9bf531SAlexander Bulekov }); 834da9bf531SAlexander Bulekov } 835da9bf531SAlexander Bulekov 836da9bf531SAlexander Bulekov fuzz_target_init(register_generic_fuzz_targets); 837