xref: /qemu/tests/qtest/fuzz/generic_fuzz.c (revision edf5ca5dbe8031e7814ea34eb109b8f7d4024ae5)
1da9bf531SAlexander Bulekov /*
2da9bf531SAlexander Bulekov  * Generic Virtual-Device Fuzzing Target
3da9bf531SAlexander Bulekov  *
4da9bf531SAlexander Bulekov  * Copyright Red Hat Inc., 2020
5da9bf531SAlexander Bulekov  *
6da9bf531SAlexander Bulekov  * Authors:
7da9bf531SAlexander Bulekov  *  Alexander Bulekov   <alxndr@bu.edu>
8da9bf531SAlexander Bulekov  *
9da9bf531SAlexander Bulekov  * This work is licensed under the terms of the GNU GPL, version 2 or later.
10da9bf531SAlexander Bulekov  * See the COPYING file in the top-level directory.
11da9bf531SAlexander Bulekov  */
12da9bf531SAlexander Bulekov 
13da9bf531SAlexander Bulekov #include "qemu/osdep.h"
14da9bf531SAlexander Bulekov 
15da9bf531SAlexander Bulekov #include <wordexp.h>
16da9bf531SAlexander Bulekov 
17da9bf531SAlexander Bulekov #include "hw/core/cpu.h"
18907b5105SMarc-André Lureau #include "tests/qtest/libqtest.h"
19b677001dSAlexander Bulekov #include "tests/qtest/libqos/pci-pc.h"
20da9bf531SAlexander Bulekov #include "fuzz.h"
21da9bf531SAlexander Bulekov #include "fork_fuzz.h"
22da9bf531SAlexander Bulekov #include "string.h"
23da9bf531SAlexander Bulekov #include "exec/memory.h"
24da9bf531SAlexander Bulekov #include "exec/ramblock.h"
25da9bf531SAlexander Bulekov #include "hw/qdev-core.h"
2605efbf24SAlexander Bulekov #include "hw/pci/pci.h"
27*edf5ca5dSMarkus Armbruster #include "hw/pci/pci_device.h"
2820f5a302SAlexander Bulekov #include "hw/boards.h"
297fdb5053SAlexander Bulekov #include "generic_fuzz_configs.h"
3025d309fbSAlexander Bulekov #include "hw/mem/sparse-mem.h"
31da9bf531SAlexander Bulekov 
32da9bf531SAlexander Bulekov /*
33da9bf531SAlexander Bulekov  * SEPARATOR is used to separate "operations" in the fuzz input
34da9bf531SAlexander Bulekov  */
35da9bf531SAlexander Bulekov #define SEPARATOR "FUZZ"
36da9bf531SAlexander Bulekov 
37da9bf531SAlexander Bulekov enum cmds {
38da9bf531SAlexander Bulekov     OP_IN,
39da9bf531SAlexander Bulekov     OP_OUT,
40da9bf531SAlexander Bulekov     OP_READ,
41da9bf531SAlexander Bulekov     OP_WRITE,
4205efbf24SAlexander Bulekov     OP_PCI_READ,
4305efbf24SAlexander Bulekov     OP_PCI_WRITE,
44ccbd4bc8SAlexander Bulekov     OP_DISABLE_PCI,
4520f5a302SAlexander Bulekov     OP_ADD_DMA_PATTERN,
4620f5a302SAlexander Bulekov     OP_CLEAR_DMA_PATTERNS,
47da9bf531SAlexander Bulekov     OP_CLOCK_STEP,
48da9bf531SAlexander Bulekov };
49da9bf531SAlexander Bulekov 
50da9bf531SAlexander Bulekov #define DEFAULT_TIMEOUT_US 100000
51da9bf531SAlexander Bulekov #define USEC_IN_SEC 1000000000
52da9bf531SAlexander Bulekov 
5320f5a302SAlexander Bulekov #define MAX_DMA_FILL_SIZE 0x10000
5420f5a302SAlexander Bulekov 
5505efbf24SAlexander Bulekov #define PCI_HOST_BRIDGE_CFG 0xcf8
5605efbf24SAlexander Bulekov #define PCI_HOST_BRIDGE_DATA 0xcfc
5705efbf24SAlexander Bulekov 
58da9bf531SAlexander Bulekov typedef struct {
59da9bf531SAlexander Bulekov     ram_addr_t addr;
60da9bf531SAlexander Bulekov     ram_addr_t size; /* The number of bytes until the end of the I/O region */
61da9bf531SAlexander Bulekov } address_range;
62da9bf531SAlexander Bulekov 
63da9bf531SAlexander Bulekov static useconds_t timeout = DEFAULT_TIMEOUT_US;
64da9bf531SAlexander Bulekov 
65da9bf531SAlexander Bulekov static bool qtest_log_enabled;
66da9bf531SAlexander Bulekov 
6725d309fbSAlexander Bulekov MemoryRegion *sparse_mem_mr;
6825d309fbSAlexander Bulekov 
69da9bf531SAlexander Bulekov /*
7020f5a302SAlexander Bulekov  * A pattern used to populate a DMA region or perform a memwrite. This is
7120f5a302SAlexander Bulekov  * useful for e.g. populating tables of unique addresses.
7220f5a302SAlexander Bulekov  * Example {.index = 1; .stride = 2; .len = 3; .data = "\x00\x01\x02"}
7320f5a302SAlexander Bulekov  * Renders as: 00 01 02   00 03 02   00 05 02   00 07 02 ...
7420f5a302SAlexander Bulekov  */
7520f5a302SAlexander Bulekov typedef struct {
7620f5a302SAlexander Bulekov     uint8_t index;      /* Index of a byte to increment by stride */
7720f5a302SAlexander Bulekov     uint8_t stride;     /* Increment each index'th byte by this amount */
7820f5a302SAlexander Bulekov     size_t len;
7920f5a302SAlexander Bulekov     const uint8_t *data;
8020f5a302SAlexander Bulekov } pattern;
8120f5a302SAlexander Bulekov 
8220f5a302SAlexander Bulekov /* Avoid filling the same DMA region between MMIO/PIO commands ? */
8320f5a302SAlexander Bulekov static bool avoid_double_fetches;
8420f5a302SAlexander Bulekov 
8520f5a302SAlexander Bulekov static QTestState *qts_global; /* Need a global for the DMA callback */
8620f5a302SAlexander Bulekov 
8720f5a302SAlexander Bulekov /*
88da9bf531SAlexander Bulekov  * List of memory regions that are children of QOM objects specified by the
89da9bf531SAlexander Bulekov  * user for fuzzing.
90da9bf531SAlexander Bulekov  */
91da9bf531SAlexander Bulekov static GHashTable *fuzzable_memoryregions;
9205efbf24SAlexander Bulekov static GPtrArray *fuzzable_pci_devices;
93da9bf531SAlexander Bulekov 
94da9bf531SAlexander Bulekov struct get_io_cb_info {
95da9bf531SAlexander Bulekov     int index;
96da9bf531SAlexander Bulekov     int found;
97da9bf531SAlexander Bulekov     address_range result;
98da9bf531SAlexander Bulekov };
99da9bf531SAlexander Bulekov 
100d1e8cf77SPeter Maydell static bool get_io_address_cb(Int128 start, Int128 size,
101b3566001SPeter Maydell                               const MemoryRegion *mr,
102b3566001SPeter Maydell                               hwaddr offset_in_region,
103b3566001SPeter Maydell                               void *opaque)
104b3566001SPeter Maydell {
105da9bf531SAlexander Bulekov     struct get_io_cb_info *info = opaque;
106da9bf531SAlexander Bulekov     if (g_hash_table_lookup(fuzzable_memoryregions, mr)) {
107da9bf531SAlexander Bulekov         if (info->index == 0) {
108da9bf531SAlexander Bulekov             info->result.addr = (ram_addr_t)start;
109da9bf531SAlexander Bulekov             info->result.size = (ram_addr_t)size;
110da9bf531SAlexander Bulekov             info->found = 1;
111d1e8cf77SPeter Maydell             return true;
112da9bf531SAlexander Bulekov         }
113da9bf531SAlexander Bulekov         info->index--;
114da9bf531SAlexander Bulekov     }
115d1e8cf77SPeter Maydell     return false;
116da9bf531SAlexander Bulekov }
117da9bf531SAlexander Bulekov 
118da9bf531SAlexander Bulekov /*
11920f5a302SAlexander Bulekov  * List of dma regions populated since the last fuzzing command. Used to ensure
12020f5a302SAlexander Bulekov  * that we only write to each DMA address once, to avoid race conditions when
12120f5a302SAlexander Bulekov  * building reproducers.
12220f5a302SAlexander Bulekov  */
12320f5a302SAlexander Bulekov static GArray *dma_regions;
12420f5a302SAlexander Bulekov 
12520f5a302SAlexander Bulekov static GArray *dma_patterns;
12620f5a302SAlexander Bulekov static int dma_pattern_index;
127ccbd4bc8SAlexander Bulekov static bool pci_disabled;
12820f5a302SAlexander Bulekov 
12920f5a302SAlexander Bulekov /*
13020f5a302SAlexander Bulekov  * Allocate a block of memory and populate it with a pattern.
13120f5a302SAlexander Bulekov  */
13220f5a302SAlexander Bulekov static void *pattern_alloc(pattern p, size_t len)
13320f5a302SAlexander Bulekov {
13420f5a302SAlexander Bulekov     int i;
13520f5a302SAlexander Bulekov     uint8_t *buf = g_malloc(len);
13620f5a302SAlexander Bulekov     uint8_t sum = 0;
13720f5a302SAlexander Bulekov 
13820f5a302SAlexander Bulekov     for (i = 0; i < len; ++i) {
13920f5a302SAlexander Bulekov         buf[i] = p.data[i % p.len];
14020f5a302SAlexander Bulekov         if ((i % p.len) == p.index) {
14120f5a302SAlexander Bulekov             buf[i] += sum;
14220f5a302SAlexander Bulekov             sum += p.stride;
14320f5a302SAlexander Bulekov         }
14420f5a302SAlexander Bulekov     }
14520f5a302SAlexander Bulekov     return buf;
14620f5a302SAlexander Bulekov }
14720f5a302SAlexander Bulekov 
1483123f93dSJagannathan Raman static int fuzz_memory_access_size(MemoryRegion *mr, unsigned l, hwaddr addr)
14920f5a302SAlexander Bulekov {
15020f5a302SAlexander Bulekov     unsigned access_size_max = mr->ops->valid.max_access_size;
15120f5a302SAlexander Bulekov 
15220f5a302SAlexander Bulekov     /*
15320f5a302SAlexander Bulekov      * Regions are assumed to support 1-4 byte accesses unless
15420f5a302SAlexander Bulekov      * otherwise specified.
15520f5a302SAlexander Bulekov      */
15620f5a302SAlexander Bulekov     if (access_size_max == 0) {
15720f5a302SAlexander Bulekov         access_size_max = 4;
15820f5a302SAlexander Bulekov     }
15920f5a302SAlexander Bulekov 
16020f5a302SAlexander Bulekov     /* Bound the maximum access by the alignment of the address.  */
16120f5a302SAlexander Bulekov     if (!mr->ops->impl.unaligned) {
16220f5a302SAlexander Bulekov         unsigned align_size_max = addr & -addr;
16320f5a302SAlexander Bulekov         if (align_size_max != 0 && align_size_max < access_size_max) {
16420f5a302SAlexander Bulekov             access_size_max = align_size_max;
16520f5a302SAlexander Bulekov         }
16620f5a302SAlexander Bulekov     }
16720f5a302SAlexander Bulekov 
16820f5a302SAlexander Bulekov     /* Don't attempt accesses larger than the maximum.  */
16920f5a302SAlexander Bulekov     if (l > access_size_max) {
17020f5a302SAlexander Bulekov         l = access_size_max;
17120f5a302SAlexander Bulekov     }
17220f5a302SAlexander Bulekov     l = pow2floor(l);
17320f5a302SAlexander Bulekov 
17420f5a302SAlexander Bulekov     return l;
17520f5a302SAlexander Bulekov }
17620f5a302SAlexander Bulekov 
17720f5a302SAlexander Bulekov /*
17820f5a302SAlexander Bulekov  * Call-back for functions that perform DMA reads from guest memory. Confirm
17920f5a302SAlexander Bulekov  * that the region has not already been populated since the last loop in
18020f5a302SAlexander Bulekov  * generic_fuzz(), avoiding potential race-conditions, which we don't have
18120f5a302SAlexander Bulekov  * a good way for reproducing right now.
18220f5a302SAlexander Bulekov  */
183fc1c8344SAlexander Bulekov void fuzz_dma_read_cb(size_t addr, size_t len, MemoryRegion *mr)
18420f5a302SAlexander Bulekov {
18520f5a302SAlexander Bulekov     /* Are we in the generic-fuzzer or are we using another fuzz-target? */
18620f5a302SAlexander Bulekov     if (!qts_global) {
18720f5a302SAlexander Bulekov         return;
18820f5a302SAlexander Bulekov     }
18920f5a302SAlexander Bulekov 
19020f5a302SAlexander Bulekov     /*
19120f5a302SAlexander Bulekov      * Return immediately if:
19220f5a302SAlexander Bulekov      * - We have no DMA patterns defined
19320f5a302SAlexander Bulekov      * - The length of the DMA read request is zero
19420f5a302SAlexander Bulekov      * - The DMA read is hitting an MR other than the machine's main RAM
19520f5a302SAlexander Bulekov      * - The DMA request hits past the bounds of our RAM
19620f5a302SAlexander Bulekov      */
19720f5a302SAlexander Bulekov     if (dma_patterns->len == 0
19820f5a302SAlexander Bulekov         || len == 0
19925d309fbSAlexander Bulekov         || (mr != current_machine->ram && mr != sparse_mem_mr)) {
20020f5a302SAlexander Bulekov         return;
20120f5a302SAlexander Bulekov     }
20220f5a302SAlexander Bulekov 
20320f5a302SAlexander Bulekov     /*
20420f5a302SAlexander Bulekov      * If we overlap with any existing dma_regions, split the range and only
20520f5a302SAlexander Bulekov      * populate the non-overlapping parts.
20620f5a302SAlexander Bulekov      */
20720f5a302SAlexander Bulekov     address_range region;
20820f5a302SAlexander Bulekov     bool double_fetch = false;
20920f5a302SAlexander Bulekov     for (int i = 0;
21020f5a302SAlexander Bulekov          i < dma_regions->len && (avoid_double_fetches || qtest_log_enabled);
21120f5a302SAlexander Bulekov          ++i) {
21220f5a302SAlexander Bulekov         region = g_array_index(dma_regions, address_range, i);
21320f5a302SAlexander Bulekov         if (addr < region.addr + region.size && addr + len > region.addr) {
21420f5a302SAlexander Bulekov             double_fetch = true;
21520f5a302SAlexander Bulekov             if (addr < region.addr
21620f5a302SAlexander Bulekov                 && avoid_double_fetches) {
217fc1c8344SAlexander Bulekov                 fuzz_dma_read_cb(addr, region.addr - addr, mr);
21820f5a302SAlexander Bulekov             }
21920f5a302SAlexander Bulekov             if (addr + len > region.addr + region.size
22020f5a302SAlexander Bulekov                 && avoid_double_fetches) {
22120f5a302SAlexander Bulekov                 fuzz_dma_read_cb(region.addr + region.size,
222fc1c8344SAlexander Bulekov                         addr + len - (region.addr + region.size), mr);
22320f5a302SAlexander Bulekov             }
22420f5a302SAlexander Bulekov             return;
22520f5a302SAlexander Bulekov         }
22620f5a302SAlexander Bulekov     }
22720f5a302SAlexander Bulekov 
22820f5a302SAlexander Bulekov     /* Cap the length of the DMA access to something reasonable */
22920f5a302SAlexander Bulekov     len = MIN(len, MAX_DMA_FILL_SIZE);
23020f5a302SAlexander Bulekov 
23120f5a302SAlexander Bulekov     address_range ar = {addr, len};
23220f5a302SAlexander Bulekov     g_array_append_val(dma_regions, ar);
23320f5a302SAlexander Bulekov     pattern p = g_array_index(dma_patterns, pattern, dma_pattern_index);
234a9f67c1dSAlexander Bulekov     void *buf_base = pattern_alloc(p, ar.size);
235a9f67c1dSAlexander Bulekov     void *buf = buf_base;
23620f5a302SAlexander Bulekov     hwaddr l, addr1;
23720f5a302SAlexander Bulekov     MemoryRegion *mr1;
23820f5a302SAlexander Bulekov     while (len > 0) {
23920f5a302SAlexander Bulekov         l = len;
24020f5a302SAlexander Bulekov         mr1 = address_space_translate(first_cpu->as,
24120f5a302SAlexander Bulekov                                       addr, &addr1, &l, true,
24220f5a302SAlexander Bulekov                                       MEMTXATTRS_UNSPECIFIED);
24320f5a302SAlexander Bulekov 
244af16990aSAlexander Bulekov         /*
245af16990aSAlexander Bulekov          *  If mr1 isn't RAM, address_space_translate doesn't update l. Use
2463123f93dSJagannathan Raman          *  fuzz_memory_access_size to identify the number of bytes that it
2473123f93dSJagannathan Raman          *  is safe to write without accidentally writing to another
2483123f93dSJagannathan Raman          *  MemoryRegion.
249af16990aSAlexander Bulekov          */
250af16990aSAlexander Bulekov         if (!memory_region_is_ram(mr1)) {
2513123f93dSJagannathan Raman             l = fuzz_memory_access_size(mr1, l, addr1);
252af16990aSAlexander Bulekov         }
253af16990aSAlexander Bulekov         if (memory_region_is_ram(mr1) ||
254af16990aSAlexander Bulekov             memory_region_is_romd(mr1) ||
255af16990aSAlexander Bulekov             mr1 == sparse_mem_mr) {
25620f5a302SAlexander Bulekov             /* ROM/RAM case */
25720f5a302SAlexander Bulekov             if (qtest_log_enabled) {
25820f5a302SAlexander Bulekov                 /*
25920f5a302SAlexander Bulekov                 * With QTEST_LOG, use a normal, slow QTest memwrite. Prefix the log
26020f5a302SAlexander Bulekov                 * that will be written by qtest.c with a DMA tag, so we can reorder
26120f5a302SAlexander Bulekov                 * the resulting QTest trace so the DMA fills precede the last PIO/MMIO
26220f5a302SAlexander Bulekov                 * command.
26320f5a302SAlexander Bulekov                 */
26420f5a302SAlexander Bulekov                 fprintf(stderr, "[DMA] ");
26520f5a302SAlexander Bulekov                 if (double_fetch) {
26620f5a302SAlexander Bulekov                     fprintf(stderr, "[DOUBLE-FETCH] ");
26720f5a302SAlexander Bulekov                 }
26820f5a302SAlexander Bulekov                 fflush(stderr);
26920f5a302SAlexander Bulekov             }
270a9f67c1dSAlexander Bulekov             qtest_memwrite(qts_global, addr, buf, l);
271a9f67c1dSAlexander Bulekov         }
272a9f67c1dSAlexander Bulekov         len -= l;
273a9f67c1dSAlexander Bulekov         buf += l;
274a9f67c1dSAlexander Bulekov         addr += l;
275a9f67c1dSAlexander Bulekov 
276a9f67c1dSAlexander Bulekov     }
277a9f67c1dSAlexander Bulekov     g_free(buf_base);
27820f5a302SAlexander Bulekov 
27920f5a302SAlexander Bulekov     /* Increment the index of the pattern for the next DMA access */
28020f5a302SAlexander Bulekov     dma_pattern_index = (dma_pattern_index + 1) % dma_patterns->len;
28120f5a302SAlexander Bulekov }
28220f5a302SAlexander Bulekov 
28320f5a302SAlexander Bulekov /*
284da9bf531SAlexander Bulekov  * Here we want to convert a fuzzer-provided [io-region-index, offset] to
285da9bf531SAlexander Bulekov  * a physical address. To do this, we iterate over all of the matched
286da9bf531SAlexander Bulekov  * MemoryRegions. Check whether each region exists within the particular io
287da9bf531SAlexander Bulekov  * space. Return the absolute address of the offset within the index'th region
288da9bf531SAlexander Bulekov  * that is a subregion of the io_space and the distance until the end of the
289da9bf531SAlexander Bulekov  * memory region.
290da9bf531SAlexander Bulekov  */
291da9bf531SAlexander Bulekov static bool get_io_address(address_range *result, AddressSpace *as,
292da9bf531SAlexander Bulekov                             uint8_t index,
293da9bf531SAlexander Bulekov                             uint32_t offset) {
294da9bf531SAlexander Bulekov     FlatView *view;
295da9bf531SAlexander Bulekov     view = as->current_map;
296da9bf531SAlexander Bulekov     g_assert(view);
297da9bf531SAlexander Bulekov     struct get_io_cb_info cb_info = {};
298da9bf531SAlexander Bulekov 
299da9bf531SAlexander Bulekov     cb_info.index = index;
300da9bf531SAlexander Bulekov 
301da9bf531SAlexander Bulekov     /*
302da9bf531SAlexander Bulekov      * Loop around the FlatView until we match "index" number of
303da9bf531SAlexander Bulekov      * fuzzable_memoryregions, or until we know that there are no matching
304da9bf531SAlexander Bulekov      * memory_regions.
305da9bf531SAlexander Bulekov      */
306da9bf531SAlexander Bulekov     do {
307da9bf531SAlexander Bulekov         flatview_for_each_range(view, get_io_address_cb , &cb_info);
308da9bf531SAlexander Bulekov     } while (cb_info.index != index && !cb_info.found);
309da9bf531SAlexander Bulekov 
310da9bf531SAlexander Bulekov     *result = cb_info.result;
311953e6d7cSAlexander Bulekov     if (result->size) {
312953e6d7cSAlexander Bulekov         offset = offset % result->size;
313953e6d7cSAlexander Bulekov         result->addr += offset;
314953e6d7cSAlexander Bulekov         result->size -= offset;
315953e6d7cSAlexander Bulekov     }
316da9bf531SAlexander Bulekov     return cb_info.found;
317da9bf531SAlexander Bulekov }
318da9bf531SAlexander Bulekov 
319da9bf531SAlexander Bulekov static bool get_pio_address(address_range *result,
320da9bf531SAlexander Bulekov                             uint8_t index, uint16_t offset)
321da9bf531SAlexander Bulekov {
322da9bf531SAlexander Bulekov     /*
323da9bf531SAlexander Bulekov      * PIO BARs can be set past the maximum port address (0xFFFF). Thus, result
324da9bf531SAlexander Bulekov      * can contain an addr that extends past the PIO space. When we pass this
325da9bf531SAlexander Bulekov      * address to qtest_in/qtest_out, it is cast to a uint16_t, so we might end
326da9bf531SAlexander Bulekov      * up fuzzing a completely different MemoryRegion/Device. Therefore, check
327da9bf531SAlexander Bulekov      * that the address here is within the PIO space limits.
328da9bf531SAlexander Bulekov      */
329da9bf531SAlexander Bulekov     bool found = get_io_address(result, &address_space_io, index, offset);
330da9bf531SAlexander Bulekov     return result->addr <= 0xFFFF ? found : false;
331da9bf531SAlexander Bulekov }
332da9bf531SAlexander Bulekov 
333da9bf531SAlexander Bulekov static bool get_mmio_address(address_range *result,
334da9bf531SAlexander Bulekov                              uint8_t index, uint32_t offset)
335da9bf531SAlexander Bulekov {
336da9bf531SAlexander Bulekov     return get_io_address(result, &address_space_memory, index, offset);
337da9bf531SAlexander Bulekov }
338da9bf531SAlexander Bulekov 
339da9bf531SAlexander Bulekov static void op_in(QTestState *s, const unsigned char * data, size_t len)
340da9bf531SAlexander Bulekov {
341da9bf531SAlexander Bulekov     enum Sizes {Byte, Word, Long, end_sizes};
342da9bf531SAlexander Bulekov     struct {
343da9bf531SAlexander Bulekov         uint8_t size;
344da9bf531SAlexander Bulekov         uint8_t base;
345da9bf531SAlexander Bulekov         uint16_t offset;
346da9bf531SAlexander Bulekov     } a;
347da9bf531SAlexander Bulekov     address_range abs;
348da9bf531SAlexander Bulekov 
349da9bf531SAlexander Bulekov     if (len < sizeof(a)) {
350da9bf531SAlexander Bulekov         return;
351da9bf531SAlexander Bulekov     }
352da9bf531SAlexander Bulekov     memcpy(&a, data, sizeof(a));
353da9bf531SAlexander Bulekov     if (get_pio_address(&abs, a.base, a.offset) == 0) {
354da9bf531SAlexander Bulekov         return;
355da9bf531SAlexander Bulekov     }
356da9bf531SAlexander Bulekov 
357da9bf531SAlexander Bulekov     switch (a.size %= end_sizes) {
358da9bf531SAlexander Bulekov     case Byte:
359da9bf531SAlexander Bulekov         qtest_inb(s, abs.addr);
360da9bf531SAlexander Bulekov         break;
361da9bf531SAlexander Bulekov     case Word:
362da9bf531SAlexander Bulekov         if (abs.size >= 2) {
363da9bf531SAlexander Bulekov             qtest_inw(s, abs.addr);
364da9bf531SAlexander Bulekov         }
365da9bf531SAlexander Bulekov         break;
366da9bf531SAlexander Bulekov     case Long:
367da9bf531SAlexander Bulekov         if (abs.size >= 4) {
368da9bf531SAlexander Bulekov             qtest_inl(s, abs.addr);
369da9bf531SAlexander Bulekov         }
370da9bf531SAlexander Bulekov         break;
371da9bf531SAlexander Bulekov     }
372da9bf531SAlexander Bulekov }
373da9bf531SAlexander Bulekov 
374da9bf531SAlexander Bulekov static void op_out(QTestState *s, const unsigned char * data, size_t len)
375da9bf531SAlexander Bulekov {
376da9bf531SAlexander Bulekov     enum Sizes {Byte, Word, Long, end_sizes};
377da9bf531SAlexander Bulekov     struct {
378da9bf531SAlexander Bulekov         uint8_t size;
379da9bf531SAlexander Bulekov         uint8_t base;
380da9bf531SAlexander Bulekov         uint16_t offset;
381da9bf531SAlexander Bulekov         uint32_t value;
382da9bf531SAlexander Bulekov     } a;
383da9bf531SAlexander Bulekov     address_range abs;
384da9bf531SAlexander Bulekov 
385da9bf531SAlexander Bulekov     if (len < sizeof(a)) {
386da9bf531SAlexander Bulekov         return;
387da9bf531SAlexander Bulekov     }
388da9bf531SAlexander Bulekov     memcpy(&a, data, sizeof(a));
389da9bf531SAlexander Bulekov 
390da9bf531SAlexander Bulekov     if (get_pio_address(&abs, a.base, a.offset) == 0) {
391da9bf531SAlexander Bulekov         return;
392da9bf531SAlexander Bulekov     }
393da9bf531SAlexander Bulekov 
394da9bf531SAlexander Bulekov     switch (a.size %= end_sizes) {
395da9bf531SAlexander Bulekov     case Byte:
396da9bf531SAlexander Bulekov         qtest_outb(s, abs.addr, a.value & 0xFF);
397da9bf531SAlexander Bulekov         break;
398da9bf531SAlexander Bulekov     case Word:
399da9bf531SAlexander Bulekov         if (abs.size >= 2) {
400da9bf531SAlexander Bulekov             qtest_outw(s, abs.addr, a.value & 0xFFFF);
401da9bf531SAlexander Bulekov         }
402da9bf531SAlexander Bulekov         break;
403da9bf531SAlexander Bulekov     case Long:
404da9bf531SAlexander Bulekov         if (abs.size >= 4) {
405da9bf531SAlexander Bulekov             qtest_outl(s, abs.addr, a.value);
406da9bf531SAlexander Bulekov         }
407da9bf531SAlexander Bulekov         break;
408da9bf531SAlexander Bulekov     }
409da9bf531SAlexander Bulekov }
410da9bf531SAlexander Bulekov 
411da9bf531SAlexander Bulekov static void op_read(QTestState *s, const unsigned char * data, size_t len)
412da9bf531SAlexander Bulekov {
413da9bf531SAlexander Bulekov     enum Sizes {Byte, Word, Long, Quad, end_sizes};
414da9bf531SAlexander Bulekov     struct {
415da9bf531SAlexander Bulekov         uint8_t size;
416da9bf531SAlexander Bulekov         uint8_t base;
417da9bf531SAlexander Bulekov         uint32_t offset;
418da9bf531SAlexander Bulekov     } a;
419da9bf531SAlexander Bulekov     address_range abs;
420da9bf531SAlexander Bulekov 
421da9bf531SAlexander Bulekov     if (len < sizeof(a)) {
422da9bf531SAlexander Bulekov         return;
423da9bf531SAlexander Bulekov     }
424da9bf531SAlexander Bulekov     memcpy(&a, data, sizeof(a));
425da9bf531SAlexander Bulekov 
426da9bf531SAlexander Bulekov     if (get_mmio_address(&abs, a.base, a.offset) == 0) {
427da9bf531SAlexander Bulekov         return;
428da9bf531SAlexander Bulekov     }
429da9bf531SAlexander Bulekov 
430da9bf531SAlexander Bulekov     switch (a.size %= end_sizes) {
431da9bf531SAlexander Bulekov     case Byte:
432da9bf531SAlexander Bulekov         qtest_readb(s, abs.addr);
433da9bf531SAlexander Bulekov         break;
434da9bf531SAlexander Bulekov     case Word:
435da9bf531SAlexander Bulekov         if (abs.size >= 2) {
436da9bf531SAlexander Bulekov             qtest_readw(s, abs.addr);
437da9bf531SAlexander Bulekov         }
438da9bf531SAlexander Bulekov         break;
439da9bf531SAlexander Bulekov     case Long:
440da9bf531SAlexander Bulekov         if (abs.size >= 4) {
441da9bf531SAlexander Bulekov             qtest_readl(s, abs.addr);
442da9bf531SAlexander Bulekov         }
443da9bf531SAlexander Bulekov         break;
444da9bf531SAlexander Bulekov     case Quad:
445da9bf531SAlexander Bulekov         if (abs.size >= 8) {
446da9bf531SAlexander Bulekov             qtest_readq(s, abs.addr);
447da9bf531SAlexander Bulekov         }
448da9bf531SAlexander Bulekov         break;
449da9bf531SAlexander Bulekov     }
450da9bf531SAlexander Bulekov }
451da9bf531SAlexander Bulekov 
452da9bf531SAlexander Bulekov static void op_write(QTestState *s, const unsigned char * data, size_t len)
453da9bf531SAlexander Bulekov {
454da9bf531SAlexander Bulekov     enum Sizes {Byte, Word, Long, Quad, end_sizes};
455da9bf531SAlexander Bulekov     struct {
456da9bf531SAlexander Bulekov         uint8_t size;
457da9bf531SAlexander Bulekov         uint8_t base;
458da9bf531SAlexander Bulekov         uint32_t offset;
459da9bf531SAlexander Bulekov         uint64_t value;
460da9bf531SAlexander Bulekov     } a;
461da9bf531SAlexander Bulekov     address_range abs;
462da9bf531SAlexander Bulekov 
463da9bf531SAlexander Bulekov     if (len < sizeof(a)) {
464da9bf531SAlexander Bulekov         return;
465da9bf531SAlexander Bulekov     }
466da9bf531SAlexander Bulekov     memcpy(&a, data, sizeof(a));
467da9bf531SAlexander Bulekov 
468da9bf531SAlexander Bulekov     if (get_mmio_address(&abs, a.base, a.offset) == 0) {
469da9bf531SAlexander Bulekov         return;
470da9bf531SAlexander Bulekov     }
471da9bf531SAlexander Bulekov 
472da9bf531SAlexander Bulekov     switch (a.size %= end_sizes) {
473da9bf531SAlexander Bulekov     case Byte:
474da9bf531SAlexander Bulekov             qtest_writeb(s, abs.addr, a.value & 0xFF);
475da9bf531SAlexander Bulekov         break;
476da9bf531SAlexander Bulekov     case Word:
477da9bf531SAlexander Bulekov         if (abs.size >= 2) {
478da9bf531SAlexander Bulekov             qtest_writew(s, abs.addr, a.value & 0xFFFF);
479da9bf531SAlexander Bulekov         }
480da9bf531SAlexander Bulekov         break;
481da9bf531SAlexander Bulekov     case Long:
482da9bf531SAlexander Bulekov         if (abs.size >= 4) {
483da9bf531SAlexander Bulekov             qtest_writel(s, abs.addr, a.value & 0xFFFFFFFF);
484da9bf531SAlexander Bulekov         }
485da9bf531SAlexander Bulekov         break;
486da9bf531SAlexander Bulekov     case Quad:
487da9bf531SAlexander Bulekov         if (abs.size >= 8) {
488da9bf531SAlexander Bulekov             qtest_writeq(s, abs.addr, a.value);
489da9bf531SAlexander Bulekov         }
490da9bf531SAlexander Bulekov         break;
491da9bf531SAlexander Bulekov     }
492da9bf531SAlexander Bulekov }
493da9bf531SAlexander Bulekov 
49405efbf24SAlexander Bulekov static void op_pci_read(QTestState *s, const unsigned char * data, size_t len)
49505efbf24SAlexander Bulekov {
49605efbf24SAlexander Bulekov     enum Sizes {Byte, Word, Long, end_sizes};
49705efbf24SAlexander Bulekov     struct {
49805efbf24SAlexander Bulekov         uint8_t size;
49905efbf24SAlexander Bulekov         uint8_t base;
50005efbf24SAlexander Bulekov         uint8_t offset;
50105efbf24SAlexander Bulekov     } a;
502ccbd4bc8SAlexander Bulekov     if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
50305efbf24SAlexander Bulekov         return;
50405efbf24SAlexander Bulekov     }
50505efbf24SAlexander Bulekov     memcpy(&a, data, sizeof(a));
50605efbf24SAlexander Bulekov     PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
50705efbf24SAlexander Bulekov                                   a.base % fuzzable_pci_devices->len);
50805efbf24SAlexander Bulekov     int devfn = dev->devfn;
50905efbf24SAlexander Bulekov     qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
51005efbf24SAlexander Bulekov     switch (a.size %= end_sizes) {
51105efbf24SAlexander Bulekov     case Byte:
51205efbf24SAlexander Bulekov         qtest_inb(s, PCI_HOST_BRIDGE_DATA);
51305efbf24SAlexander Bulekov         break;
51405efbf24SAlexander Bulekov     case Word:
51505efbf24SAlexander Bulekov         qtest_inw(s, PCI_HOST_BRIDGE_DATA);
51605efbf24SAlexander Bulekov         break;
51705efbf24SAlexander Bulekov     case Long:
51805efbf24SAlexander Bulekov         qtest_inl(s, PCI_HOST_BRIDGE_DATA);
51905efbf24SAlexander Bulekov         break;
52005efbf24SAlexander Bulekov     }
52105efbf24SAlexander Bulekov }
52205efbf24SAlexander Bulekov 
52305efbf24SAlexander Bulekov static void op_pci_write(QTestState *s, const unsigned char * data, size_t len)
52405efbf24SAlexander Bulekov {
52505efbf24SAlexander Bulekov     enum Sizes {Byte, Word, Long, end_sizes};
52605efbf24SAlexander Bulekov     struct {
52705efbf24SAlexander Bulekov         uint8_t size;
52805efbf24SAlexander Bulekov         uint8_t base;
52905efbf24SAlexander Bulekov         uint8_t offset;
53005efbf24SAlexander Bulekov         uint32_t value;
53105efbf24SAlexander Bulekov     } a;
532ccbd4bc8SAlexander Bulekov     if (len < sizeof(a) || fuzzable_pci_devices->len == 0 || pci_disabled) {
53305efbf24SAlexander Bulekov         return;
53405efbf24SAlexander Bulekov     }
53505efbf24SAlexander Bulekov     memcpy(&a, data, sizeof(a));
53605efbf24SAlexander Bulekov     PCIDevice *dev = g_ptr_array_index(fuzzable_pci_devices,
53705efbf24SAlexander Bulekov                                   a.base % fuzzable_pci_devices->len);
53805efbf24SAlexander Bulekov     int devfn = dev->devfn;
53905efbf24SAlexander Bulekov     qtest_outl(s, PCI_HOST_BRIDGE_CFG, (1U << 31) | (devfn << 8) | a.offset);
54005efbf24SAlexander Bulekov     switch (a.size %= end_sizes) {
54105efbf24SAlexander Bulekov     case Byte:
54205efbf24SAlexander Bulekov         qtest_outb(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFF);
54305efbf24SAlexander Bulekov         break;
54405efbf24SAlexander Bulekov     case Word:
54505efbf24SAlexander Bulekov         qtest_outw(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFF);
54605efbf24SAlexander Bulekov         break;
54705efbf24SAlexander Bulekov     case Long:
54805efbf24SAlexander Bulekov         qtest_outl(s, PCI_HOST_BRIDGE_DATA, a.value & 0xFFFFFFFF);
54905efbf24SAlexander Bulekov         break;
55005efbf24SAlexander Bulekov     }
55105efbf24SAlexander Bulekov }
55205efbf24SAlexander Bulekov 
55320f5a302SAlexander Bulekov static void op_add_dma_pattern(QTestState *s,
55420f5a302SAlexander Bulekov                                const unsigned char *data, size_t len)
55520f5a302SAlexander Bulekov {
55620f5a302SAlexander Bulekov     struct {
55720f5a302SAlexander Bulekov         /*
55820f5a302SAlexander Bulekov          * index and stride can be used to increment the index-th byte of the
55920f5a302SAlexander Bulekov          * pattern by the value stride, for each loop of the pattern.
56020f5a302SAlexander Bulekov          */
56120f5a302SAlexander Bulekov         uint8_t index;
56220f5a302SAlexander Bulekov         uint8_t stride;
56320f5a302SAlexander Bulekov     } a;
56420f5a302SAlexander Bulekov 
56520f5a302SAlexander Bulekov     if (len < sizeof(a) + 1) {
56620f5a302SAlexander Bulekov         return;
56720f5a302SAlexander Bulekov     }
56820f5a302SAlexander Bulekov     memcpy(&a, data, sizeof(a));
56920f5a302SAlexander Bulekov     pattern p = {a.index, a.stride, len - sizeof(a), data + sizeof(a)};
57020f5a302SAlexander Bulekov     p.index = a.index % p.len;
57120f5a302SAlexander Bulekov     g_array_append_val(dma_patterns, p);
57220f5a302SAlexander Bulekov     return;
57320f5a302SAlexander Bulekov }
57420f5a302SAlexander Bulekov 
57520f5a302SAlexander Bulekov static void op_clear_dma_patterns(QTestState *s,
57620f5a302SAlexander Bulekov                                   const unsigned char *data, size_t len)
57720f5a302SAlexander Bulekov {
57820f5a302SAlexander Bulekov     g_array_set_size(dma_patterns, 0);
57920f5a302SAlexander Bulekov     dma_pattern_index = 0;
58020f5a302SAlexander Bulekov }
58120f5a302SAlexander Bulekov 
582da9bf531SAlexander Bulekov static void op_clock_step(QTestState *s, const unsigned char *data, size_t len)
583da9bf531SAlexander Bulekov {
584da9bf531SAlexander Bulekov     qtest_clock_step_next(s);
585da9bf531SAlexander Bulekov }
586da9bf531SAlexander Bulekov 
587ccbd4bc8SAlexander Bulekov static void op_disable_pci(QTestState *s, const unsigned char *data, size_t len)
588ccbd4bc8SAlexander Bulekov {
589ccbd4bc8SAlexander Bulekov     pci_disabled = true;
590ccbd4bc8SAlexander Bulekov }
591ccbd4bc8SAlexander Bulekov 
592da9bf531SAlexander Bulekov static void handle_timeout(int sig)
593da9bf531SAlexander Bulekov {
594da9bf531SAlexander Bulekov     if (qtest_log_enabled) {
595da9bf531SAlexander Bulekov         fprintf(stderr, "[Timeout]\n");
596da9bf531SAlexander Bulekov         fflush(stderr);
597da9bf531SAlexander Bulekov     }
598d0614b8eSAlexander Bulekov 
599d0614b8eSAlexander Bulekov     /*
600d0614b8eSAlexander Bulekov      * If there is a crash, libfuzzer/ASAN forks a child to run an
601d0614b8eSAlexander Bulekov      * "llvm-symbolizer" process for printing out a pretty stacktrace. It
602d0614b8eSAlexander Bulekov      * communicates with this child using a pipe.  If we timeout+Exit, while
603d0614b8eSAlexander Bulekov      * libfuzzer is still communicating with the llvm-symbolizer child, we will
604d0614b8eSAlexander Bulekov      * be left with an orphan llvm-symbolizer process. Sometimes, this appears
605d0614b8eSAlexander Bulekov      * to lead to a deadlock in the forkserver. Use waitpid to check if there
606d0614b8eSAlexander Bulekov      * are any waitable children. If so, exit out of the signal-handler, and
607d0614b8eSAlexander Bulekov      * let libfuzzer finish communicating with the child, and exit, on its own.
608d0614b8eSAlexander Bulekov      */
609d0614b8eSAlexander Bulekov     if (waitpid(-1, NULL, WNOHANG) == 0) {
610d0614b8eSAlexander Bulekov         return;
611d0614b8eSAlexander Bulekov     }
612d0614b8eSAlexander Bulekov 
613da9bf531SAlexander Bulekov     _Exit(0);
614da9bf531SAlexander Bulekov }
615da9bf531SAlexander Bulekov 
616da9bf531SAlexander Bulekov /*
617da9bf531SAlexander Bulekov  * Here, we interpret random bytes from the fuzzer, as a sequence of commands.
618da9bf531SAlexander Bulekov  * Some commands can be variable-width, so we use a separator, SEPARATOR, to
619da9bf531SAlexander Bulekov  * specify the boundaries between commands. SEPARATOR is used to separate
620da9bf531SAlexander Bulekov  * "operations" in the fuzz input. Why use a separator, instead of just using
621da9bf531SAlexander Bulekov  * the operations' length to identify operation boundaries?
622da9bf531SAlexander Bulekov  *   1. This is a simple way to support variable-length operations
623da9bf531SAlexander Bulekov  *   2. This adds "stability" to the input.
624da9bf531SAlexander Bulekov  *      For example take the input "AbBcgDefg", where there is no separator and
625da9bf531SAlexander Bulekov  *      Opcodes are capitalized.
626da9bf531SAlexander Bulekov  *      Simply, by removing the first byte, we end up with a very different
627da9bf531SAlexander Bulekov  *      sequence:
628da9bf531SAlexander Bulekov  *      BbcGdefg...
629da9bf531SAlexander Bulekov  *      By adding a separator, we avoid this problem:
630da9bf531SAlexander Bulekov  *      Ab SEP Bcg SEP Defg -> B SEP Bcg SEP Defg
631da9bf531SAlexander Bulekov  *      Since B uses two additional bytes as operands, the first "B" will be
632da9bf531SAlexander Bulekov  *      ignored. The fuzzer actively tries to reduce inputs, so such unused
633da9bf531SAlexander Bulekov  *      bytes are likely to be pruned, eventually.
634da9bf531SAlexander Bulekov  *
635da9bf531SAlexander Bulekov  *  SEPARATOR is trivial for the fuzzer to discover when using ASan. Optionally,
636da9bf531SAlexander Bulekov  *  SEPARATOR can be manually specified as a dictionary value (see libfuzzer's
637da9bf531SAlexander Bulekov  *  -dict), though this should not be necessary.
638da9bf531SAlexander Bulekov  *
639da9bf531SAlexander Bulekov  * As a result, the stream of bytes is converted into a sequence of commands.
640da9bf531SAlexander Bulekov  * In a simplified example where SEPARATOR is 0xFF:
641da9bf531SAlexander Bulekov  * 00 01 02 FF 03 04 05 06 FF 01 FF ...
642da9bf531SAlexander Bulekov  * becomes this sequence of commands:
643da9bf531SAlexander Bulekov  * 00 01 02    -> op00 (0102)   -> in (0102, 2)
644da9bf531SAlexander Bulekov  * 03 04 05 06 -> op03 (040506) -> write (040506, 3)
645da9bf531SAlexander Bulekov  * 01          -> op01 (-,0)    -> out (-,0)
646da9bf531SAlexander Bulekov  * ...
647da9bf531SAlexander Bulekov  *
648da9bf531SAlexander Bulekov  * Note here that it is the job of the individual opcode functions to check
649da9bf531SAlexander Bulekov  * that enough data was provided. I.e. in the last command out (,0), out needs
650da9bf531SAlexander Bulekov  * to check that there is not enough data provided to select an address/value
651da9bf531SAlexander Bulekov  * for the operation.
652da9bf531SAlexander Bulekov  */
653da9bf531SAlexander Bulekov static void generic_fuzz(QTestState *s, const unsigned char *Data, size_t Size)
654da9bf531SAlexander Bulekov {
655da9bf531SAlexander Bulekov     void (*ops[]) (QTestState *s, const unsigned char* , size_t) = {
656da9bf531SAlexander Bulekov         [OP_IN]                 = op_in,
657da9bf531SAlexander Bulekov         [OP_OUT]                = op_out,
658da9bf531SAlexander Bulekov         [OP_READ]               = op_read,
659da9bf531SAlexander Bulekov         [OP_WRITE]              = op_write,
66005efbf24SAlexander Bulekov         [OP_PCI_READ]           = op_pci_read,
66105efbf24SAlexander Bulekov         [OP_PCI_WRITE]          = op_pci_write,
662ccbd4bc8SAlexander Bulekov         [OP_DISABLE_PCI]        = op_disable_pci,
66320f5a302SAlexander Bulekov         [OP_ADD_DMA_PATTERN]    = op_add_dma_pattern,
66420f5a302SAlexander Bulekov         [OP_CLEAR_DMA_PATTERNS] = op_clear_dma_patterns,
665da9bf531SAlexander Bulekov         [OP_CLOCK_STEP]         = op_clock_step,
666da9bf531SAlexander Bulekov     };
667da9bf531SAlexander Bulekov     const unsigned char *cmd = Data;
668da9bf531SAlexander Bulekov     const unsigned char *nextcmd;
669da9bf531SAlexander Bulekov     size_t cmd_len;
670da9bf531SAlexander Bulekov     uint8_t op;
671da9bf531SAlexander Bulekov 
672da9bf531SAlexander Bulekov     if (fork() == 0) {
673993f52f4SAlexander Bulekov         struct sigaction sact;
674993f52f4SAlexander Bulekov         struct itimerval timer;
675aaa94a1bSAlexander Bulekov         sigset_t set;
676da9bf531SAlexander Bulekov         /*
677da9bf531SAlexander Bulekov          * Sometimes the fuzzer will find inputs that take quite a long time to
678da9bf531SAlexander Bulekov          * process. Often times, these inputs do not result in new coverage.
679da9bf531SAlexander Bulekov          * Even if these inputs might be interesting, they can slow down the
680993f52f4SAlexander Bulekov          * fuzzer, overall. Set a timeout for each command to avoid hurting
681993f52f4SAlexander Bulekov          * performance, too much
682da9bf531SAlexander Bulekov          */
683da9bf531SAlexander Bulekov         if (timeout) {
684da9bf531SAlexander Bulekov 
685da9bf531SAlexander Bulekov             sigemptyset(&sact.sa_mask);
686da9bf531SAlexander Bulekov             sact.sa_flags   = SA_NODEFER;
687da9bf531SAlexander Bulekov             sact.sa_handler = handle_timeout;
688da9bf531SAlexander Bulekov             sigaction(SIGALRM, &sact, NULL);
689da9bf531SAlexander Bulekov 
690aaa94a1bSAlexander Bulekov             sigemptyset(&set);
691aaa94a1bSAlexander Bulekov             sigaddset(&set, SIGALRM);
692aaa94a1bSAlexander Bulekov             pthread_sigmask(SIG_UNBLOCK, &set, NULL);
693aaa94a1bSAlexander Bulekov 
694da9bf531SAlexander Bulekov             memset(&timer, 0, sizeof(timer));
695da9bf531SAlexander Bulekov             timer.it_value.tv_sec = timeout / USEC_IN_SEC;
696da9bf531SAlexander Bulekov             timer.it_value.tv_usec = timeout % USEC_IN_SEC;
697da9bf531SAlexander Bulekov         }
698da9bf531SAlexander Bulekov 
69920f5a302SAlexander Bulekov         op_clear_dma_patterns(s, NULL, 0);
700ccbd4bc8SAlexander Bulekov         pci_disabled = false;
70120f5a302SAlexander Bulekov 
702da9bf531SAlexander Bulekov         while (cmd && Size) {
703993f52f4SAlexander Bulekov             /* Reset the timeout, each time we run a new command */
704993f52f4SAlexander Bulekov             if (timeout) {
70540c0d963SAlexander Bulekov                 setitimer(ITIMER_REAL, &timer, NULL);
706993f52f4SAlexander Bulekov             }
707993f52f4SAlexander Bulekov 
708da9bf531SAlexander Bulekov             /* Get the length until the next command or end of input */
709da9bf531SAlexander Bulekov             nextcmd = memmem(cmd, Size, SEPARATOR, strlen(SEPARATOR));
710da9bf531SAlexander Bulekov             cmd_len = nextcmd ? nextcmd - cmd : Size;
711da9bf531SAlexander Bulekov 
712da9bf531SAlexander Bulekov             if (cmd_len > 0) {
713da9bf531SAlexander Bulekov                 /* Interpret the first byte of the command as an opcode */
714da9bf531SAlexander Bulekov                 op = *cmd % (sizeof(ops) / sizeof((ops)[0]));
715da9bf531SAlexander Bulekov                 ops[op](s, cmd + 1, cmd_len - 1);
716da9bf531SAlexander Bulekov 
717da9bf531SAlexander Bulekov                 /* Run the main loop */
718da9bf531SAlexander Bulekov                 flush_events(s);
719da9bf531SAlexander Bulekov             }
720da9bf531SAlexander Bulekov             /* Advance to the next command */
721da9bf531SAlexander Bulekov             cmd = nextcmd ? nextcmd + sizeof(SEPARATOR) - 1 : nextcmd;
722da9bf531SAlexander Bulekov             Size = Size - (cmd_len + sizeof(SEPARATOR) - 1);
72320f5a302SAlexander Bulekov             g_array_set_size(dma_regions, 0);
724da9bf531SAlexander Bulekov         }
725da9bf531SAlexander Bulekov         _Exit(0);
726da9bf531SAlexander Bulekov     } else {
727da9bf531SAlexander Bulekov         flush_events(s);
728da9bf531SAlexander Bulekov         wait(0);
729da9bf531SAlexander Bulekov     }
730da9bf531SAlexander Bulekov }
731da9bf531SAlexander Bulekov 
732da9bf531SAlexander Bulekov static void usage(void)
733da9bf531SAlexander Bulekov {
734da9bf531SAlexander Bulekov     printf("Please specify the following environment variables:\n");
735da9bf531SAlexander Bulekov     printf("QEMU_FUZZ_ARGS= the command line arguments passed to qemu\n");
736da9bf531SAlexander Bulekov     printf("QEMU_FUZZ_OBJECTS= "
737da9bf531SAlexander Bulekov             "a space separated list of QOM type names for objects to fuzz\n");
73820f5a302SAlexander Bulekov     printf("Optionally: QEMU_AVOID_DOUBLE_FETCH= "
73920f5a302SAlexander Bulekov             "Try to avoid racy DMA double fetch bugs? %d by default\n",
74020f5a302SAlexander Bulekov             avoid_double_fetches);
741da9bf531SAlexander Bulekov     printf("Optionally: QEMU_FUZZ_TIMEOUT= Specify a custom timeout (us). "
742da9bf531SAlexander Bulekov             "0 to disable. %d by default\n", timeout);
743da9bf531SAlexander Bulekov     exit(0);
744da9bf531SAlexander Bulekov }
745da9bf531SAlexander Bulekov 
746da9bf531SAlexander Bulekov static int locate_fuzz_memory_regions(Object *child, void *opaque)
747da9bf531SAlexander Bulekov {
748da9bf531SAlexander Bulekov     MemoryRegion *mr;
749da9bf531SAlexander Bulekov     if (object_dynamic_cast(child, TYPE_MEMORY_REGION)) {
750da9bf531SAlexander Bulekov         mr = MEMORY_REGION(child);
751da9bf531SAlexander Bulekov         if ((memory_region_is_ram(mr) ||
752da9bf531SAlexander Bulekov             memory_region_is_ram_device(mr) ||
753da9bf531SAlexander Bulekov             memory_region_is_rom(mr)) == false) {
754da9bf531SAlexander Bulekov             /*
755da9bf531SAlexander Bulekov              * We don't want duplicate pointers to the same MemoryRegion, so
756da9bf531SAlexander Bulekov              * try to remove copies of the pointer, before adding it.
757da9bf531SAlexander Bulekov              */
758da9bf531SAlexander Bulekov             g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
759da9bf531SAlexander Bulekov         }
760da9bf531SAlexander Bulekov     }
761da9bf531SAlexander Bulekov     return 0;
762da9bf531SAlexander Bulekov }
763da9bf531SAlexander Bulekov 
764da9bf531SAlexander Bulekov static int locate_fuzz_objects(Object *child, void *opaque)
765da9bf531SAlexander Bulekov {
766f2e8b87aSAlexander Bulekov     GString *type_name;
767f2e8b87aSAlexander Bulekov     GString *path_name;
768da9bf531SAlexander Bulekov     char *pattern = opaque;
769f2e8b87aSAlexander Bulekov 
770f2e8b87aSAlexander Bulekov     type_name = g_string_new(object_get_typename(child));
771f2e8b87aSAlexander Bulekov     g_string_ascii_down(type_name);
772f2e8b87aSAlexander Bulekov     if (g_pattern_match_simple(pattern, type_name->str)) {
773da9bf531SAlexander Bulekov         /* Find and save ptrs to any child MemoryRegions */
774da9bf531SAlexander Bulekov         object_child_foreach_recursive(child, locate_fuzz_memory_regions, NULL);
775da9bf531SAlexander Bulekov 
77605efbf24SAlexander Bulekov         /*
77705efbf24SAlexander Bulekov          * We matched an object. If its a PCI device, store a pointer to it so
77805efbf24SAlexander Bulekov          * we can map BARs and fuzz its config space.
77905efbf24SAlexander Bulekov          */
78005efbf24SAlexander Bulekov         if (object_dynamic_cast(OBJECT(child), TYPE_PCI_DEVICE)) {
78105efbf24SAlexander Bulekov             /*
78205efbf24SAlexander Bulekov              * Don't want duplicate pointers to the same PCIDevice, so remove
78305efbf24SAlexander Bulekov              * copies of the pointer, before adding it.
78405efbf24SAlexander Bulekov              */
78505efbf24SAlexander Bulekov             g_ptr_array_remove_fast(fuzzable_pci_devices, PCI_DEVICE(child));
78605efbf24SAlexander Bulekov             g_ptr_array_add(fuzzable_pci_devices, PCI_DEVICE(child));
78705efbf24SAlexander Bulekov         }
788da9bf531SAlexander Bulekov     } else if (object_dynamic_cast(OBJECT(child), TYPE_MEMORY_REGION)) {
789f2e8b87aSAlexander Bulekov         path_name = g_string_new(object_get_canonical_path_component(child));
790f2e8b87aSAlexander Bulekov         g_string_ascii_down(path_name);
791f2e8b87aSAlexander Bulekov         if (g_pattern_match_simple(pattern, path_name->str)) {
792da9bf531SAlexander Bulekov             MemoryRegion *mr;
793da9bf531SAlexander Bulekov             mr = MEMORY_REGION(child);
794da9bf531SAlexander Bulekov             if ((memory_region_is_ram(mr) ||
795da9bf531SAlexander Bulekov                  memory_region_is_ram_device(mr) ||
796da9bf531SAlexander Bulekov                  memory_region_is_rom(mr)) == false) {
797da9bf531SAlexander Bulekov                 g_hash_table_insert(fuzzable_memoryregions, mr, (gpointer)true);
798da9bf531SAlexander Bulekov             }
799da9bf531SAlexander Bulekov         }
800f2e8b87aSAlexander Bulekov         g_string_free(path_name, true);
801da9bf531SAlexander Bulekov     }
802f2e8b87aSAlexander Bulekov     g_string_free(type_name, true);
803da9bf531SAlexander Bulekov     return 0;
804da9bf531SAlexander Bulekov }
805da9bf531SAlexander Bulekov 
806b677001dSAlexander Bulekov 
807b677001dSAlexander Bulekov static void pci_enum(gpointer pcidev, gpointer bus)
808b677001dSAlexander Bulekov {
809b677001dSAlexander Bulekov     PCIDevice *dev = pcidev;
810b677001dSAlexander Bulekov     QPCIDevice *qdev;
811b677001dSAlexander Bulekov     int i;
812b677001dSAlexander Bulekov 
813b677001dSAlexander Bulekov     qdev = qpci_device_find(bus, dev->devfn);
814b677001dSAlexander Bulekov     g_assert(qdev != NULL);
815b677001dSAlexander Bulekov     for (i = 0; i < 6; i++) {
816b677001dSAlexander Bulekov         if (dev->io_regions[i].size) {
817b677001dSAlexander Bulekov             qpci_iomap(qdev, i, NULL);
818b677001dSAlexander Bulekov         }
819b677001dSAlexander Bulekov     }
820b677001dSAlexander Bulekov     qpci_device_enable(qdev);
821b677001dSAlexander Bulekov     g_free(qdev);
822b677001dSAlexander Bulekov }
823b677001dSAlexander Bulekov 
824da9bf531SAlexander Bulekov static void generic_pre_fuzz(QTestState *s)
825da9bf531SAlexander Bulekov {
826da9bf531SAlexander Bulekov     GHashTableIter iter;
827da9bf531SAlexander Bulekov     MemoryRegion *mr;
828b677001dSAlexander Bulekov     QPCIBus *pcibus;
829da9bf531SAlexander Bulekov     char **result;
830f2e8b87aSAlexander Bulekov     GString *name_pattern;
831da9bf531SAlexander Bulekov 
832da9bf531SAlexander Bulekov     if (!getenv("QEMU_FUZZ_OBJECTS")) {
833da9bf531SAlexander Bulekov         usage();
834da9bf531SAlexander Bulekov     }
835da9bf531SAlexander Bulekov     if (getenv("QTEST_LOG")) {
836da9bf531SAlexander Bulekov         qtest_log_enabled = 1;
837da9bf531SAlexander Bulekov     }
83820f5a302SAlexander Bulekov     if (getenv("QEMU_AVOID_DOUBLE_FETCH")) {
83920f5a302SAlexander Bulekov         avoid_double_fetches = 1;
84020f5a302SAlexander Bulekov     }
841da9bf531SAlexander Bulekov     if (getenv("QEMU_FUZZ_TIMEOUT")) {
842da9bf531SAlexander Bulekov         timeout = g_ascii_strtoll(getenv("QEMU_FUZZ_TIMEOUT"), NULL, 0);
843da9bf531SAlexander Bulekov     }
84420f5a302SAlexander Bulekov     qts_global = s;
84520f5a302SAlexander Bulekov 
84625d309fbSAlexander Bulekov     /*
84725d309fbSAlexander Bulekov      * Create a special device that we can use to back DMA buffers at very
84825d309fbSAlexander Bulekov      * high memory addresses
84925d309fbSAlexander Bulekov      */
85025d309fbSAlexander Bulekov     sparse_mem_mr = sparse_mem_init(0, UINT64_MAX);
85125d309fbSAlexander Bulekov 
85220f5a302SAlexander Bulekov     dma_regions = g_array_new(false, false, sizeof(address_range));
85320f5a302SAlexander Bulekov     dma_patterns = g_array_new(false, false, sizeof(pattern));
854da9bf531SAlexander Bulekov 
855da9bf531SAlexander Bulekov     fuzzable_memoryregions = g_hash_table_new(NULL, NULL);
85605efbf24SAlexander Bulekov     fuzzable_pci_devices   = g_ptr_array_new();
857da9bf531SAlexander Bulekov 
858da9bf531SAlexander Bulekov     result = g_strsplit(getenv("QEMU_FUZZ_OBJECTS"), " ", -1);
859da9bf531SAlexander Bulekov     for (int i = 0; result[i] != NULL; i++) {
860f2e8b87aSAlexander Bulekov         name_pattern = g_string_new(result[i]);
861f2e8b87aSAlexander Bulekov         /*
862f2e8b87aSAlexander Bulekov          * Make the pattern lowercase. We do the same for all the MemoryRegion
863f2e8b87aSAlexander Bulekov          * and Type names so the configs are case-insensitive.
864f2e8b87aSAlexander Bulekov          */
865f2e8b87aSAlexander Bulekov         g_string_ascii_down(name_pattern);
866da9bf531SAlexander Bulekov         printf("Matching objects by name %s\n", result[i]);
867da9bf531SAlexander Bulekov         object_child_foreach_recursive(qdev_get_machine(),
868da9bf531SAlexander Bulekov                                     locate_fuzz_objects,
869f2e8b87aSAlexander Bulekov                                     name_pattern->str);
870f2e8b87aSAlexander Bulekov         g_string_free(name_pattern, true);
871da9bf531SAlexander Bulekov     }
872da9bf531SAlexander Bulekov     g_strfreev(result);
873da9bf531SAlexander Bulekov     printf("This process will try to fuzz the following MemoryRegions:\n");
874da9bf531SAlexander Bulekov 
875da9bf531SAlexander Bulekov     g_hash_table_iter_init(&iter, fuzzable_memoryregions);
876da9bf531SAlexander Bulekov     while (g_hash_table_iter_next(&iter, (gpointer)&mr, NULL)) {
877a8fbec7eSPhilippe Mathieu-Daudé         printf("  * %s (size 0x%" PRIx64 ")\n",
878da9bf531SAlexander Bulekov                object_get_canonical_path_component(&(mr->parent_obj)),
879a8fbec7eSPhilippe Mathieu-Daudé                memory_region_size(mr));
880da9bf531SAlexander Bulekov     }
881da9bf531SAlexander Bulekov 
882da9bf531SAlexander Bulekov     if (!g_hash_table_size(fuzzable_memoryregions)) {
883da9bf531SAlexander Bulekov         printf("No fuzzable memory regions found...\n");
884da9bf531SAlexander Bulekov         exit(1);
885da9bf531SAlexander Bulekov     }
886da9bf531SAlexander Bulekov 
887b677001dSAlexander Bulekov     pcibus = qpci_new_pc(s, NULL);
888b677001dSAlexander Bulekov     g_ptr_array_foreach(fuzzable_pci_devices, pci_enum, pcibus);
889b677001dSAlexander Bulekov     qpci_free_pc(pcibus);
890b677001dSAlexander Bulekov 
891da9bf531SAlexander Bulekov     counter_shm_init();
892da9bf531SAlexander Bulekov }
893da9bf531SAlexander Bulekov 
894a2539322SAlexander Bulekov /*
895a2539322SAlexander Bulekov  * When libfuzzer gives us two inputs to combine, return a new input with the
896a2539322SAlexander Bulekov  * following structure:
897a2539322SAlexander Bulekov  *
898a2539322SAlexander Bulekov  * Input 1 (data1)
899a2539322SAlexander Bulekov  * SEPARATOR
900a2539322SAlexander Bulekov  * Clear out the DMA Patterns
901a2539322SAlexander Bulekov  * SEPARATOR
902a2539322SAlexander Bulekov  * Disable the pci_read/write instructions
903a2539322SAlexander Bulekov  * SEPARATOR
904a2539322SAlexander Bulekov  * Input 2 (data2)
905a2539322SAlexander Bulekov  *
906a2539322SAlexander Bulekov  * The idea is to collate the core behaviors of the two inputs.
907a2539322SAlexander Bulekov  * For example:
908a2539322SAlexander Bulekov  * Input 1: maps a device's BARs, sets up three DMA patterns, and triggers
909a2539322SAlexander Bulekov  *          device functionality A
910a2539322SAlexander Bulekov  * Input 2: maps a device's BARs, sets up one DMA pattern, and triggers device
911a2539322SAlexander Bulekov  *          functionality B
912a2539322SAlexander Bulekov  *
913a2539322SAlexander Bulekov  * This function attempts to produce an input that:
914a2539322SAlexander Bulekov  * Ouptut: maps a device's BARs, set up three DMA patterns, triggers
915a2539322SAlexander Bulekov  *          functionality A device, replaces the DMA patterns with a single
916a2539322SAlexander Bulekov  *          patten, and triggers device functionality B.
917a2539322SAlexander Bulekov  */
918a2539322SAlexander Bulekov static size_t generic_fuzz_crossover(const uint8_t *data1, size_t size1, const
919a2539322SAlexander Bulekov                                      uint8_t *data2, size_t size2, uint8_t *out,
920a2539322SAlexander Bulekov                                      size_t max_out_size, unsigned int seed)
921a2539322SAlexander Bulekov {
922a2539322SAlexander Bulekov     size_t copy_len = 0, size = 0;
923a2539322SAlexander Bulekov 
924a2539322SAlexander Bulekov     /* Check that we have enough space for data1 and at least part of data2 */
925a2539322SAlexander Bulekov     if (max_out_size <= size1 + strlen(SEPARATOR) * 3 + 2) {
926a2539322SAlexander Bulekov         return 0;
927a2539322SAlexander Bulekov     }
928a2539322SAlexander Bulekov 
929a2539322SAlexander Bulekov     /* Copy_Len in the first input */
930a2539322SAlexander Bulekov     copy_len = size1;
931a2539322SAlexander Bulekov     memcpy(out + size, data1, copy_len);
932a2539322SAlexander Bulekov     size += copy_len;
933a2539322SAlexander Bulekov     max_out_size -= copy_len;
934a2539322SAlexander Bulekov 
935a2539322SAlexander Bulekov     /* Append a separator */
936a2539322SAlexander Bulekov     copy_len = strlen(SEPARATOR);
937a2539322SAlexander Bulekov     memcpy(out + size, SEPARATOR, copy_len);
938a2539322SAlexander Bulekov     size += copy_len;
939a2539322SAlexander Bulekov     max_out_size -= copy_len;
940a2539322SAlexander Bulekov 
941a2539322SAlexander Bulekov     /* Clear out the DMA Patterns */
942a2539322SAlexander Bulekov     copy_len = 1;
943a2539322SAlexander Bulekov     if (copy_len) {
944a2539322SAlexander Bulekov         out[size] = OP_CLEAR_DMA_PATTERNS;
945a2539322SAlexander Bulekov     }
946a2539322SAlexander Bulekov     size += copy_len;
947a2539322SAlexander Bulekov     max_out_size -= copy_len;
948a2539322SAlexander Bulekov 
949a2539322SAlexander Bulekov     /* Append a separator */
950a2539322SAlexander Bulekov     copy_len = strlen(SEPARATOR);
951a2539322SAlexander Bulekov     memcpy(out + size, SEPARATOR, copy_len);
952a2539322SAlexander Bulekov     size += copy_len;
953a2539322SAlexander Bulekov     max_out_size -= copy_len;
954a2539322SAlexander Bulekov 
955a2539322SAlexander Bulekov     /* Disable PCI ops. Assume data1 took care of setting up PCI */
956a2539322SAlexander Bulekov     copy_len = 1;
957a2539322SAlexander Bulekov     if (copy_len) {
958a2539322SAlexander Bulekov         out[size] = OP_DISABLE_PCI;
959a2539322SAlexander Bulekov     }
960a2539322SAlexander Bulekov     size += copy_len;
961a2539322SAlexander Bulekov     max_out_size -= copy_len;
962a2539322SAlexander Bulekov 
963a2539322SAlexander Bulekov     /* Append a separator */
964a2539322SAlexander Bulekov     copy_len = strlen(SEPARATOR);
965a2539322SAlexander Bulekov     memcpy(out + size, SEPARATOR, copy_len);
966a2539322SAlexander Bulekov     size += copy_len;
967a2539322SAlexander Bulekov     max_out_size -= copy_len;
968a2539322SAlexander Bulekov 
969a2539322SAlexander Bulekov     /* Copy_Len over the second input */
970a2539322SAlexander Bulekov     copy_len = MIN(size2, max_out_size);
971a2539322SAlexander Bulekov     memcpy(out + size, data2, copy_len);
972a2539322SAlexander Bulekov     size += copy_len;
973a2539322SAlexander Bulekov     max_out_size -= copy_len;
974a2539322SAlexander Bulekov 
975a2539322SAlexander Bulekov     return  size;
976a2539322SAlexander Bulekov }
977a2539322SAlexander Bulekov 
978a2539322SAlexander Bulekov 
979da9bf531SAlexander Bulekov static GString *generic_fuzz_cmdline(FuzzTarget *t)
980da9bf531SAlexander Bulekov {
981da9bf531SAlexander Bulekov     GString *cmd_line = g_string_new(TARGET_NAME);
982da9bf531SAlexander Bulekov     if (!getenv("QEMU_FUZZ_ARGS")) {
983da9bf531SAlexander Bulekov         usage();
984da9bf531SAlexander Bulekov     }
985da9bf531SAlexander Bulekov     g_string_append_printf(cmd_line, " -display none \
986da9bf531SAlexander Bulekov                                       -machine accel=qtest, \
987da9bf531SAlexander Bulekov                                       -m 512M %s ", getenv("QEMU_FUZZ_ARGS"));
988da9bf531SAlexander Bulekov     return cmd_line;
989da9bf531SAlexander Bulekov }
990da9bf531SAlexander Bulekov 
9917fdb5053SAlexander Bulekov static GString *generic_fuzz_predefined_config_cmdline(FuzzTarget *t)
9927fdb5053SAlexander Bulekov {
9938630b43fSAlexander Bulekov     gchar *args;
9947fdb5053SAlexander Bulekov     const generic_fuzz_config *config;
9957fdb5053SAlexander Bulekov     g_assert(t->opaque);
9967fdb5053SAlexander Bulekov 
9977fdb5053SAlexander Bulekov     config = t->opaque;
998a47ea61dSBin Meng     g_setenv("QEMU_AVOID_DOUBLE_FETCH", "1", 1);
9998630b43fSAlexander Bulekov     if (config->argfunc) {
10008630b43fSAlexander Bulekov         args = config->argfunc();
1001a47ea61dSBin Meng         g_setenv("QEMU_FUZZ_ARGS", args, 1);
10028630b43fSAlexander Bulekov         g_free(args);
10038630b43fSAlexander Bulekov     } else {
10048630b43fSAlexander Bulekov         g_assert_nonnull(config->args);
1005a47ea61dSBin Meng         g_setenv("QEMU_FUZZ_ARGS", config->args, 1);
10068630b43fSAlexander Bulekov     }
1007a47ea61dSBin Meng     g_setenv("QEMU_FUZZ_OBJECTS", config->objects, 1);
10087fdb5053SAlexander Bulekov     return generic_fuzz_cmdline(t);
10097fdb5053SAlexander Bulekov }
10107fdb5053SAlexander Bulekov 
1011da9bf531SAlexander Bulekov static void register_generic_fuzz_targets(void)
1012da9bf531SAlexander Bulekov {
1013da9bf531SAlexander Bulekov     fuzz_add_target(&(FuzzTarget){
1014da9bf531SAlexander Bulekov             .name = "generic-fuzz",
1015da9bf531SAlexander Bulekov             .description = "Fuzz based on any qemu command-line args. ",
1016da9bf531SAlexander Bulekov             .get_init_cmdline = generic_fuzz_cmdline,
1017da9bf531SAlexander Bulekov             .pre_fuzz = generic_pre_fuzz,
1018da9bf531SAlexander Bulekov             .fuzz = generic_fuzz,
1019a2539322SAlexander Bulekov             .crossover = generic_fuzz_crossover
1020da9bf531SAlexander Bulekov     });
10217fdb5053SAlexander Bulekov 
10227fdb5053SAlexander Bulekov     GString *name;
10237fdb5053SAlexander Bulekov     const generic_fuzz_config *config;
10247fdb5053SAlexander Bulekov 
10257fdb5053SAlexander Bulekov     for (int i = 0;
10267fdb5053SAlexander Bulekov          i < sizeof(predefined_configs) / sizeof(generic_fuzz_config);
10277fdb5053SAlexander Bulekov          i++) {
10287fdb5053SAlexander Bulekov         config = predefined_configs + i;
10297fdb5053SAlexander Bulekov         name = g_string_new("generic-fuzz");
10307fdb5053SAlexander Bulekov         g_string_append_printf(name, "-%s", config->name);
10317fdb5053SAlexander Bulekov         fuzz_add_target(&(FuzzTarget){
10327fdb5053SAlexander Bulekov                 .name = name->str,
10337fdb5053SAlexander Bulekov                 .description = "Predefined generic-fuzz config.",
10347fdb5053SAlexander Bulekov                 .get_init_cmdline = generic_fuzz_predefined_config_cmdline,
10357fdb5053SAlexander Bulekov                 .pre_fuzz = generic_pre_fuzz,
10367fdb5053SAlexander Bulekov                 .fuzz = generic_fuzz,
10377fdb5053SAlexander Bulekov                 .crossover = generic_fuzz_crossover,
10387fdb5053SAlexander Bulekov                 .opaque = (void *)config
10397fdb5053SAlexander Bulekov         });
10407fdb5053SAlexander Bulekov     }
1041da9bf531SAlexander Bulekov }
1042da9bf531SAlexander Bulekov 
1043da9bf531SAlexander Bulekov fuzz_target_init(register_generic_fuzz_targets);
1044