| /linux/kernel/livepatch/ |
| H A D | shadow.c | 70 static inline bool klp_shadow_match(struct klp_shadow *shadow, void *obj, in klp_shadow_match() argument 73 return shadow->obj == obj && shadow->id == id; in klp_shadow_match() 85 struct klp_shadow *shadow; in klp_shadow_get() local 89 hash_for_each_possible_rcu(klp_shadow_hash, shadow, node, in klp_shadow_get() 92 if (klp_shadow_match(shadow, obj, id)) { in klp_shadow_get() 94 return shadow->data; in klp_shadow_get() 234 static void klp_shadow_free_struct(struct klp_shadow *shadow, in klp_shadow_free_struct() argument 237 hash_del_rcu(&shadow->node); in klp_shadow_free_struct() 239 dtor(shadow->obj, shadow->data); in klp_shadow_free_struct() 240 kfree_rcu(shadow, rcu_head); in klp_shadow_free_struct() [all …]
|
| /linux/Documentation/arch/riscv/ |
| H A D | zicfiss.rst | 11 to enable shadow stacks for user mode applications on RISC-V. 29 provides an alternate stack (the "shadow stack") on which return 34 - PTE encodings for shadow stack virtual memory 36 PTE.R=0, PTE.W=1, PTE.X=0 becomes the PTE encoding for shadow stack pages. 38 - The ``sspush x1/x5`` instruction pushes (stores) ``x1/x5`` to shadow stack. 40 - The ``sspopchk x1/x5`` instruction pops (loads) from shadow stack and compares 45 x1/x5`` to save the return address on shadow stack in addition to the 48 from the regular stack matches with the popped value from the shadow 54 As mentioned earlier, shadow stacks get new page table encodings that 56 that operate on the shadow stacks: [all …]
|
| /linux/drivers/scsi/ |
| H A D | xen-scsifront.c | 121 struct vscsifrnt_shadow *shadow[VSCSIIF_MAX_REQS]; member 163 info->shadow[id] = NULL; in _scsifront_put_rqid() 182 struct vscsifrnt_shadow *shadow) in scsifront_do_request() argument 186 struct scsi_cmnd *sc = shadow->sc; in scsifront_do_request() 197 info->shadow[id] = shadow; in scsifront_do_request() 198 shadow->rqid = id; in scsifront_do_request() 204 ring_req->act = shadow->act; in scsifront_do_request() 205 ring_req->ref_rqid = shadow->ref_rqid; in scsifront_do_request() 206 ring_req->nr_segments = shadow->nr_segments; in scsifront_do_request() 220 for (i = 0; i < (shadow->nr_segments & ~VSCSIIF_SG_GRANT); i++) in scsifront_do_request() [all …]
|
| /linux/Documentation/livepatch/ |
| H A D | shadow-vars.rst | 6 additional "shadow" data with existing data structures. Shadow data is 8 unmodified. The shadow variable API described in this document is used 9 to allocate/add and remove/free shadow variables to/from their parents. 13 shadow data. The numeric identifier is a simple enumeration that may be 14 used to describe shadow variable version, class or type, etc. More 16 numeric id subsequently filters hashtable queries. Multiple shadow 24 (See the full API usage docbook notes in livepatch/shadow.c.) 26 A hashtable references all shadow variables. These references are 30 meta-data and shadow-data: 37 - data[] - storage for shadow data [all …]
|
| /linux/Documentation/arch/x86/ |
| H A D | shstk.rst | 14 CET introduces shadow stack and indirect branch tracking (IBT). A shadow stack 17 return address to both the normal stack and the shadow stack. Upon 18 function return, the processor pops the shadow stack copy and compares it 23 shadow stack and kernel IBT are supported. 28 To use userspace shadow stack you need HW that supports it, a kernel 31 The kernel Kconfig option is X86_USER_SHADOW_STACK. When compiled in, shadow 34 To build a user shadow stack enabled kernel, Binutils v2.29 or LLVM v6 or later 38 CET. "user_shstk" means that userspace shadow stack is supported on the current 101 Currently shadow stack and WRSS are supported via this interface. WRSS 102 can only be enabled with shadow stack, and is automatically disabled [all …]
|
| /linux/mm/kmsan/ |
| H A D | init.c | 99 struct page *shadow, *origin; member 118 struct page *shadow, *origin; in kmsan_memblock_free_pages() local 120 if (!held_back[order].shadow) { in kmsan_memblock_free_pages() 121 held_back[order].shadow = page; in kmsan_memblock_free_pages() 128 shadow = held_back[order].shadow; in kmsan_memblock_free_pages() 130 kmsan_setup_meta(page, shadow, origin, order); in kmsan_memblock_free_pages() 132 held_back[order].shadow = NULL; in kmsan_memblock_free_pages() 170 struct page *page, *shadow, *origin; in do_collection() local 174 shadow = smallstack_pop(&collect); in do_collection() 176 kmsan_setup_meta(page, shadow, origin, collect.order); in do_collection() [all …]
|
| H A D | shadow.c | 85 void *shadow; in kmsan_get_shadow_origin_ptr() local 97 shadow = kmsan_get_metadata(address, KMSAN_META_SHADOW); in kmsan_get_shadow_origin_ptr() 98 if (!shadow) in kmsan_get_shadow_origin_ptr() 101 ret.shadow = shadow; in kmsan_get_shadow_origin_ptr() 108 ret.shadow = dummy_store_page; in kmsan_get_shadow_origin_ptr() 112 ret.shadow = dummy_load_page; in kmsan_get_shadow_origin_ptr() 173 struct page *shadow, *origin; in kmsan_alloc_page() local 180 shadow = shadow_page_for(page); in kmsan_alloc_page() 184 __memset(page_address(shadow), 0, PAGE_SIZE * pages); in kmsan_alloc_page() 193 __memset(page_address(shadow), -1, PAGE_SIZE * pages); in kmsan_alloc_page() [all …]
|
| H A D | instrumentation.c | 139 static inline void get_param0_metadata(u64 *shadow, in get_param0_metadata() argument 144 *shadow = *(u64 *)(ctx->cstate.param_tls); in get_param0_metadata() 148 static inline void set_retval_metadata(u64 shadow, depot_stack_handle_t origin) in set_retval_metadata() argument 152 *(u64 *)(ctx->cstate.retval_tls) = shadow; in set_retval_metadata() 162 u64 shadow; in __msan_memmove() local 164 get_param0_metadata(&shadow, &origin); in __msan_memmove() 176 set_retval_metadata(shadow, origin); in __msan_memmove() 187 u64 shadow; in __msan_memcpy() local 189 get_param0_metadata(&shadow, &origin); in __msan_memcpy() 203 set_retval_metadata(shadow, origin); in __msan_memcpy() [all …]
|
| H A D | hooks.c | 154 struct page *shadow, *origin; in kmsan_ioremap_page_range() local 164 shadow = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range() 166 if (!shadow || !origin) { in kmsan_ioremap_page_range() 172 vmalloc_shadow(start + off + PAGE_SIZE), prot, &shadow, in kmsan_ioremap_page_range() 178 shadow = NULL; in kmsan_ioremap_page_range() 202 if (shadow) in kmsan_ioremap_page_range() 203 __free_pages(shadow, 1); in kmsan_ioremap_page_range() 222 struct page *shadow, *origin; in kmsan_iounmap_page_range() local 234 shadow = kmsan_vmalloc_to_page_or_null((void *)v_shadow); in kmsan_iounmap_page_range() 238 if (shadow) in kmsan_iounmap_page_range() [all …]
|
| /linux/fs/nilfs2/ |
| H A D | mdt.c | 480 struct nilfs_shadow_map *shadow = mdi->mi_shadow; in nilfs_mdt_clear() local 485 if (shadow) { in nilfs_mdt_clear() 486 struct inode *s_inode = shadow->inode; in nilfs_mdt_clear() 488 shadow->inode = NULL; in nilfs_mdt_clear() 524 struct nilfs_shadow_map *shadow) in nilfs_mdt_setup_shadow_map() argument 529 INIT_LIST_HEAD(&shadow->frozen_buffers); in nilfs_mdt_setup_shadow_map() 535 shadow->inode = s_inode; in nilfs_mdt_setup_shadow_map() 536 mi->mi_shadow = shadow; in nilfs_mdt_setup_shadow_map() 550 struct nilfs_shadow_map *shadow = mi->mi_shadow; in nilfs_mdt_save_to_shadow_map() local 551 struct inode *s_inode = shadow->inode; in nilfs_mdt_save_to_shadow_map() [all …]
|
| /linux/drivers/gpio/ |
| H A D | gpio-janz-ttl.c | 60 u8 *shadow; in ttl_get_value() local 64 shadow = &mod->porta_shadow; in ttl_get_value() 66 shadow = &mod->portb_shadow; in ttl_get_value() 69 shadow = &mod->portc_shadow; in ttl_get_value() 74 ret = *shadow & BIT(offset); in ttl_get_value() 83 u8 *shadow; in ttl_set_value() local 87 shadow = &mod->porta_shadow; in ttl_set_value() 90 shadow = &mod->portb_shadow; in ttl_set_value() 94 shadow = &mod->portc_shadow; in ttl_set_value() 100 *shadow |= BIT(offset); in ttl_set_value() [all …]
|
| H A D | gpio-mm-lantiq.c | 31 u16 shadow; /* shadow the latches state */ member 47 __raw_writew(chip->shadow, chip->regs); in ltq_mm_apply() 65 chip->shadow |= (1 << offset); in ltq_mm_set() 67 chip->shadow &= ~(1 << offset); in ltq_mm_set() 104 u32 shadow; in ltq_mm_probe() local 129 if (!of_property_read_u32(pdev->dev.of_node, "lantiq,shadow", &shadow)) in ltq_mm_probe() 130 chip->shadow = shadow; in ltq_mm_probe()
|
| H A D | gpio-stp-xway.c | 84 u32 shadow; /* shadow the shift registers state */ member 121 chip->shadow |= BIT(gpio); in xway_stp_set() 123 chip->shadow &= ~BIT(gpio); in xway_stp_set() 124 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set() 235 u32 shadow, groups, dsl, phy; in xway_stp_probe() local 258 if (!of_property_read_u32(pdev->dev.of_node, "lantiq,shadow", &shadow)) in xway_stp_probe() 259 chip->shadow = shadow; in xway_stp_probe()
|
| /linux/include/linux/ |
| H A D | scx200_gpio.h | 13 #define __SCx200_GPIO_SHADOW unsigned long *shadow = scx200_gpio_shadow+bank 16 #define __SCx200_GPIO_OUT __asm__ __volatile__("outsl":"=mS" (shadow):"d" (ioaddr), "0" (shadow)) 46 set_bit(index, shadow); /* __set_bit()? */ in scx200_gpio_set_high() 57 clear_bit(index, shadow); /* __clear_bit()? */ in scx200_gpio_set_low() 69 set_bit(index, shadow); in scx200_gpio_set() 71 clear_bit(index, shadow); in scx200_gpio_set() 81 change_bit(index, shadow); in scx200_gpio_change()
|
| /linux/drivers/block/ |
| H A D | xen-blkfront.c | 190 struct blk_shadow shadow[]; member 291 rinfo->shadow_free = rinfo->shadow[free].req.u.rw.id; in get_id_from_freelist() 292 rinfo->shadow[free].req.u.rw.id = 0x0fffffee; /* debug */ in get_id_from_freelist() 299 if (rinfo->shadow[id].req.u.rw.id != id) in add_id_to_freelist() 301 if (rinfo->shadow[id].request == NULL) in add_id_to_freelist() 303 rinfo->shadow[id].req.u.rw.id = rinfo->shadow_free; in add_id_to_freelist() 304 rinfo->shadow[id].request = NULL; in add_id_to_freelist() 543 rinfo->shadow[id].request = req; in blkif_ring_get_request() 544 rinfo->shadow[id].status = REQ_PROCESSING; in blkif_ring_get_request() 545 rinfo->shadow[id].associated_id = NO_ASSOCIATED_ID; in blkif_ring_get_request() [all …]
|
| /linux/drivers/irqchip/ |
| H A D | irq-econet-en751221.c | 74 u8 shadow; in econet_chmask() local 83 shadow = econet_intc.interrupt_shadows[hwirq]; in econet_chmask() 84 if (WARN_ON_ONCE(shadow == IS_SHADOW)) in econet_chmask() 86 else if (shadow != NOT_PERCPU && smp_processor_id() == 1) in econet_chmask() 87 hwirq = shadow; in econet_chmask() 219 u32 shadow = shadows[i + 1]; in get_shadow_interrupts() local 222 if (shadow > IRQ_COUNT) { in get_shadow_interrupts() 224 node, field, i + 1, shadow); in get_shadow_interrupts() 239 if (econet_intc.interrupt_shadows[shadow] != NOT_PERCPU) { in get_shadow_interrupts() 241 node, field, i + 1, shadow); in get_shadow_interrupts() [all …]
|
| /linux/mm/kasan/ |
| H A D | report_sw_tags.c | 51 u8 *shadow; in kasan_get_alloc_size() local 62 shadow = (u8 *)kasan_mem_to_shadow(object); in kasan_get_alloc_size() 64 if (*shadow != KASAN_TAG_INVALID) in kasan_get_alloc_size() 68 shadow++; in kasan_get_alloc_size() 81 u8 *shadow = (u8 *)kasan_mem_to_shadow(addr); in kasan_print_tags() local 83 pr_err("Pointer tag: [%02x], memory tag: [%02x]\n", addr_tag, *shadow); in kasan_print_tags()
|
| H A D | sw_tags.c | 79 u8 *shadow_first, *shadow_last, *shadow; in kasan_check_range() local 115 for (shadow = shadow_first; shadow <= shadow_last; shadow++) { in kasan_check_range() 116 if (*shadow != tag) { in kasan_check_range()
|
| H A D | report_generic.c | 49 u8 *shadow; in kasan_get_alloc_size() local 60 shadow = (u8 *)kasan_mem_to_shadow(object); in kasan_get_alloc_size() 62 if (*shadow == 0) in kasan_get_alloc_size() 64 else if (*shadow >= 1 && *shadow <= KASAN_GRANULE_SIZE - 1) in kasan_get_alloc_size() 65 return size + *shadow; in kasan_get_alloc_size() 68 shadow++; in kasan_get_alloc_size()
|
| /linux/mm/ |
| H A D | workingset.c | 210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, in unpack_shadow() argument 213 unsigned long entry = xa_to_value(shadow); in unpack_shadow() 264 static bool lru_gen_test_recent(void *shadow, struct lruvec **lruvec, in lru_gen_test_recent() argument 272 unpack_shadow(shadow, &memcg_id, &pgdat, token, workingset); in lru_gen_test_recent() 283 static void lru_gen_refault(struct folio *folio, void *shadow) in lru_gen_refault() argument 296 recent = lru_gen_test_recent(shadow, &lruvec, &token, &workingset); in lru_gen_refault() 333 static bool lru_gen_test_recent(void *shadow, struct lruvec **lruvec, in lru_gen_test_recent() argument 339 static void lru_gen_refault(struct folio *folio, void *shadow) in lru_gen_refault() argument 418 bool workingset_test_recent(void *shadow, bool file, bool *workingset, in workingset_test_recent() argument 434 recent = lru_gen_test_recent(shadow, &eviction_lruvec, &eviction, workingset); in workingset_test_recent() [all …]
|
| H A D | swap_table.h | 40 static inline unsigned long shadow_swp_to_tb(void *shadow) in shadow_swp_to_tb() argument 44 VM_WARN_ON_ONCE(shadow && !xa_is_value(shadow)); in shadow_swp_to_tb() 45 return (unsigned long)shadow; in shadow_swp_to_tb()
|
| /linux/Documentation/dev-tools/ |
| H A D | kmsan.rst | 95 incorrect shadow/origin values, likely leading to false positives. Functions 132 KMSAN shadow memory 135 KMSAN associates a metadata byte (also called shadow byte) with every byte of 136 kernel memory. A bit in the shadow byte is set if the corresponding bit of the 138 setting its shadow bytes to ``0xff``) is called poisoning, marking it 139 initialized (setting the shadow bytes to ``0x00``) is called unpoisoning. 146 Compiler instrumentation also tracks the shadow values as they are used along 148 ``mm/kmsan/`` to persist shadow values. 150 The shadow value of a basic or compound type is an array of bytes of the same 152 When a value is read from memory, its shadow memory is also obtained and [all …]
|
| /linux/drivers/ps3/ |
| H A D | ps3-lpm.c | 128 struct ps3_lpm_shadow_regs shadow; member 394 return lpm_priv->shadow.pm_control; in ps3_read_pm() 398 return lpm_priv->shadow.pm_start_stop; in ps3_read_pm() 409 return lpm_priv->shadow.group_control; in ps3_read_pm() 411 return lpm_priv->shadow.debug_bus_control; in ps3_read_pm() 446 if (val != lpm_priv->shadow.group_control) in ps3_write_pm() 451 lpm_priv->shadow.group_control = val; in ps3_write_pm() 454 if (val != lpm_priv->shadow.debug_bus_control) in ps3_write_pm() 459 lpm_priv->shadow.debug_bus_control = val; in ps3_write_pm() 465 if (val != lpm_priv->shadow.pm_control) in ps3_write_pm() [all …]
|
| /linux/Documentation/virt/kvm/x86/ |
| H A D | mmu.rst | 4 The x86 kvm shadow mmu 55 spte shadow pte (referring to pfns) 87 direct mode; otherwise it operates in shadow mode (see below). 121 The principal data structure is the shadow page, 'struct kvm_mmu_page'. A 122 shadow page contains 512 sptes, which can be either leaf or nonleaf sptes. A 123 shadow page may contain a mix of leaf and nonleaf sptes. 126 is not related to a translation directly. It points to other shadow pages. 152 The level in the shadow paging hierarchy that this shadow page belongs to. 166 so multiple shadow pages are needed to shadow one guest page. 167 For first-level shadow pages, role.quadrant can be 0 or 1 and denotes the [all …]
|
| /linux/drivers/usb/host/ |
| H A D | xen-hcd.c | 84 struct usb_shadow shadow[XENUSB_URB_RING_SIZE]; member 565 info->shadow_free = info->shadow[free].req.id; in xenhcd_get_id_from_freelist() 566 info->shadow[free].req.id = 0x0fff; /* debug */ in xenhcd_get_id_from_freelist() 573 info->shadow[id].req.id = info->shadow_free; in xenhcd_add_id_to_freelist() 574 info->shadow[id].urb = NULL; in xenhcd_add_id_to_freelist() 718 struct usb_shadow *shadow = info->shadow + id; in xenhcd_gnttab_done() local 722 if (!shadow->in_flight) { in xenhcd_gnttab_done() 726 shadow->in_flight = false; in xenhcd_gnttab_done() 728 nr_segs = shadow->req.nr_buffer_segs; in xenhcd_gnttab_done() 730 if (xenusb_pipeisoc(shadow->req.pipe)) in xenhcd_gnttab_done() [all …]
|