| /linux/arch/arm64/mm/ |
| H A D | mmap.c | 22 [VM_EXEC] = PAGE_READONLY_EXEC, 23 [VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 24 [VM_EXEC | VM_WRITE] = PAGE_READONLY_EXEC, 25 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_READONLY_EXEC, 31 [VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC, 32 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 33 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC, 34 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC 72 protection_map[VM_EXEC] = PAGE_EXECONLY; in adjust_protection_map() 73 protection_map[VM_EXEC | VM_SHARED] = PAGE_EXECONLY; in adjust_protection_map() [all …]
|
| /linux/arch/x86/mm/ |
| H A D | pgprot.c | 13 [VM_EXEC] = PAGE_READONLY_EXEC, 14 [VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 15 [VM_EXEC | VM_WRITE] = PAGE_COPY_EXEC, 16 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_EXEC, 21 [VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC, 22 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 23 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC, 24 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC 38 (VM_READ|VM_WRITE|VM_EXEC|VM_SHARED)]); in vm_get_page_prot()
|
| /linux/arch/arc/mm/ |
| H A D | mmap.c | 67 [VM_EXEC] = PAGE_U_X_R, 68 [VM_EXEC | VM_READ] = PAGE_U_X_R, 69 [VM_EXEC | VM_WRITE] = PAGE_U_X_R, 70 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_U_X_R, 75 [VM_SHARED | VM_EXEC] = PAGE_U_X_R, 76 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_U_X_R, 77 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_U_X_W_R, 78 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_U_X_W_R
|
| /linux/arch/nios2/mm/ |
| H A D | init.c | 105 VM_READ | VM_EXEC | VM_MAYREAD | in arch_setup_additional_pages() 123 [VM_EXEC] = MKP(1, 0, 0), 124 [VM_EXEC | VM_READ] = MKP(1, 0, 1), 125 [VM_EXEC | VM_WRITE] = MKP(1, 0, 0), 126 [VM_EXEC | VM_WRITE | VM_READ] = MKP(1, 0, 1), 131 [VM_SHARED | VM_EXEC] = MKP(1, 0, 0), 132 [VM_SHARED | VM_EXEC | VM_READ] = MKP(1, 0, 1), 133 [VM_SHARED | VM_EXEC | VM_WRITE] = MKP(1, 1, 0), 134 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = MKP(1, 1, 1)
|
| H A D | cacheflush.c | 138 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 159 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 236 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 268 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 279 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
| /linux/arch/m68k/mm/ |
| H A D | sun3mmu.c | 98 [VM_EXEC] = PAGE_READONLY, 99 [VM_EXEC | VM_READ] = PAGE_READONLY, 100 [VM_EXEC | VM_WRITE] = PAGE_COPY, 101 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY, 106 [VM_SHARED | VM_EXEC] = PAGE_READONLY, 107 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY, 108 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED, 109 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
|
| H A D | mcfmmu.c | 233 [VM_EXEC] = __pgprot(CF_PAGE_VALID | 236 [VM_EXEC | VM_READ] = __pgprot(CF_PAGE_VALID | 240 [VM_EXEC | VM_WRITE] = __pgprot(CF_PAGE_VALID | 244 [VM_EXEC | VM_WRITE | VM_READ] = __pgprot(CF_PAGE_VALID | 258 [VM_SHARED | VM_EXEC] = __pgprot(CF_PAGE_VALID | 261 [VM_SHARED | VM_EXEC | VM_READ] = __pgprot(CF_PAGE_VALID | 265 [VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(CF_PAGE_VALID | 269 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(CF_PAGE_VALID |
|
| H A D | motorola.c | 411 [VM_EXEC] = PAGE_READONLY_C, 412 [VM_EXEC | VM_READ] = PAGE_READONLY_C, 413 [VM_EXEC | VM_WRITE] = PAGE_COPY_C, 414 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_C, 419 [VM_SHARED | VM_EXEC] = PAGE_READONLY_C, 420 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_C, 421 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_C, 422 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_C
|
| /linux/arch/xtensa/mm/ |
| H A D | init.c | 188 [VM_EXEC] = PAGE_READONLY_EXEC, 189 [VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 190 [VM_EXEC | VM_WRITE] = PAGE_COPY_EXEC, 191 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_EXEC, 196 [VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC, 197 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC, 198 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC, 199 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC
|
| /linux/arch/csky/mm/ |
| H A D | init.c | 139 [VM_EXEC] = PAGE_READ, 140 [VM_EXEC | VM_READ] = PAGE_READ, 141 [VM_EXEC | VM_WRITE] = PAGE_READ, 142 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_READ, 147 [VM_SHARED | VM_EXEC] = PAGE_READ, 148 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READ, 149 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_WRITE, 150 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_WRITE
|
| /linux/arch/microblaze/mm/ |
| H A D | init.c | 248 [VM_EXEC] = PAGE_READONLY, 249 [VM_EXEC | VM_READ] = PAGE_READONLY_X, 250 [VM_EXEC | VM_WRITE] = PAGE_COPY, 251 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X, 256 [VM_SHARED | VM_EXEC] = PAGE_READONLY, 257 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X, 258 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED, 259 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
|
| /linux/arch/sh/mm/ |
| H A D | mmap.c | 27 [VM_EXEC] = PAGE_EXECREAD, 28 [VM_EXEC | VM_READ] = PAGE_EXECREAD, 29 [VM_EXEC | VM_WRITE] = PAGE_COPY, 30 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY, 35 [VM_SHARED | VM_EXEC] = PAGE_EXECREAD, 36 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_EXECREAD, 37 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_RWX, 38 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_RWX
|
| /linux/arch/openrisc/mm/ |
| H A D | init.c | 251 [VM_EXEC] = PAGE_READONLY, 252 [VM_EXEC | VM_READ] = PAGE_READONLY_X, 253 [VM_EXEC | VM_WRITE] = PAGE_COPY, 254 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X, 259 [VM_SHARED | VM_EXEC] = PAGE_READONLY, 260 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X, 261 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED, 262 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
|
| /linux/arch/hexagon/mm/ |
| H A D | init.c | 222 [VM_EXEC] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 224 [VM_EXEC | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 227 [VM_EXEC | VM_WRITE] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 229 [VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 241 [VM_SHARED | VM_EXEC] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 243 [VM_SHARED | VM_EXEC | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 246 [VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(_PAGE_PRESENT | _PAGE_USER | 249 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
|
| /linux/arch/loongarch/mm/ |
| H A D | cache.c | 174 [VM_EXEC] = __pgprot(_CACHE_CC | _PAGE_VALID | 176 [VM_EXEC | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 178 [VM_EXEC | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID | 180 [VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 194 [VM_SHARED | VM_EXEC] = __pgprot(_CACHE_CC | _PAGE_VALID | 196 [VM_SHARED | VM_EXEC | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID | 198 [VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID | 201 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID |
|
| /linux/arch/s390/mm/ |
| H A D | mmap.c | 202 pm[VM_EXEC] = PAGE_RX; in setup_protection_map() 203 pm[VM_EXEC | VM_READ] = PAGE_RX; in setup_protection_map() 204 pm[VM_EXEC | VM_WRITE] = PAGE_RX; in setup_protection_map() 205 pm[VM_EXEC | VM_WRITE | VM_READ] = PAGE_RX; in setup_protection_map() 210 pm[VM_SHARED | VM_EXEC] = PAGE_RX; in setup_protection_map() 211 pm[VM_SHARED | VM_EXEC | VM_READ] = PAGE_RX; in setup_protection_map() 212 pm[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_RWX; in setup_protection_map() 213 pm[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_RWX; in setup_protection_map()
|
| /linux/arch/um/kernel/ |
| H A D | mem.c | 241 [VM_EXEC] = PAGE_READONLY, 242 [VM_EXEC | VM_READ] = PAGE_READONLY, 243 [VM_EXEC | VM_WRITE] = PAGE_COPY, 244 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY, 249 [VM_SHARED | VM_EXEC] = PAGE_READONLY, 250 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY, 251 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED, 252 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
|
| /linux/arch/alpha/mm/ |
| H A D | init.c | 255 [VM_EXEC] = _PAGE_P(_PAGE_FOW | _PAGE_FOR), 256 [VM_EXEC | VM_READ] = _PAGE_P(_PAGE_FOW), 257 [VM_EXEC | VM_WRITE] = _PAGE_P(0), 258 [VM_EXEC | VM_WRITE | VM_READ] = _PAGE_P(0), 264 [VM_SHARED | VM_EXEC] = _PAGE_S(_PAGE_FOW | _PAGE_FOR), 265 [VM_SHARED | VM_EXEC | VM_READ] = _PAGE_S(_PAGE_FOW), 266 [VM_SHARED | VM_EXEC | VM_WRITE] = _PAGE_S(0), 267 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = _PAGE_S(0)
|
| /linux/arch/sparc/mm/ |
| H A D | init_32.c | 290 [VM_EXEC] = PAGE_READONLY, 291 [VM_EXEC | VM_READ] = PAGE_READONLY, 292 [VM_EXEC | VM_WRITE] = PAGE_COPY, 293 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY, 298 [VM_SHARED | VM_EXEC] = PAGE_READONLY, 299 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY, 300 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED, 301 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
|
| /linux/arch/powerpc/mm/ |
| H A D | pgtable.c | 168 if (WARN_ON(!(vma->vm_flags & VM_EXEC))) in set_access_flags_filter() 539 [VM_EXEC] = PAGE_EXECONLY_X, 540 [VM_EXEC | VM_READ] = PAGE_READONLY_X, 541 [VM_EXEC | VM_WRITE] = PAGE_COPY_X, 542 [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X, 547 [VM_SHARED | VM_EXEC] = PAGE_EXECONLY_X, 548 [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X, 549 [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_X, 550 [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
|
| /linux/include/linux/ |
| H A D | mman.h | 145 _calc_vm_trans(prot, PROT_EXEC, VM_EXEC) | in calc_vm_prot_bits() 208 if (!(new & VM_EXEC)) in map_deny_write_exec() 216 if (!(old & VM_EXEC)) in map_deny_write_exec()
|
| /linux/arch/arm/mm/ |
| H A D | cache-v4wt.S | 67 mov r2, #VM_EXEC 70 tst r2, #VM_EXEC 92 tst r2, #VM_EXEC
|
| H A D | cache-fa.S | 64 mov r2, #VM_EXEC 67 tst r2, #VM_EXEC 91 1: tst r2, #VM_EXEC 97 tst r2, #VM_EXEC
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | find_vma.c | 13 #define VM_EXEC 0x00000004 macro 31 if (vma->vm_flags & VM_EXEC) in check_vma()
|
| /linux/arch/csky/abiv1/ |
| H A D | cacheflush.c | 63 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 73 if (vma->vm_flags & VM_EXEC) in flush_cache_range()
|