| /linux/lib/tests/ |
| H A D | usercopy_kunit.c | 36 char *kmem; member 52 char *kmem = priv->kmem; in usercopy_test_check_nonzero_user() local 66 kmem += start; in usercopy_test_check_nonzero_user() 83 memset(kmem, 0x0, size); in usercopy_test_check_nonzero_user() 85 kmem[i] = 0xff; in usercopy_test_check_nonzero_user() 87 kmem[i] = 0xff; in usercopy_test_check_nonzero_user() 89 KUNIT_EXPECT_EQ_MSG(test, copy_to_user(umem, kmem, size), 0, in usercopy_test_check_nonzero_user() 96 int expected = is_zeroed(kmem + start, len); in usercopy_test_check_nonzero_user() 111 char *kmem = priv->kmem; in usercopy_test_copy_struct_from_user() local 132 memset(kmem, 0x0, size); in usercopy_test_copy_struct_from_user() [all …]
|
| /linux/tools/perf/Documentation/ |
| H A D | examples.txt | 14 kmem:kmalloc [Tracepoint event] 15 kmem:kmem_cache_alloc [Tracepoint event] 16 kmem:kmalloc_node [Tracepoint event] 17 kmem:kmem_cache_alloc_node [Tracepoint event] 18 kmem:kfree [Tracepoint event] 19 kmem:kmem_cache_free [Tracepoint event] 20 kmem:mm_page_free [Tracepoint event] 21 kmem:mm_page_free_batched [Tracepoint event] 22 kmem:mm_page_alloc [Tracepoint event] 23 kmem:mm_page_alloc_zone_locked [Tracepoint event] [all …]
|
| H A D | perf-kmem.txt | 1 perf-kmem(1) 6 perf-kmem - Tool to trace/measure kernel memory properties 11 'perf kmem' [<options>] {record|stat} 15 There are two variants of perf kmem: 17 'perf kmem [<options>] record [<perf-record-options>] <command>' to 18 record the kmem events of an arbitrary workload. Additional 'perf 22 'perf kmem [<options>] stat' to report kernel memory statistics. 67 Show live page stat. The perf kmem shows total allocation stat by
|
| /linux/Documentation/trace/ |
| H A D | tracepoint-analysis.rst | 98 -e kmem:mm_page_alloc -e kmem:mm_page_free \ 99 -e kmem:mm_page_free_batched \ 103 9630 kmem:mm_page_alloc 104 2143 kmem:mm_page_free 105 7424 kmem:mm_page_free_batched 125 $ perf stat -e kmem:mm_page_alloc -e kmem:mm_page_free \ 126 -e kmem:mm_page_free_batched ./hackbench 10 131 17803 kmem:mm_page_alloc 132 12398 kmem:mm_page_free 133 4827 kmem:mm_page_free_batched [all …]
|
| H A D | events.rst | 461 # echo 'enable_event:kmem:kmalloc:1' > \ 468 # echo 'disable_event:kmem:kmalloc' > \ 478 # echo '!enable_event:kmem:kmalloc:1' > \ 481 # echo '!disable_event:kmem:kmalloc' > \ 487 kmem:kmalloc and sched:sched_switch, but can't have two kmem:kmalloc 488 versions such as kmem:kmalloc and kmem:kmalloc:1 or 'kmem:kmalloc if 489 bytes_req == 256' and 'kmem:kmalloc if bytes_alloc == 256' (they 490 could be combined into a single filter on kmem:kmalloc though). 501 /sys/kernel/tracing/events/kmem/kmalloc/trigger 507 /sys/kernel/tracing/events/kmem/kmalloc/trigger [all …]
|
| /linux/drivers/md/dm-pcache/ |
| H A D | backing_dev.c | 97 if (backing_req->kmem.bvecs != backing_req->kmem.inline_bvecs) in backing_dev_req_end() 98 mempool_free(backing_req->kmem.bvecs, &backing_dev->bvec_pool); in backing_dev_req_end() 227 u32 n_vecs = bio_add_max_vecs(opts->kmem.data, opts->kmem.len); in kmem_type_req_alloc() 236 backing_req->kmem.bvecs = mempool_alloc(&backing_dev->bvec_pool, opts->gfp_mask); in kmem_type_req_alloc() 237 if (!backing_req->kmem.bvecs) in kmem_type_req_alloc() 240 backing_req->kmem.bvecs = backing_req->kmem.inline_bvecs; in kmem_type_req_alloc() 243 backing_req->kmem.n_vecs = n_vecs; in kmem_type_req_alloc() 298 bio_init(&backing_req->bio, backing_dev->dm_dev->bdev, backing_req->kmem.bvecs, in kmem_type_req_init() 299 backing_req->kmem.n_vecs, opts->kmem.opf); in kmem_type_req_init() 302 bio_map(backing_bio, opts->kmem.data, opts->kmem.len); in kmem_type_req_init() [all …]
|
| H A D | cache_writeback.c | 126 writeback_req_opts.kmem.data = addr; in cache_key_writeback() 127 writeback_req_opts.kmem.opf = REQ_OP_WRITE; in cache_key_writeback() 128 writeback_req_opts.kmem.len = req_len; in cache_key_writeback() 129 writeback_req_opts.kmem.backing_off = key->off + done; in cache_key_writeback()
|
| H A D | backing_dev.h | 38 } kmem; member 80 } kmem; member
|
| /linux/tools/testing/selftests/ftrace/test.d/trigger/ |
| H A D | trigger-hist.tc | 44 echo 'hist:keys=call_site.sym' > events/kmem/kmalloc/trigger 46 grep '{ call_site: \[[0-9a-f][0-9a-f]*\] [_a-zA-Z][_a-zA-Z]* *}' events/kmem/kmalloc/hist > /dev/nu… 53 echo 'hist:keys=call_site.sym-offset' > events/kmem/kmalloc/trigger 55 grep '{ call_site: \[[0-9a-f][0-9a-f]*\] [_a-zA-Z][_a-zA-Z]*+0x[0-9a-f][0-9a-f]*' events/kmem/kmall…
|
| H A D | trigger-hist-mod.tc | 45 echo 'hist:keys=bytes_req.log2' > events/kmem/kmalloc/trigger 47 grep 'bytes_req: ~ 2^[0-9]*' events/kmem/kmalloc/hist > /dev/null || \
|
| /linux/tools/testing/selftests/ftrace/test.d/filter/ |
| H A D | event-filter-function.tc | 13 echo 1 > events/kmem/kmem_cache_free/enable 17 echo 0 > events/kmem/kmem_cache_free/enable 38 echo "call_site.function == $target_func" > events/kmem/kmem_cache_free/filter 64 echo "call_site.function == 0x$address" > events/kmem/kmem_cache_free/filter
|
| /linux/drivers/infiniband/hw/mana/ |
| H A D | cq.c | 147 if (cq->queue.kmem) in mana_ib_install_cq_cb() 148 gdma_cq = cq->queue.kmem; in mana_ib_install_cq_cb() 169 if (cq->queue.kmem) in mana_ib_remove_cq_cb() 180 struct gdma_queue *gdma_cq = cq->queue.kmem; in mana_ib_arm_cq() 192 struct gdma_queue *wq = qp->ud_qp.queues[MANA_UD_SEND_QUEUE].kmem; in handle_ud_sq_cqe() 208 struct gdma_queue *wq = qp->ud_qp.queues[MANA_UD_RECV_QUEUE].kmem; in handle_ud_rq_cqe() 326 struct gdma_queue *queue = cq->queue.kmem; in mana_ib_poll_cq()
|
| H A D | wr.c | 13 struct gdma_queue *queue = qp->ud_qp.queues[MANA_UD_RECV_QUEUE].kmem; in mana_ib_post_recv_ud() 80 struct gdma_queue *queue = qp->ud_qp.queues[MANA_UD_SEND_QUEUE].kmem; in mana_ib_post_send_ud()
|
| /linux/tools/perf/scripts/perl/bin/ |
| H A D | check-perf-trace-record | 2 perf record -a -e kmem:kmalloc -e irq:softirq_entry -e kmem:kfree
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_mr.c | 577 if (mtr->kmem) { in mtr_free_bufs() 578 hns_roce_buf_free(hr_dev, mtr->kmem); in mtr_free_bufs() 579 mtr->kmem = NULL; in mtr_free_bufs() 593 mtr->kmem = NULL; in mtr_alloc_bufs() 603 mtr->kmem = hns_roce_buf_alloc(hr_dev, total_size, in mtr_alloc_bufs() 607 if (IS_ERR(mtr->kmem)) { in mtr_alloc_bufs() 609 mtr->kmem); in mtr_alloc_bufs() 610 return PTR_ERR(mtr->kmem); in mtr_alloc_bufs() 662 mtr->kmem, page_shift); in mtr_map_bufs() 1049 mtr->kmem = NULL; in hns_roce_mtr_create()
|
| /linux/fs/resctrl/ |
| H A D | pseudo_lock.c | 197 kfree(plr->kmem); in pseudo_lock_region_clear() 198 plr->kmem = NULL; in pseudo_lock_region_clear() 311 plr->kmem = kzalloc(plr->size, GFP_KERNEL); in pseudo_lock_region_alloc() 312 if (!plr->kmem) { in pseudo_lock_region_alloc() 1035 physical = __pa(plr->kmem) >> PAGE_SHIFT; in pseudo_lock_dev_mmap_prepare() 1057 memset(plr->kmem + off, 0, vsize); in pseudo_lock_dev_mmap_prepare()
|
| /linux/Documentation/driver-api/cxl/linux/ |
| H A D | overview.rst | 82 * kmem - dax device is converted to hotplug memory blocks 84 * DAX kmem IO Resource creation 90 b) DAX kmem surfaces memory region to Memory Hotplug to add to page
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | pseudo_lock.c | 169 mem_r = plr->kmem; in resctrl_arch_pseudo_lock_fn() 255 mem_r = READ_ONCE(plr->kmem); in resctrl_arch_measure_cycles_lat_fn() 360 mem_r = READ_ONCE(plr->kmem); in measure_residency_fn()
|
| /linux/drivers/crypto/nx/ |
| H A D | nx.c | 285 memset(nx_ctx->kmem, 0, nx_ctx->kmem_len); in nx_ctx_init() 646 nx_ctx->kmem = kmalloc(nx_ctx->kmem_len, GFP_KERNEL); in nx_crypto_ctx_init() 647 if (!nx_ctx->kmem) in nx_crypto_ctx_init() 651 nx_ctx->csbcpb = (struct nx_csbcpb *)(round_up((u64)nx_ctx->kmem, in nx_crypto_ctx_init() 726 kfree_sensitive(nx_ctx->kmem); in nx_crypto_ctx_exit() 742 kfree_sensitive(nx_ctx->kmem); in nx_crypto_ctx_aead_exit()
|
| /linux/mm/kasan/ |
| H A D | kasan_test_c.c | 2151 char *kmem; in copy_user_test_oob() local 2157 kmem = kunit_kmalloc(test, size, GFP_KERNEL); in copy_user_test_oob() 2158 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, kmem); in copy_user_test_oob() 2172 unused = copy_from_user(kmem, usermem, size + 1)); in copy_user_test_oob() 2174 unused = copy_to_user(usermem, kmem, size + 1)); in copy_user_test_oob() 2176 unused = __copy_from_user(kmem, usermem, size + 1)); in copy_user_test_oob() 2178 unused = __copy_to_user(usermem, kmem, size + 1)); in copy_user_test_oob() 2180 unused = __copy_from_user_inatomic(kmem, usermem, size + 1)); in copy_user_test_oob() 2182 unused = __copy_to_user_inatomic(usermem, kmem, size + 1)); in copy_user_test_oob() 2188 memset(kmem, 'a', size); in copy_user_test_oob() [all …]
|
| /linux/drivers/dax/ |
| H A D | Makefile | 4 obj-$(CONFIG_DEV_DAX_KMEM) += kmem.o
|
| /linux/tools/testing/selftests/ftrace/test.d/dynevent/ |
| H A D | add_remove_tprobe.tc | 9 SUBSYSTEM=kmem
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| H A D | memory.rst | 100 memory.kmem.limit_in_bytes Deprecated knob to set and read the kernel 106 by memory.kmem.usage_in_bytes. 107 memory.kmem.usage_in_bytes show current kernel memory allocation 108 memory.kmem.failcnt show the number of kernel memory usage 110 memory.kmem.max_usage_in_bytes show max kernel memory usage recorded 112 memory.kmem.tcp.limit_in_bytes set/show hard limit for tcp buf memory 115 memory.kmem.tcp.usage_in_bytes show current tcp buf memory allocation 118 memory.kmem.tcp.failcnt show the number of tcp buf memory usage 122 memory.kmem.tcp.max_usage_in_bytes show max tcp buf memory usage recorded 337 memory.kmem.usage_in_bytes, or in a separate counter when it makes sense. [all …]
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.h | 129 struct irdma_dma_mem kmem; member 148 struct irdma_dma_mem kmem; member
|
| /linux/tools/perf/scripts/perl/ |
| H A D | check-perf-trace.pl | 43 sub kmem::kmalloc subroutine
|