/linux/mm/ |
H A D | page_frag_cache.c | 79 nc->pagecnt_bias); in page_frag_cache_drain() 115 nc->pagecnt_bias = PAGE_FRAG_CACHE_MAX_SIZE + 1; in __page_frag_alloc_align() 137 if (!page_ref_sub_and_test(page, nc->pagecnt_bias)) in __page_frag_alloc_align() 150 nc->pagecnt_bias = PAGE_FRAG_CACHE_MAX_SIZE + 1; in __page_frag_alloc_align() 154 nc->pagecnt_bias--; in __page_frag_alloc_align()
|
/linux/include/linux/ |
H A D | mm_types_task.h | 60 __u16 pagecnt_bias; member 63 __u32 pagecnt_bias; member
|
/linux/drivers/net/ethernet/google/gve/ |
H A D | gve_utils.c | 94 page_info->pagecnt_bias--; in gve_dec_pagecnt_bias() 95 if (page_info->pagecnt_bias == 0) { in gve_dec_pagecnt_bias() 101 page_info->pagecnt_bias = INT_MAX - pagecount; in gve_dec_pagecnt_bias()
|
H A D | gve_rx.c | 22 page_ref_sub(page_info->page, page_info->pagecnt_bias - 1); in gve_rx_free_buffer() 43 rx->data.page_info[i].pagecnt_bias - 1); in gve_rx_unfill_pages() 47 rx->qpl_copy_pool[i].pagecnt_bias - 1); in gve_rx_unfill_pages() 156 page_info->pagecnt_bias = INT_MAX; in gve_setup_rx_buffer() 234 rx->qpl_copy_pool[j].pagecnt_bias = INT_MAX; in gve_rx_prefill_pages() 244 rx->qpl_copy_pool[j].pagecnt_bias - 1); in gve_rx_prefill_pages() 253 rx->data.page_info[i].pagecnt_bias - 1); in gve_rx_prefill_pages() 507 if (pagecount == page_info->pagecnt_bias) in gve_rx_can_recycle_buffer() 510 else if (pagecount > page_info->pagecnt_bias) in gve_rx_can_recycle_buffer() 512 WARN(pagecount < page_info->pagecnt_bias, in gve_rx_can_recycle_buffer() [all...] |
H A D | gve_buffer_mgmt_dqo.c | 13 return page_count(bs->page_info.page) - bs->page_info.pagecnt_bias; in gve_buf_ref_cnt() 153 buf_state->page_info.pagecnt_bias = INT_MAX; in gve_alloc_qpl_page_dqo() 164 buf_state->page_info.pagecnt_bias - 1); in gve_free_qpl_page_dqo()
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_txrx.c | 421 __page_frag_cache_drain(rx_buf->page, rx_buf->pagecnt_bias); in ice_clean_rx_ring() 703 bi->pagecnt_bias = USHRT_MAX; in ice_alloc_mapped_page() 837 unsigned int pagecnt_bias = rx_buf->pagecnt_bias; in ice_can_reuse_rx_page() local 845 if (unlikely(rx_buf->pgcnt - pagecnt_bias > 1)) in ice_can_reuse_rx_page() 855 * the pagecnt_bias and page count so that we fully restock the in ice_can_reuse_rx_page() 858 if (unlikely(pagecnt_bias == 1)) { in ice_can_reuse_rx_page() 860 rx_buf->pagecnt_bias = USHRT_MAX; in ice_can_reuse_rx_page() 934 new_buf->pagecnt_bias = old_buf->pagecnt_bias; in ice_reuse_rx_page() [all...] |
H A D | ice_txrx.h | 205 unsigned int pagecnt_bias; member
|
/linux/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.c | 1381 new_buff->pagecnt_bias = old_buff->pagecnt_bias; in i40e_reuse_rx_page() 1503 __page_frag_cache_drain(rx_bi->page, rx_bi->pagecnt_bias); in i40e_clean_rx_ring() 1668 bi->pagecnt_bias = USHRT_MAX; in i40e_alloc_mapped_page() 1925 unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; in i40e_can_reuse_rx_page() local 1936 if (unlikely((rx_buffer->page_count - pagecnt_bias) > 1)) { in i40e_can_reuse_rx_page() 1950 * the pagecnt_bias and page count so that we fully restock the in i40e_can_reuse_rx_page() 1953 if (unlikely(pagecnt_bias == 1)) { in i40e_can_reuse_rx_page() 1955 rx_buffer->pagecnt_bias in i40e_can_reuse_rx_page() [all...] |
H A D | i40e_txrx.h | 282 __u16 pagecnt_bias; member
|
/linux/drivers/net/ethernet/meta/fbnic/ |
H A D | fbnic_txrx.h | 98 long pagecnt_bias; member
|
H A D | fbnic_txrx.c | 665 rx_buf->pagecnt_bias = FBNIC_PAGECNT_BIAS_MAX; in fbnic_page_pool_init() 674 rx_buf->pagecnt_bias--; in fbnic_page_pool_get() 685 if (!page_pool_unref_page(page, rx_buf->pagecnt_bias)) in fbnic_page_pool_drain()
|
/linux/drivers/net/ethernet/intel/ixgbevf/ |
H A D | ixgbevf_main.c | 547 rx_buffer->pagecnt_bias--; in ixgbevf_get_rx_buffer() 569 rx_buffer->pagecnt_bias); in ixgbevf_put_rx_buffer() 643 bi->pagecnt_bias = 1; in ixgbevf_alloc_mapped_page() 781 new_buff->pagecnt_bias = old_buff->pagecnt_bias; in ixgbevf_reuse_rx_page() 786 unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; in ixgbevf_can_reuse_rx_page() local 795 if (unlikely((page_ref_count(page) - pagecnt_bias) > 1)) in ixgbevf_can_reuse_rx_page() 807 * the pagecnt_bias and page count so that we fully restock the in ixgbevf_can_reuse_rx_page() 810 if (unlikely(!pagecnt_bias)) { in ixgbevf_can_reuse_rx_page() [all...] |
/linux/drivers/net/ethernet/intel/igc/ |
H A D | igc_main.c | 435 buffer_info->pagecnt_bias); in igc_clean_rx_ring_page_shared() 1904 rx_buffer->pagecnt_bias--; in igc_get_rx_buffer() 2039 rx_buffer->pagecnt_bias++; in igc_construct_skb() 2071 new_buff->pagecnt_bias = old_buff->pagecnt_bias; in igc_reuse_rx_page() 2077 unsigned int pagecnt_bias = rx_buffer->pagecnt_bias; in igc_can_reuse_rx_page() local 2086 if (unlikely((rx_buffer_pgcnt - pagecnt_bias) > 1)) in igc_can_reuse_rx_page() 2097 * the pagecnt_bias and page count so that we fully restock the in igc_can_reuse_rx_page() 2100 if (unlikely(pagecnt_bias in igc_can_reuse_rx_page() [all...] |
H A D | igc.h | 587 __u16 pagecnt_bias; member
|
/linux/drivers/net/ethernet/intel/igb/ |
H A D | igb.h | 292 __u16 pagecnt_bias; member
|
/linux/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe.h | 281 __u16 pagecnt_bias; member
|