| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | fs_pool.c | 12 fs_bulk->bitmask = kvcalloc(BITS_TO_LONGS(fs_bulk->bulk_len), in mlx5_fs_bulk_bitmap_alloc() 17 for (i = 0; i < fs_bulk->bulk_len; i++) in mlx5_fs_bulk_bitmap_alloc() 23 void mlx5_fs_bulk_init(struct mlx5_fs_bulk *fs_bulk, int bulk_len) in mlx5_fs_bulk_init() argument 25 fs_bulk->bulk_len = bulk_len; in mlx5_fs_bulk_init() 35 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount() 41 int free_index = find_first_bit(fs_bulk->bitmask, fs_bulk->bulk_len); in mlx5_fs_bulk_acquire_index() 44 if (free_index >= fs_bulk->bulk_len) in mlx5_fs_bulk_acquire_index() 101 fs_pool->available_units += new_bulk->bulk_len; in mlx5_fs_pool_alloc_new_bulk() 111 fs_pool->available_units -= bulk->bulk_len; in mlx5_fs_pool_free_bulk() 188 if (bulk_free_amount == bulk->bulk_len) { in mlx5_fs_pool_release_index()
|
| H A D | fs_counters.c | 114 u32 bulk_len = fc_stats->bulk_query_len; in mlx5_fc_stats_query_all_counters() local 131 last_bulk_id = bulk_base_id + bulk_len; in mlx5_fc_stats_query_all_counters() 137 err = mlx5_cmd_fc_bulk_query(dev, bulk_base_id, bulk_len, data); in mlx5_fc_stats_query_all_counters() 456 int bulk_len; in mlx5_fc_bulk_create() local 461 bulk_len = alloc_bitmask > 0 ? MLX5_FC_BULK_NUM_FCS(alloc_bitmask) : 1; in mlx5_fc_bulk_create() 463 fc_bulk = kvzalloc_flex(*fc_bulk, fcs, bulk_len); in mlx5_fc_bulk_create() 467 mlx5_fs_bulk_init(&fc_bulk->fs_bulk, bulk_len); in mlx5_fc_bulk_create() 476 for (i = 0; i < bulk_len; i++) in mlx5_fc_bulk_create() 495 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fc_bulk_destroy()
|
| H A D | fs_pool.h | 11 int bulk_len; member 42 void mlx5_fs_bulk_init(struct mlx5_fs_bulk *fs_bulk, int bulk_len);
|
| H A D | fs_cmd.h | 118 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len); 119 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len,
|
| H A D | fs_cmd.c | 882 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len) in mlx5_cmd_fc_get_bulk_query_out_len() argument 885 MLX5_ST_SZ_BYTES(traffic_counter) * bulk_len; in mlx5_cmd_fc_get_bulk_query_out_len() 888 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len, in mlx5_cmd_fc_bulk_query() argument 891 int outlen = mlx5_cmd_fc_get_bulk_query_out_len(bulk_len); in mlx5_cmd_fc_bulk_query() 897 MLX5_SET(query_flow_counter_in, in, num_of_counters, bulk_len); in mlx5_cmd_fc_bulk_query()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | fs_hws_pools.c | 113 int bulk_len; in mlx5_fs_hws_pr_bulk_create() local 119 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_pr_bulk_create() 120 pr_bulk = kvzalloc_flex(*pr_bulk, prs_data, bulk_len); in mlx5_fs_hws_pr_bulk_create() 124 mlx5_fs_bulk_init(&pr_bulk->fs_bulk, bulk_len); in mlx5_fs_hws_pr_bulk_create() 129 for (i = 0; i < bulk_len; i++) { in mlx5_fs_hws_pr_bulk_create() 153 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fs_hws_pr_bulk_destroy() 261 int bulk_len; in mlx5_fs_hws_mh_bulk_create() local 275 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_mh_bulk_create() 276 mh_bulk = kvzalloc_flex(*mh_bulk, mhs_data, bulk_len); in mlx5_fs_hws_mh_bulk_create() 280 mlx5_fs_bulk_init(&mh_bulk->fs_bulk, bulk_len); in mlx5_fs_hws_mh_bulk_create() [all …]
|
| /linux/lib/crypto/ |
| H A D | poly1305.c | 41 unsigned int bulk_len; in poly1305_update() local 55 bulk_len = round_down(nbytes, POLY1305_BLOCK_SIZE); in poly1305_update() 58 if (bulk_len) { in poly1305_update() 59 poly1305_blocks(&desc->state, src, bulk_len, 1); in poly1305_update() 60 src += bulk_len; in poly1305_update()
|
| /linux/crypto/ |
| H A D | adiantum.c | 210 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_header() local 215 .message_bits = cpu_to_le64((u64)bulk_len * 8) in adiantum_hash_header() 373 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_crypt() local 407 nhpoly1305_update(&rctx->u.hash_ctx, tctx, virt, bulk_len); in adiantum_crypt() 409 memcpy(&rbuf.bignum, virt + bulk_len, sizeof(le128)); in adiantum_crypt() 414 memcpy_from_sglist(&rbuf.bignum, src, bulk_len, sizeof(le128)); in adiantum_crypt() 441 stream_len = bulk_len; in adiantum_crypt() 472 nhpoly1305_update(&rctx->u.hash_ctx, tctx, virt, bulk_len); in adiantum_crypt() 475 memcpy(virt + bulk_len, &rbuf.bignum, sizeof(le128)); in adiantum_crypt() 482 memcpy_to_sglist(dst, bulk_len, &rbuf.bignum, sizeof(le128)); in adiantum_crypt()
|
| H A D | hctr2.c | 159 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message() local 166 for (i = 0; i < bulk_len; i += n) { in hctr2_hash_message() 168 n = min_t(unsigned int, miter.length, bulk_len - i); in hctr2_hash_message() 215 int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_crypt() local 254 rctx->bulk_part_dst, bulk_len, in hctr2_crypt()
|
| /linux/net/core/ |
| H A D | page_pool.c | 925 u32 bulk_len) in page_pool_recycle_ring_bulk() argument 933 for (i = 0; i < bulk_len; i++) { in page_pool_recycle_ring_bulk() 945 if (likely(i == bulk_len)) in page_pool_recycle_ring_bulk() 952 for (; i < bulk_len; i++) in page_pool_recycle_ring_bulk() 972 u32 bulk_len = 0; in page_pool_put_netmem_bulk() local 978 data[bulk_len++] = netmem; in page_pool_put_netmem_bulk() 981 count = bulk_len; in page_pool_put_netmem_bulk() 988 bulk_len = 0; in page_pool_put_netmem_bulk() 1011 bulk[bulk_len++] = netmem; in page_pool_put_netmem_bulk() 1014 if (bulk_len) in page_pool_put_netmem_bulk() [all …]
|
| /linux/drivers/gpu/drm/gud/ |
| H A D | gud_drv.c | 318 string_get_size(gdrm->bulk_len, 1, STRING_UNITS_2, buf, sizeof(buf)); in gud_stats_debugfs() 405 gdrm->bulk_buf = vmalloc_32(gdrm->bulk_len); in gud_alloc_bulk_buffer() 409 num_pages = DIV_ROUND_UP(gdrm->bulk_len, PAGE_SIZE); in gud_alloc_bulk_buffer() 418 0, gdrm->bulk_len, GFP_KERNEL); in gud_alloc_bulk_buffer() 594 gdrm->bulk_len = max_buffer_size; in gud_probe() 605 gdrm->compress_buf = vmalloc(gdrm->bulk_len); in gud_probe()
|
| H A D | gud_pipe.c | 165 if (len > gdrm->bulk_len) in gud_prep_flush() 342 if (gdrm->bulk_len < lines * pitch) in gud_flush_damage() 343 lines = gdrm->bulk_len / pitch; in gud_flush_damage()
|
| H A D | gud_internal.h | 28 size_t bulk_len; member
|