/linux/drivers/gpu/drm/xe/ |
H A D | xe_guc_log.c | 83 snapshot->num_chunks = DIV_ROUND_UP(snapshot->size, GUC_LOG_CHUNK_SIZE); in xe_guc_log_snapshot_alloc() 85 snapshot->copy = kcalloc(snapshot->num_chunks, sizeof(*snapshot->copy), in xe_guc_log_snapshot_alloc() 91 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_alloc() 103 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_alloc() 126 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_free() 160 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_capture() 209 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_print() 212 char suffix = i == snapshot->num_chunks - 1 ? '\n' : 0; in xe_guc_log_snapshot_print()
|
H A D | xe_hmm.c | 26 unsigned long num_chunks = 0; in xe_alloc_sg() local 51 num_chunks++; in xe_alloc_sg() 55 return sg_alloc_table(st, num_chunks, GFP_KERNEL); in xe_alloc_sg()
|
H A D | xe_guc_log_types.h | 25 /** @num_chunks: Number of chunks within @copy */ 26 int num_chunks; member
|
/linux/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_virtchnl.c | 1197 u16 num_chunks = le16_to_cpu(chunks->num_chunks); in idpf_vport_get_q_reg() local 1201 while (num_chunks--) { in idpf_vport_get_q_reg() 1205 chunk = &chunks->chunks[num_chunks]; in idpf_vport_get_q_reg() 1576 int totqs, num_msgs, num_chunks; in idpf_send_config_tx_queues_msg() local 1648 num_chunks = min_t(u32, IDPF_NUM_CHUNKS_PER_MSG(config_sz, chunk_sz), in idpf_send_config_tx_queues_msg() 1650 num_msgs = DIV_ROUND_UP(totqs, num_chunks); in idpf_send_config_tx_queues_msg() 1652 buf_sz = struct_size(ctq, qinfo, num_chunks); in idpf_send_config_tx_queues_msg() 1663 ctq->num_qinfo = cpu_to_le16(num_chunks); in idpf_send_config_tx_queues_msg() 1664 memcpy(ctq->qinfo, &qi[k], chunk_sz * num_chunks); in idpf_send_config_tx_queues_msg() 1695 int totqs, num_msgs, num_chunks; idpf_send_config_rx_queues_msg() local 1847 u32 num_msgs, num_chunks, num_txq, num_rxq, num_q; idpf_send_ena_dis_queues_msg() local 1995 u32 num_msgs, num_chunks, num_q; idpf_send_map_unmap_queue_vector_msg() local 2158 idpf_convert_reg_to_queue_chunks(struct virtchnl2_queue_chunk * dchunks,struct virtchnl2_queue_reg_chunk * schunks,u16 num_chunks) idpf_convert_reg_to_queue_chunks() argument 2185 u16 num_chunks; idpf_send_delete_queues_msg() local 3389 u16 num_chunks = le16_to_cpu(chunks->num_vchunks); idpf_get_vec_ids() local 3431 u16 num_chunks = le16_to_cpu(chunks->num_chunks); idpf_vport_get_queue_ids() local [all...] |
H A D | virtchnl2.h | 592 * @num_chunks: Number of chunks. 597 __le16 num_chunks; member 599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks); 661 * num_chunks entries of virtchnl2_queue_chunk structures. 873 * followed by num_chunks of virtchnl2_queue_chunk structures. 1197 * @num_chunks: Number of chunks. 1202 __le16 num_chunks; member 1204 struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks);
|
/linux/drivers/platform/x86/intel/ifs/ |
H A D | load.c | 123 int i, num_chunks, chunk_size; in copy_hashes_authenticate_chunks() local 135 num_chunks = hashes_status.num_chunks; in copy_hashes_authenticate_chunks() 149 for (i = 0; i < num_chunks; i++) { in copy_hashes_authenticate_chunks() 171 return gen >= IFS_GEN_STRIDE_AWARE ? status.chunks_in_stride : status.num_chunks; in get_num_chunks() 187 int i, num_chunks, chunk_size; in copy_hashes_authenticate_chunks_gen2() local 206 num_chunks = get_num_chunks(ifsd->generation, hashes_status); in copy_hashes_authenticate_chunks_gen2() 215 num_chunks = ifsd->valid_chunks; in copy_hashes_authenticate_chunks_gen2() 234 for (i = 0; i < num_chunks; i++) { in copy_hashes_authenticate_chunks_gen2()
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | icp_qat_uclo.h | 260 unsigned short num_chunks; member 275 short num_chunks; member 520 unsigned short num_chunks; member 546 unsigned short num_chunks; member 562 unsigned short num_chunks; member
|
H A D | qat_uclo.c | 117 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format() 458 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk() 505 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk() 1208 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof() 1899 uobj_chunk_num = uobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1901 sobj_chunk_num = sobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1981 if (mof_hdr->num_chunks <= 0x1) { in qat_uclo_check_mof_format() 2024 chunks_num = mof_ptr->num_chunks; in qat_uclo_map_mof_obj()
|
/linux/drivers/accel/habanalabs/common/ |
H A D | command_submission.c | 1351 u32 cs_type_flags, num_chunks; in hl_cs_sanity_checks() local 1382 num_chunks = args->in.num_chunks_execute; in hl_cs_sanity_checks() 1393 if (!num_chunks) { in hl_cs_sanity_checks() 1397 } else if (is_sync_stream && num_chunks != 1) { in hl_cs_sanity_checks() 1409 void __user *chunks, u32 num_chunks, in hl_cs_copy_chunk_array() argument 1414 if (num_chunks > HL_MAX_JOBS_PER_CS) { in hl_cs_copy_chunk_array() 1423 *cs_chunk_array = kmalloc_array(num_chunks, sizeof(**cs_chunk_array), in hl_cs_copy_chunk_array() 1426 *cs_chunk_array = kmalloc_array(num_chunks, in hl_cs_copy_chunk_array() 1434 size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); in hl_cs_copy_chunk_array() 1490 u32 num_chunks, u6 in cs_ioctl_default() argument 1695 u32 num_chunks, tmp; hl_cs_ctx_switch() local 2199 cs_ioctl_signal_wait(struct hl_fpriv * hpriv,enum hl_cs_type cs_type,void __user * chunks,u32 num_chunks,u64 * cs_seq,u32 flags,u32 timeout,u32 * signal_sob_addr_offset,u16 * signal_initial_sob_count) cs_ioctl_signal_wait() argument 2564 u32 num_chunks, flags, timeout, hl_cs_ioctl() local [all...] |
/linux/include/drm/ |
H A D | drm_gpusvm.h | 175 * @num_chunks: Number of chunks 199 int num_chunks; member 242 const unsigned long *chunk_sizes, int num_chunks);
|
/linux/drivers/infiniband/sw/siw/ |
H A D | siw_mem.c | 341 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 348 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get() 355 kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); in siw_umem_get()
|
/linux/drivers/gpu/drm/lima/ |
H A D | lima_dump.h | 47 __u32 num_chunks; member
|
H A D | lima_sched.c | 336 dt->num_chunks++; in lima_sched_build_error_task_list() 343 dt->num_chunks++; in lima_sched_build_error_task_list() 349 dt->num_chunks++; in lima_sched_build_error_task_list() 388 dt->num_chunks++; in lima_sched_build_error_task_list()
|
/linux/drivers/gpu/drm/ |
H A D | drm_gpusvm.c | 429 * @num_chunks: Number of chunks. 441 const unsigned long *chunk_sizes, int num_chunks) in drm_gpusvm_init() argument 443 if (!ops->invalidate || !num_chunks) in drm_gpusvm_init() 455 gpusvm->num_chunks = num_chunks; in drm_gpusvm_init() 806 for (; i < gpusvm->num_chunks; ++i) { in drm_gpusvm_range_chunk_size() 817 if (i == gpusvm->num_chunks) in drm_gpusvm_range_chunk_size()
|
/linux/drivers/crypto/marvell/octeontx/ |
H A D | otx_cptvf.h | 37 u32 num_chunks; /* Number of command chunks */ member
|
H A D | otx_cptvf_main.c | 179 queue->num_chunks = 0; in free_command_queues() 226 i, queue->num_chunks); in alloc_command_queues() 231 if (queue->num_chunks == 0) { in alloc_command_queues() 238 queue->num_chunks++; in alloc_command_queues()
|
/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_cs.c | 277 if (!cs->num_chunks) { in radeon_cs_parser_init() 289 p->chunks_array = kvmalloc_array(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); in radeon_cs_parser_init() 295 sizeof(uint64_t)*cs->num_chunks)) { in radeon_cs_parser_init() 299 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
|
/linux/drivers/net/dsa/sja1105/ |
H A D | sja1105_spi.c | 44 int num_chunks; in sja1105_xfer() local 47 num_chunks = DIV_ROUND_UP(len, priv->max_xfer_len); in sja1105_xfer() 56 for (i = 0; i < num_chunks; i++) { in sja1105_xfer()
|
/linux/drivers/vfio/pci/mlx5/ |
H A D | main.c | 385 int num_chunks; in mlx5vf_prep_stop_copy() local 408 num_chunks = mvdev->chunk_mode ? MAX_NUM_CHUNKS : 1; in mlx5vf_prep_stop_copy() 409 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy() 444 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
|
/linux/drivers/staging/media/atomisp/pci/ |
H A D | ia_css_acc_types.h | 111 u32 num_chunks; member
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_cs.c | 51 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init() 188 chunk_array = kvmalloc_array(cs->in.num_chunks, sizeof(uint64_t), in amdgpu_cs_pass1() 196 sizeof(uint64_t)*cs->in.num_chunks)) { in amdgpu_cs_pass1() 201 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1()
|
/linux/drivers/media/i2c/ |
H A D | thp7312.c | 1669 unsigned int num_chunks; in thp7312_fw_load_to_ram() local 1677 num_chunks = DIV_ROUND_UP(bank_size, THP7312_FW_DOWNLOAD_UNIT); in thp7312_fw_load_to_ram() 1680 __func__, bank_size, i, num_chunks); in thp7312_fw_load_to_ram() 1682 for (j = 0 ; j < num_chunks; j++) { in thp7312_fw_load_to_ram()
|
/linux/net/sctp/ |
H A D | socket.c | 6973 u32 num_chunks = 0; in sctp_getsockopt_peer_auth_chunks() local 6995 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_peer_auth_chunks() 6996 if (len < num_chunks) in sctp_getsockopt_peer_auth_chunks() 6999 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_peer_auth_chunks() 7002 len = sizeof(struct sctp_authchunks) + num_chunks; in sctp_getsockopt_peer_auth_chunks() 7005 if (put_user(num_chunks, &p->gauth_number_of_chunks)) in sctp_getsockopt_peer_auth_chunks() 7018 u32 num_chunks = 0; in sctp_getsockopt_local_auth_chunks() local 7045 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_local_auth_chunks() 7046 if (len < sizeof(struct sctp_authchunks) + num_chunks) in sctp_getsockopt_local_auth_chunks() 7049 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_local_auth_chunks() [all...] |
/linux/include/uapi/drm/ |
H A D | radeon_drm.h | 984 __u32 num_chunks; member
|
H A D | amdgpu_drm.h | 917 __u32 num_chunks; member
|