Home
last modified time | relevance | path

Searched refs:nent (Results 1 – 25 of 57) sorted by relevance

123

/linux/tools/testing/selftests/kvm/lib/x86/
H A Dhyperv.c32 int i, nent = 0; in vcpu_set_hv_cpuid() local
38 cpuid_full = allocate_kvm_cpuid2(cpuid_sys->nent + cpuid_hv->nent); in vcpu_set_hv_cpuid()
45 for (i = 0; i < cpuid_sys->nent; i++) { in vcpu_set_hv_cpuid()
49 cpuid_full->entries[nent] = cpuid_sys->entries[i]; in vcpu_set_hv_cpuid()
50 nent++; in vcpu_set_hv_cpuid()
53 memcpy(&cpuid_full->entries[nent], cpuid_hv->entries, in vcpu_set_hv_cpuid()
54 cpuid_hv->nent * sizeof(struct kvm_cpuid_entry2)); in vcpu_set_hv_cpuid()
55 cpuid_full->nent = nent in vcpu_set_hv_cpuid()
[all...]
H A Dprocessor.c756 for (i = 0; i < cpuid->nent; i++) { in __kvm_cpu_has()
847 if (vcpu->cpuid && vcpu->cpuid->nent < cpuid->nent) { in vcpu_init_cpuid()
853 vcpu->cpuid = allocate_kvm_cpuid2(cpuid->nent); in vcpu_init_cpuid()
855 memcpy(vcpu->cpuid, cpuid, kvm_cpuid2_size(cpuid->nent)); in vcpu_init_cpuid()
1173 for (i = 0; i < cpuid->nent; i++) { in get_cpuid_entry()
/linux/tools/testing/selftests/kvm/x86/
H A Dcpuid_test.c32 for (i = 0; i < guest_cpuid->nent; i++) { in test_guest_cpuids()
89 TEST_ASSERT(cpuid1->nent == cpuid2->nent, in compare_cpuids()
90 "CPUID nent mismatch: %d vs. %d", cpuid1->nent, cpuid2->nent); in compare_cpuids()
92 for (i = 0; i < cpuid1->nent; i++) { in compare_cpuids()
145 int size = sizeof(*cpuid) + cpuid->nent * sizeof(cpuid->entries[0]); in vcpu_alloc_cpuid()
185 struct kvm_cpuid2 *cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent + 1); in test_get_cpuid2()
189 TEST_ASSERT(cpuid->nent in test_get_cpuid2()
[all...]
H A Dhyperv_cpuid.c38 TEST_ASSERT(hv_cpuid_entries->nent == nent_expected, in test_hv_cpuid()
41 nent_expected, hv_cpuid_entries->nent); in test_hv_cpuid()
43 for (i = 0; i < hv_cpuid_entries->nent; i++) { in test_hv_cpuid()
117 static struct kvm_cpuid2 cpuid = {.nent = 0}; in test_hv_cpuid_e2big()
H A Dhyperv_features.c154 prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent); in guest_test_msrs_access()
494 memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent)); in guest_test_msrs_access()
546 prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent); in guest_test_hcalls_access()
663 memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent)); in guest_test_hcalls_access()
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dgk20a.c41 int nent; in gk20a_gr_av_to_init_() local
44 nent = (blob->size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_init_()
45 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_init_()
52 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_init_()
84 int nent; in gk20a_gr_aiv_to_init() local
87 nent = (blob->size / sizeof(struct gk20a_fw_aiv)); in gk20a_gr_aiv_to_init()
88 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_aiv_to_init()
95 for (i = 0; i < nent; i++) { in gk20a_gr_aiv_to_init()
117 int nent; in gk20a_gr_av_to_method() local
120 nent in gk20a_gr_av_to_method()
[all...]
H A Dga102.c245 int nent; in ga102_gr_av64_to_init() local
248 nent = (blob->size / sizeof(struct netlist_av64)); in ga102_gr_av64_to_init()
249 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in ga102_gr_av64_to_init()
257 for (i = 0; i < nent; i++) { in ga102_gr_av64_to_init()
/linux/arch/x86/kvm/
H A Dcpuid.c85 struct kvm_cpuid_entry2 *entries, int nent, u32 function, u64 index) in kvm_find_cpuid_entry2() argument
102 for (i = 0; i < nent; i++) { in kvm_find_cpuid_entry2()
174 int nent) in kvm_cpuid_check_equal() argument
183 * Note! @e2 and @nent track the _old_ CPUID entries! in kvm_cpuid_check_equal()
188 if (nent != vcpu->arch.cpuid_nent) in kvm_cpuid_check_equal()
191 for (i = 0; i < nent; i++) { in kvm_cpuid_check_equal()
492 int nent) in kvm_set_cpuid() argument
509 swap(vcpu->arch.cpuid_nent, nent); in kvm_set_cpuid()
526 r = kvm_cpuid_check_equal(vcpu, e2, nent); in kvm_set_cpuid()
556 swap(vcpu->arch.cpuid_nent, nent); in kvm_set_cpuid()
1240 int nent; global() member
[all...]
/linux/drivers/tty/vt/
H A Dconmakehash.c82 int i, nuni, nent; in main() local
269 nent = 0; in main()
272 while ( nent >= unicount[fp0] ) in main()
275 nent = 0; in main()
277 printf("0x%04x", unitable[fp0][nent++]); in main()
/linux/drivers/hid/intel-thc-hid/intel-thc/
H A Dintel-thc-dma.c225 unsigned int i, nent = PRD_ENTRIES_NUM; in setup_dma_buffers() local
249 config->sgls[i] = sgl_alloc(buf_sz, GFP_KERNEL, &nent); in setup_dma_buffers()
250 if (!config->sgls[i] || nent > PRD_ENTRIES_NUM) { in setup_dma_buffers()
251 dev_err_once(dev->dev, "sgl_alloc (%uth) failed, nent %u\n", in setup_dma_buffers()
252 i, nent); in setup_dma_buffers()
255 count = dma_map_sg(dev->dev, config->sgls[i], nent, dir); in setup_dma_buffers()
448 size_t mes_len, u8 *nent) in calc_prd_entries_num() argument
450 *nent = DIV_ROUND_UP(mes_len, THC_MIN_BYTES_PER_SG_LIST_ENTRY); in calc_prd_entries_num()
451 if (*nent > PRD_ENTRIES_NUM) in calc_prd_entries_num()
457 static size_t calc_message_len(struct thc_prd_table *prd_tbl, u8 *nent) in calc_message_len() argument
569 u8 nent; read_dma_buffer() local
870 u8 nent; write_dma_buffer() local
[all...]
/linux/drivers/infiniband/hw/mthca/
H A Dmthca_allocator.c155 int mthca_array_init(struct mthca_array *array, int nent) in mthca_array_init() argument
157 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init()
173 void mthca_array_cleanup(struct mthca_array *array, int nent) in mthca_array_cleanup() argument
177 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i) in mthca_array_cleanup()
H A Dmthca_eq.c184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci()
230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe()
466 int nent, in mthca_create_eq() argument
479 eq->nent = roundup_pow_of_two(max(nent, 2)); in mthca_create_eq()
480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; in mthca_create_eq()
511 for (i = 0; i < eq->nent; ++i) in mthca_create_eq()
535 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); in mthca_create_eq()
560 eq->eqn, eq->nent); in mthca_create_eq()
593 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZ in mthca_free_eq()
[all...]
H A Dmthca_cq.c348 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) in mthca_alloc_cq_buf() argument
353 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, in mthca_alloc_cq_buf()
360 for (i = 0; i < nent; ++i) in mthca_alloc_cq_buf()
768 int mthca_init_cq(struct mthca_dev *dev, int nent, in mthca_init_cq() argument
776 cq->ibcq.cqe = nent - 1; in mthca_init_cq()
814 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); in mthca_init_cq()
828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
H A Dmthca_dev.h421 int mthca_array_init(struct mthca_array *array, int nent);
422 void mthca_array_cleanup(struct mthca_array *array, int nent);
487 int mthca_init_cq(struct mthca_dev *dev, int nent,
498 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
/linux/fs/nfs/
H A Dnfs42xattr.c73 atomic_long_t nent; member
297 atomic_long_set(&cache->nent, 0); in nfs4_xattr_alloc_cache()
406 atomic_long_set(&cache->nent, 0); in nfs4_xattr_discard_cache()
548 atomic_long_inc(&cache->nent); in nfs4_xattr_hash_add()
577 atomic_long_dec(&cache->nent); in nfs4_xattr_hash_remove()
812 if (atomic_long_read(&cache->nent) > 1) in cache_lru_isolate()
898 atomic_long_dec(&cache->nent); in entry_lru_isolate()
968 atomic_long_set(&cache->nent, 0); in nfs4_xattr_cache_init_once()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Deq.c259 u8 log_eq_size = order_base_2(param->nent + MLX5_NUM_SPARE_EQE); in create_map_eq()
664 .nent = MLX5_NUM_CMD_EQE, in create_async_eqs()
677 .nent = async_eq_depth_devlink_param_get(dev), in create_async_eqs()
693 .nent = /* TODO: sriov max_vf + */ 1, in create_async_eqs()
790 u32 nent = eq_get_size(eq); in mlx5_eq_get_eqe() local
793 eqe = get_eqe(eq, ci & (nent - 1)); in mlx5_eq_get_eqe()
794 eqe = ((eqe->owner & 1) ^ !!(ci & nent)) ? NULL : eqe; in mlx5_eq_get_eqe()
993 int nent; in create_comp_eq() local
1007 nent = comp_eq_depth_devlink_param_get(dev); in create_comp_eq()
1024 .nent in create_comp_eq()
[all...]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Deq.c109 /* (entry & (eq->nent - 1)) gives us a cyclic array */ in get_eqe()
110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe()
124 return !!(eqe->owner & 0x80) ^ !!(eq->cons_index & eq->nent) ? NULL : eqe; in next_eqe_sw()
781 mlx4_warn(dev, "Unhandled event FATAL WARNING (%02x), subtype %02x on EQ %d at index %u. owner=%x, nent=0x%x, slave=%x, ownership=%s\n", in mlx4_eq_int()
783 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
786 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
806 "Unhandled recoverable error event detected: %02x(%02x) on EQ %d at index %u. owner=%x, nent=0x%x, ownership=%s\n", in mlx4_eq_int()
808 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int()
810 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int()
818 mlx4_warn(dev, "Unhandled event %02x(%02x) on EQ %d at index %u. owner=%x, nent in mlx4_eq_int()
970 mlx4_create_eq(struct mlx4_dev * dev,int nent,u8 intr,struct mlx4_eq * eq) mlx4_create_eq() argument
[all...]
H A Dcq.c342 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, in mlx4_cq_alloc() argument
382 cpu_to_be32((ilog2(nent) << 24) | in mlx4_cq_alloc()
394 err = mlx4_init_user_cqes(buf_addr, nent, in mlx4_cq_alloc()
399 mlx4_init_kernel_cqes(buf_addr, nent, in mlx4_cq_alloc()
/linux/Documentation/translations/zh_CN/mm/
H A Dfree_page_reporting.rst29 nent-1中。 当页面被报告函数处理时,分配器将无法访问它们。一旦报告函数完成,这些
/linux/drivers/tty/serial/
H A Dpch_uart.c231 int nent; member
743 for (i = 0; i < priv->nent; i++, sg++) in pch_dma_tx_complete()
749 priv->nent = 0; in pch_dma_tx_complete()
855 int nent; in dma_handle_tx() local
931 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); in dma_handle_tx()
932 if (!nent) { in dma_handle_tx()
937 priv->nent = nent; in dma_handle_tx()
939 for (i = 0; i < nent; i++, sg++) { in dma_handle_tx()
943 if (i == (nent in dma_handle_tx()
[all...]
/linux/drivers/iommu/
H A Domap-iommu.c704 int nent = 1; in iopgtable_clear_entry_core() local
718 nent *= 16; in iopgtable_clear_entry_core()
722 bytes *= nent; in iopgtable_clear_entry_core()
723 memset(iopte, 0, nent * sizeof(*iopte)); in iopgtable_clear_entry_core()
725 flush_iopte_range(obj->dev, pt_dma, pt_offset, nent); in iopgtable_clear_entry_core()
736 nent = 1; /* for the next L1 entry */ in iopgtable_clear_entry_core()
740 nent *= 16; in iopgtable_clear_entry_core()
744 bytes *= nent; in iopgtable_clear_entry_core()
746 memset(iopgd, 0, nent * sizeof(*iopgd)); in iopgtable_clear_entry_core()
747 flush_iopte_range(obj->dev, obj->pd_dma, pd_offset, nent); in iopgtable_clear_entry_core()
[all...]
/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c77 static u8 sw_ownership_bit(int n, int nent) in sw_ownership_bit() argument
79 return (n & nent) ? 1 : 0; in sw_ownership_bit()
673 int nent, in alloc_cq_frag_buf() argument
682 nent * cqe_size, in alloc_cq_frag_buf()
691 buf->nent = nent; in alloc_cq_frag_buf()
880 for (i = 0; i < buf->nent; i++) { in init_cq_frag_buf()
1258 (i + 1) & cq->resize_buf->nent); in copy_resize_cqes()
1260 sw_own = sw_ownership_bit(i + 1, cq->resize_buf->nent); in copy_resize_cqes()
/linux/include/linux/mlx5/
H A Deq.h16 int nent; member
/linux/lib/
H A Dscatterlist.c595 unsigned int nent, nalloc; in sgl_alloc_order() local
598 nent = round_up(length, PAGE_SIZE << order) >> (PAGE_SHIFT + order); in sgl_alloc_order()
600 if (length > (nent << (PAGE_SHIFT + order))) in sgl_alloc_order()
602 nalloc = nent; in sgl_alloc_order()
630 *nent_p = nent; in sgl_alloc_order()
/linux/drivers/vfio/pci/mlx5/
H A Dcmd.c1054 struct mlx5_vhca_cq_buf *buf, int nent, in alloc_cq_frag_buf() argument
1062 err = mlx5_frag_buf_alloc_node(mdev, nent * cqe_size, frag_buf, in alloc_cq_frag_buf()
1069 buf->nent = nent; in alloc_cq_frag_buf()
1079 for (i = 0; i < buf->nent; i++) { in init_cq_frag_buf()
1617 u32 nent = size / entry_size; in set_report_output() local
1631 (nent > qp->max_msg_size / entry_size))) in set_report_output()
1641 nent_to_set = min(nent, nent_in_page); in set_report_output()
1651 nent -= nent_to_set; in set_report_output()
1652 } while (nent); in set_report_output()
[all...]

123