Lines Matching defs:bcnt

696 			     u64 user_va, size_t bcnt, u32 *bytes_mapped,
718 np = ib_umem_odp_map_dma_and_lock(odp, user_va, bcnt, access_mask, fault);
740 *bytes_mapped += min_t(u32, new_mappings, bcnt);
751 size_t bcnt, u32 *bytes_mapped, u32 flags)
753 unsigned long end_idx = (user_va + bcnt - 1) >> MLX5_IMR_MTT_SHIFT;
761 mlx5_imr_ksm_entries * MLX5_IMR_MTT_SIZE - user_va < bcnt))
765 while (bcnt) {
788 len = min_t(u64, user_va + bcnt, ib_umem_end(umem_odp)) -
799 bcnt -= len;
835 static int pagefault_dmabuf_mr(struct mlx5_ib_mr *mr, size_t bcnt,
886 *bytes_mapped += bcnt;
893 * -EFAULT: The io_virt->bcnt is not within the MR, it covers pages that are
900 static int pagefault_mr(struct mlx5_ib_mr *mr, u64 io_virt, size_t bcnt,
909 return pagefault_dmabuf_mr(mr, bcnt, bytes_mapped, flags);
922 if ((user_va + bcnt) > ib_umem_end(odp))
923 bcnt = ib_umem_end(odp) - user_va;
925 ib_umem_end(odp) - user_va < bcnt))
927 return pagefault_real_mr(mr, odp, user_va, bcnt, bytes_mapped,
930 return pagefault_implicit_mr(mr, odp, io_virt, bcnt, bytes_mapped,
958 size_t bcnt;
1005 u64 io_virt, size_t bcnt,
1018 bcnt -= *bytes_committed;
1029 *bytes_mapped += bcnt;
1044 pages_in_range = (ALIGN(io_virt + bcnt, PAGE_SIZE) -
1047 ret = pagefault_mr(mr, io_virt, bcnt, bytes_mapped, 0, false);
1092 for (i = 0; bcnt && i < mmkey->ndescs; i++, pklm++) {
1106 frame->bcnt = min_t(size_t, bcnt,
1112 bcnt -= frame->bcnt;
1129 bcnt = frame->bcnt;
1179 size_t bcnt;
1194 bcnt = byte_count & ~MLX5_INLINE_SEG;
1197 bcnt = bcnt & MLX5_WQE_INLINE_SEG_BYTE_COUNT_MASK;
1198 wqe += ALIGN(sizeof(struct mlx5_wqe_inline_seg) + bcnt,
1205 if (receive_queue && bcnt == 0 &&
1211 *total_wqe_bytes += bcnt - min_t(size_t, bcnt,
1216 if (bcnt == 0)
1217 bcnt = 1U << 31;
1219 if (inline_segment || bcnt <= pfault->bytes_committed) {
1221 min_t(size_t, bcnt,
1227 io_virt, bcnt,