/linux/include/linux/ |
H A D | agpgart.h | 53 size_t pg_count; /* number of pages */ member 59 size_t pg_count; member 71 size_t pg_count; /* number of pages */ member
|
H A D | nfs_page.h | 91 size_t pg_count; member
|
/linux/arch/alpha/kernel/ |
H A D | pci_iommu.c | 826 iommu_reserve(struct pci_iommu_arena *arena, long pg_count, long align_mask) in iommu_reserve() 838 p = iommu_arena_find_pages(NULL, arena, pg_count, align_mask); in iommu_reserve() 847 for (i = 0; i < pg_count; ++i) in iommu_reserve() 850 arena->next_entry = p + pg_count; in iommu_reserve() 857 iommu_release(struct pci_iommu_arena *arena, long pg_start, long pg_count) in iommu_release() 867 for(i = pg_start; i < pg_start + pg_count; i++) in iommu_release() 871 iommu_arena_free(arena, pg_start, pg_count); in iommu_release() 876 iommu_bind(struct pci_iommu_arena *arena, long pg_start, long pg_count, in iommu_bind() 889 for(j = pg_start; j < pg_start + pg_count; j++) { in iommu_bind() 896 for(i = 0, j = pg_start; i < pg_count; in iommu_bind() 824 iommu_reserve(struct pci_iommu_arena * arena,long pg_count,long align_mask) iommu_reserve() argument 855 iommu_release(struct pci_iommu_arena * arena,long pg_start,long pg_count) iommu_release() argument 874 iommu_bind(struct pci_iommu_arena * arena,long pg_start,long pg_count,struct page ** pages) iommu_bind() argument 903 iommu_unbind(struct pci_iommu_arena * arena,long pg_start,long pg_count) iommu_unbind() argument [all...] |
H A D | core_titan.c | 591 long pg_count; in titan_agp_setup() 607 aper->pg_count = alpha_agpgart_size / PAGE_SIZE; in titan_agp_setup() 608 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in titan_agp_setup() 609 aper->pg_count - 1); in titan_agp_setup() 618 agp->aperture.size = aper->pg_count * PAGE_SIZE; 630 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup() 634 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in titan_agp_cleanup() 636 aper->pg_count); in titan_agp_cleanup() 586 long pg_count; global() member
|
H A D | core_marvel.c | 862 long pg_count; in marvel_agp_setup() 877 aper->pg_count = alpha_agpgart_size / PAGE_SIZE; in marvel_agp_setup() 878 aper->pg_start = iommu_reserve(aper->arena, aper->pg_count, in marvel_agp_setup() 879 aper->pg_count - 1); in marvel_agp_setup() 889 agp->aperture.size = aper->pg_count * PAGE_SIZE; in marvel_agp_cleanup() 901 status = iommu_release(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup() 905 iommu_unbind(aper->arena, aper->pg_start, aper->pg_count); in marvel_agp_cleanup() 907 aper->pg_count); 853 long pg_count; global() member
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_res.c | 70 for (i = 0; i < pbl->pg_count; i++) { in __free_pbl() 87 pbl->pg_count = 0; in __free_pbl() 100 pbl->pg_count++; in bnxt_qplib_fill_user_dma_pages() 131 pbl->pg_count = 0; in __alloc_pbl() 142 pbl->pg_count++; in __alloc_pbl() 260 for (i = 0; i < hwq->pbl[PBL_LVL_1].pg_count; i++) in bnxt_qplib_alloc_init_hwq() 275 for (i = 0; i < hwq->pbl[PBL_LVL_2].pg_count; i++) { in bnxt_qplib_alloc_init_hwq() 281 i = hwq->pbl[PBL_LVL_2].pg_count; in bnxt_qplib_alloc_init_hwq() 315 for (i = 0; i < hwq->pbl[PBL_LVL_1].pg_count; i++) in bnxt_qplib_alloc_init_hwq() 320 i = hwq->pbl[PBL_LVL_1].pg_count; in bnxt_qplib_alloc_init_hwq() 416 int pg_count; bnxt_qplib_map_tqm_pgtbl() local [all...] |
H A D | qplib_res.h | 157 u32 pg_count; member
|
H A D | main.c | 1776 rattr.pages = nq->hwq.pbl[rdev->nqr->nq[i].hwq.level].pg_count; in bnxt_re_alloc_res() 2144 rattr.pages = creq->hwq.pbl[creq->hwq.level].pg_count; in bnxt_re_dev_init()
|
/linux/include/uapi/linux/ |
H A D | agpgart.h | 81 __kernel_size_t pg_count; /* number of pages */ member 93 __kernel_size_t pg_count;/* number of pages */ member
|
/linux/drivers/char/agp/ |
H A D | intel-gtt.c | 241 static struct agp_memory *alloc_agpphysmem_i8xx(size_t pg_count, int type) in alloc_agpphysmem_i8xx() argument 246 switch (pg_count) { in alloc_agpphysmem_i8xx() 260 new = agp_create_memory(pg_count); in alloc_agpphysmem_i8xx() 265 if (pg_count == 4) { in alloc_agpphysmem_i8xx() 271 new->page_count = pg_count; in alloc_agpphysmem_i8xx() 272 new->num_scratch_pages = pg_count; in alloc_agpphysmem_i8xx() 1013 static struct agp_memory *intel_fake_agp_alloc_by_type(size_t pg_count, in intel_fake_agp_alloc_by_type() argument 1019 if (pg_count != intel_private.num_dcache_entries) in intel_fake_agp_alloc_by_type() 1027 new->page_count = pg_count; in intel_fake_agp_alloc_by_type() 1033 return alloc_agpphysmem_i8xx(pg_count, typ in intel_fake_agp_alloc_by_type() [all...] |
/linux/drivers/hv/ |
H A D | mshv_root_hv_call.c | 22 #define HV_PAGE_COUNT_2M_ALIGNED(pg_count) (!((pg_count) & (0x200 - 1))) argument
|
H A D | hv_balloon.c | 836 unsigned long pg_count) in handle_pg_range() argument 839 unsigned long pfn_cnt = pg_count; in handle_pg_range() 846 pr_debug("Hot adding %lu pages starting at pfn 0x%lx.\n", pg_count, in handle_pg_range()
|
/linux/drivers/md/ |
H A D | dm-mpath.c | 1194 unsigned int pg_count = 0; in multipath_ctr() local 1248 pg_count++; in multipath_ctr() 1249 pg->pg_num = pg_count; in multipath_ctr() 1254 if (pg_count != m->nr_priority_groups) { in multipath_ctr()
|