Searched refs:cur_page (Results 1 – 12 of 12) sorted by relevance
| /linux/drivers/xen/ |
| H A D | xen-front-pgdir-shbuf.c | 251 int ret, cur_gref, cur_dir_page, cur_page, grefs_left; in backend_map() local 272 cur_page = 0; in backend_map() 284 addr = xen_page_to_vaddr(buf->pages[cur_page]); in backend_map() 285 gnttab_set_map_op(&map_ops[cur_page], addr, in backend_map() 289 cur_page++; in backend_map() 298 for (cur_page = 0; cur_page < buf->num_pages; cur_page++) { in backend_map() 299 if (likely(map_ops[cur_page].status == GNTST_okay)) { in backend_map() 300 buf->backend_map_handles[cur_page] = in backend_map() 301 map_ops[cur_page].handle; in backend_map() 303 buf->backend_map_handles[cur_page] = in backend_map() [all …]
|
| /linux/fs/ |
| H A D | direct-io.c | 99 struct page *cur_page; /* The page */ member 692 ret = bio_add_page(sdio->bio, sdio->cur_page, in dio_bio_add_page() 700 dio_pin_page(dio, sdio->cur_page); in dio_bio_add_page() 803 if (sdio->cur_page == page && in submit_page_section() 814 if (sdio->cur_page) { in submit_page_section() 816 dio_unpin_page(dio, sdio->cur_page); in submit_page_section() 817 sdio->cur_page = NULL; in submit_page_section() 823 sdio->cur_page = page; in submit_page_section() 837 dio_unpin_page(dio, sdio->cur_page); in submit_page_section() 838 sdio->cur_page = NULL; in submit_page_section() [all …]
|
| /linux/include/linux/ |
| H A D | mm_inline.h | 638 struct page *cur_page = pages[0]; in num_pages_contiguous() local 639 unsigned long section = memdesc_section(cur_page->flags); in num_pages_contiguous() 643 if (++cur_page != pages[i]) in num_pages_contiguous() 649 if (memdesc_section(cur_page->flags) != section) in num_pages_contiguous()
|
| /linux/drivers/net/wwan/t7xx/ |
| H A D | t7xx_hif_dpmaif_rx.c | 340 struct dpmaif_bat_page *cur_page = bat_skb + cur_bat_idx; in t7xx_dpmaif_rx_frag_alloc() local 344 if (!cur_page->page) { in t7xx_dpmaif_rx_frag_alloc() 364 cur_page->page = page; in t7xx_dpmaif_rx_frag_alloc() 365 cur_page->data_bus_addr = data_base_addr; in t7xx_dpmaif_rx_frag_alloc() 366 cur_page->offset = offset; in t7xx_dpmaif_rx_frag_alloc() 367 cur_page->data_len = bat_req->pkt_buf_sz; in t7xx_dpmaif_rx_frag_alloc() 370 data_base_addr = cur_page->data_bus_addr; in t7xx_dpmaif_rx_frag_alloc()
|
| /linux/lib/ |
| H A D | scatterlist.c | 437 unsigned int chunks, cur_page, seg_len, i, prv_len = 0; in sg_alloc_append_table_from_pages() local 490 cur_page = 0; in sg_alloc_append_table_from_pages() 496 for (j = cur_page + 1; j < n_pages; j++) { in sg_alloc_append_table_from_pages() 515 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_append_table_from_pages() 516 sg_set_page(s, pages[cur_page], in sg_alloc_append_table_from_pages() 521 cur_page = j; in sg_alloc_append_table_from_pages()
|
| /linux/arch/x86/include/asm/ |
| H A D | sev-common.h | 167 u64 cur_page : 12, member
|
| /linux/fs/ntfs3/ |
| H A D | fslog.c | 1809 struct RECORD_PAGE_HDR *cur_page = in last_log_lsn() local 1813 tail_page = cur_page; in last_log_lsn() 1825 cur_page->rhdr.lsn == page->rhdr.lsn && in last_log_lsn() 1826 cur_page->record_hdr.next_record_off == in last_log_lsn() 1833 cur_page = NULL; in last_log_lsn() 1841 lsn_cur = le64_to_cpu(cur_page->rhdr.lsn); in last_log_lsn() 1844 le64_to_cpu(cur_page->record_hdr.last_end_lsn) && in last_log_lsn() 1853 if (!is_log_record_end(cur_page)) { in last_log_lsn() 1861 log->last_lsn = le64_to_cpu(cur_page->record_hdr.last_end_lsn); in last_log_lsn() 1862 log->ra->current_lsn = cur_page->record_hdr.last_end_lsn; in last_log_lsn() [all …]
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | icp_qat_uclo.h | 162 struct icp_qat_uclo_page *cur_page[ICP_QAT_UCLO_MAX_CTX]; member
|
| H A D | qat_uclo.c | 2205 aed->ae_slices[s].cur_page[ctx] = in qat_uclo_wr_uimage_page()
|
| /linux/drivers/gpu/drm/i915/gvt/ |
| H A D | kvmgt.c | 150 struct page *cur_page; in gvt_pin_guest_page() local 153 IOMMU_READ | IOMMU_WRITE, &cur_page); in gvt_pin_guest_page() 161 base_page = cur_page; in gvt_pin_guest_page() 162 else if (page_to_pfn(base_page) + npage != page_to_pfn(cur_page)) { in gvt_pin_guest_page()
|
| /linux/mm/ |
| H A D | rmap.c | 1449 struct page *cur_page = page + i; in __folio_add_anon_rmap() local 1453 PageAnonExclusive(cur_page), folio); in __folio_add_anon_rmap() 1461 VM_WARN_ON_FOLIO(atomic_read(&cur_page->_mapcount) > 0 && in __folio_add_anon_rmap() 1462 PageAnonExclusive(cur_page), folio); in __folio_add_anon_rmap()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | sev.c | 3799 entry->cur_page = entry->pagesize ? 512 : 1; in __snp_complete_one_psc() 3865 if (entry_start.cur_page > npages || !IS_ALIGNED(gfn, npages)) { in snp_begin_psc() 3870 if (entry_start.cur_page) { in snp_begin_psc() 3878 npages -= entry_start.cur_page; in snp_begin_psc() 3879 gfn += entry_start.cur_page; in snp_begin_psc() 3907 entry.cur_page || !!entry.pagesize != huge) in snp_begin_psc()
|