| /linux/fs/udf/ |
| H A D | directory.c | 19 static int udf_verify_fi(struct udf_fileident_iter *iter) in udf_verify_fi() argument 23 if (iter->fi.descTag.tagIdent != cpu_to_le16(TAG_IDENT_FID)) { in udf_verify_fi() 24 udf_err(iter->dir->i_sb, in udf_verify_fi() 26 iter->dir->i_ino, (unsigned long long)iter->pos, in udf_verify_fi() 27 le16_to_cpu(iter->fi.descTag.tagIdent)); in udf_verify_fi() 30 len = udf_dir_entry_len(&iter->fi); in udf_verify_fi() 31 if (le16_to_cpu(iter->fi.lengthOfImpUse) & 3) { in udf_verify_fi() 32 udf_err(iter->dir->i_sb, in udf_verify_fi() 34 iter->dir->i_ino, (unsigned long long)iter->pos); in udf_verify_fi() 42 if (len > 1 << iter->dir->i_blkbits) { in udf_verify_fi() [all …]
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | tdp_iter.c | 12 static void tdp_iter_refresh_sptep(struct tdp_iter *iter) in tdp_iter_refresh_sptep() argument 14 iter->sptep = iter->pt_path[iter->level - 1] + in tdp_iter_refresh_sptep() 15 SPTE_INDEX((iter->gfn | iter->gfn_bits) << PAGE_SHIFT, iter->level); in tdp_iter_refresh_sptep() 16 iter->old_spte = kvm_tdp_mmu_read_spte(iter->sptep); in tdp_iter_refresh_sptep() 23 void tdp_iter_restart(struct tdp_iter *iter) in tdp_iter_restart() argument 25 iter->yielded = false; in tdp_iter_restart() 26 iter->yielded_gfn = iter->next_last_level_gfn; in tdp_iter_restart() 27 iter->level = iter->root_level; in tdp_iter_restart() 29 iter->gfn = gfn_round_for_level(iter->next_last_level_gfn, iter->level); in tdp_iter_restart() 30 tdp_iter_refresh_sptep(iter); in tdp_iter_restart() [all …]
|
| H A D | tdp_mmu.c | 240 struct tdp_iter *iter) in tdp_mmu_init_child_sp() argument 245 parent_sp = sptep_to_sp(rcu_dereference(iter->sptep)); in tdp_mmu_init_child_sp() 250 tdp_mmu_init_sp(child_sp, iter->sptep, iter->gfn, role); in tdp_mmu_init_child_sp() 651 struct tdp_iter *iter, in __tdp_mmu_set_spte_atomic() argument 660 WARN_ON_ONCE(iter->yielded || is_frozen_spte(iter->old_spte)); in __tdp_mmu_set_spte_atomic() 662 if (is_mirror_sptep(iter->sptep) && !is_frozen_spte(new_spte)) { in __tdp_mmu_set_spte_atomic() 672 ret = set_external_spte_present(kvm, iter->sptep, iter->gfn, in __tdp_mmu_set_spte_atomic() 673 iter->old_spte, new_spte, iter->level); in __tdp_mmu_set_spte_atomic() 677 u64 *sptep = rcu_dereference(iter->sptep); in __tdp_mmu_set_spte_atomic() 687 if (!try_cmpxchg64(sptep, &iter->old_spte, new_spte)) in __tdp_mmu_set_spte_atomic() [all …]
|
| /linux/lib/ |
| H A D | interval_tree.c | 50 void interval_tree_span_iter_first(struct interval_tree_span_iter *iter, in interval_tree_span_iter_first() argument 55 iter->first_index = first_index; in interval_tree_span_iter_first() 56 iter->last_index = last_index; in interval_tree_span_iter_first() 57 iter->nodes[0] = NULL; in interval_tree_span_iter_first() 58 iter->nodes[1] = in interval_tree_span_iter_first() 60 if (!iter->nodes[1]) { in interval_tree_span_iter_first() 62 iter->start_hole = first_index; in interval_tree_span_iter_first() 63 iter->last_hole = last_index; in interval_tree_span_iter_first() 64 iter->is_hole = 1; in interval_tree_span_iter_first() 67 if (iter->nodes[1]->start > first_index) { in interval_tree_span_iter_first() [all …]
|
| /linux/fs/iomap/ |
| H A D | iter.c | 9 static inline void iomap_iter_reset_iomap(struct iomap_iter *iter) in iomap_iter_reset_iomap() argument 11 if (iter->iomap.flags & IOMAP_F_FOLIO_BATCH) { in iomap_iter_reset_iomap() 12 folio_batch_release(iter->fbatch); in iomap_iter_reset_iomap() 13 folio_batch_reinit(iter->fbatch); in iomap_iter_reset_iomap() 14 iter->iomap.flags &= ~IOMAP_F_FOLIO_BATCH; in iomap_iter_reset_iomap() 17 iter->status = 0; in iomap_iter_reset_iomap() 18 memset(&iter->iomap, 0, sizeof(iter->iomap)); in iomap_iter_reset_iomap() 19 memset(&iter->srcmap, 0, sizeof(iter->srcmap)); in iomap_iter_reset_iomap() 23 int iomap_iter_advance(struct iomap_iter *iter, u64 count) in iomap_iter_advance() argument 25 if (WARN_ON_ONCE(count > iomap_length(iter))) in iomap_iter_advance() [all …]
|
| H A D | seek.c | 9 static int iomap_seek_hole_iter(struct iomap_iter *iter, in iomap_seek_hole_iter() argument 12 loff_t length = iomap_length(iter); in iomap_seek_hole_iter() 14 switch (iter->iomap.type) { in iomap_seek_hole_iter() 16 *hole_pos = mapping_seek_hole_data(iter->inode->i_mapping, in iomap_seek_hole_iter() 17 iter->pos, iter->pos + length, SEEK_HOLE); in iomap_seek_hole_iter() 18 if (*hole_pos == iter->pos + length) in iomap_seek_hole_iter() 19 return iomap_iter_advance(iter, length); in iomap_seek_hole_iter() 22 *hole_pos = iter->pos; in iomap_seek_hole_iter() 25 return iomap_iter_advance(iter, length); in iomap_seek_hole_iter() 33 struct iomap_iter iter = { in iomap_seek_hole() local [all …]
|
| H A D | buffered-io.c | 351 static inline bool iomap_block_needs_zeroing(const struct iomap_iter *iter, in iomap_block_needs_zeroing() argument 354 const struct iomap *srcmap = iomap_iter_srcmap(iter); in iomap_block_needs_zeroing() 358 pos >= i_size_read(iter->inode); in iomap_block_needs_zeroing() 370 static int iomap_read_inline_data(const struct iomap_iter *iter, in iomap_read_inline_data() argument 373 const struct iomap *iomap = iomap_iter_srcmap(iter); in iomap_read_inline_data() 374 size_t size = i_size_read(iter->inode) - iomap->offset; in iomap_read_inline_data() 384 fserror_report_io(iter->inode, FSERR_BUFFERED_READ, in iomap_read_inline_data() 389 ifs_alloc(iter->inode, folio, iter->flags); in iomap_read_inline_data() 509 static int iomap_read_folio_iter(struct iomap_iter *iter, in iomap_read_folio_iter() argument 512 const struct iomap *iomap = &iter->iomap; in iomap_read_folio_iter() [all …]
|
| /linux/block/ |
| H A D | blk-mq-dma.c | 9 static bool __blk_map_iter_next(struct blk_map_iter *iter) in __blk_map_iter_next() argument 11 if (iter->iter.bi_size) in __blk_map_iter_next() 13 if (!iter->bio || !iter->bio->bi_next) in __blk_map_iter_next() 16 iter->bio = iter->bio->bi_next; in __blk_map_iter_next() 17 if (iter->is_integrity) { in __blk_map_iter_next() 18 iter->iter = bio_integrity(iter->bio)->bip_iter; in __blk_map_iter_next() 19 iter->bvecs = bio_integrity(iter->bio)->bip_vec; in __blk_map_iter_next() 21 iter->iter = iter->bio->bi_iter; in __blk_map_iter_next() 22 iter->bvecs = iter->bio->bi_io_vec; in __blk_map_iter_next() 27 static bool blk_map_iter_next(struct request *req, struct blk_map_iter *iter, in blk_map_iter_next() argument [all …]
|
| H A D | t10-pi.c | 37 static void t10_pi_generate(struct blk_integrity_iter *iter, in t10_pi_generate() argument 43 for (i = 0 ; i < iter->data_size ; i += iter->interval) { in t10_pi_generate() 44 struct t10_pi_tuple *pi = iter->prot_buf + offset; in t10_pi_generate() 46 pi->guard_tag = t10_pi_csum(0, iter->data_buf, iter->interval, in t10_pi_generate() 50 iter->prot_buf, offset, bi->csum_type); in t10_pi_generate() 54 pi->ref_tag = cpu_to_be32(lower_32_bits(iter->seed)); in t10_pi_generate() 58 iter->data_buf += iter->interval; in t10_pi_generate() 59 iter->prot_buf += bi->metadata_size; in t10_pi_generate() 60 iter->seed++; in t10_pi_generate() 64 static blk_status_t t10_pi_verify(struct blk_integrity_iter *iter, in t10_pi_verify() argument [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_displayid.c | 82 static const u8 *find_next_displayid_extension(struct displayid_iter *iter) in find_next_displayid_extension() argument 86 bool ignore_checksum = iter->quirks & BIT(QUIRK_IGNORE_CHECKSUM); in find_next_displayid_extension() 88 displayid = drm_edid_find_extension(iter->drm_edid, DISPLAYID_EXT, &iter->ext_index); in find_next_displayid_extension() 93 iter->length = EDID_LENGTH - 1; in find_next_displayid_extension() 94 iter->idx = 1; in find_next_displayid_extension() 96 base = validate_displayid(displayid, iter->length, iter->idx, ignore_checksum); in find_next_displayid_extension() 100 iter->length = iter->idx + sizeof(*base) + base->bytes; in find_next_displayid_extension() 106 struct displayid_iter *iter) in displayid_iter_edid_begin() argument 108 memset(iter, 0, sizeof(*iter)); in displayid_iter_edid_begin() 110 iter->drm_edid = drm_edid; in displayid_iter_edid_begin() [all …]
|
| H A D | drm_vma_manager.c | 146 struct rb_node *iter; in drm_vma_offset_lookup_locked() local 149 iter = mgr->vm_addr_space_mm.interval_tree.rb_root.rb_node; in drm_vma_offset_lookup_locked() 152 while (likely(iter)) { in drm_vma_offset_lookup_locked() 153 node = rb_entry(iter, struct drm_mm_node, rb); in drm_vma_offset_lookup_locked() 156 iter = iter->rb_right; in drm_vma_offset_lookup_locked() 161 iter = iter->rb_left; in drm_vma_offset_lookup_locked() 247 struct rb_node **iter; in vma_node_allow() local 260 iter = &node->vm_files.rb_node; in vma_node_allow() 262 while (likely(*iter)) { in vma_node_allow() 263 parent = *iter; in vma_node_allow() [all …]
|
| /linux/arch/arm64/kvm/vgic/ |
| H A D | vgic-debug.c | 33 static void iter_next(struct kvm *kvm, struct vgic_state_iter *iter) in iter_next() argument 37 if (iter->dist_id == 0) { in iter_next() 38 iter->dist_id++; in iter_next() 46 if (iter->intid >= (iter->nr_spis + VGIC_NR_PRIVATE_IRQS - 1)) { in iter_next() 47 if (iter->intid == VGIC_LPI_MAX_INTID + 1) in iter_next() 51 if (!xa_find_after(&dist->lpi_xa, &iter->intid, in iter_next() 53 iter->intid = VGIC_LPI_MAX_INTID + 1; in iter_next() 58 iter->intid++; in iter_next() 59 if (iter->intid == VGIC_NR_PRIVATE_IRQS && in iter_next() 60 ++iter->vcpu_id < iter->nr_cpus) in iter_next() [all …]
|
| /linux/include/linux/ |
| H A D | bvec.h | 111 #define __bvec_iter_bvec(bvec, iter) (&(bvec)[(iter).bi_idx]) argument 114 #define mp_bvec_iter_page(bvec, iter) \ argument 115 (__bvec_iter_bvec((bvec), (iter))->bv_page) 117 #define mp_bvec_iter_len(bvec, iter) \ argument 118 min((iter).bi_size, \ 119 __bvec_iter_bvec((bvec), (iter))->bv_len - (iter).bi_bvec_done) 121 #define mp_bvec_iter_offset(bvec, iter) \ argument 122 (__bvec_iter_bvec((bvec), (iter))->bv_offset + (iter).bi_bvec_done) 124 #define mp_bvec_iter_page_idx(bvec, iter) \ argument 125 (mp_bvec_iter_offset((bvec), (iter)) / PAGE_SIZE) [all …]
|
| H A D | iov_iter.h | 24 size_t iterate_ubuf(struct iov_iter *iter, size_t len, void *priv, void *priv2, in iterate_ubuf() argument 27 void __user *base = iter->ubuf; in iterate_ubuf() 30 remain = step(base + iter->iov_offset, 0, len, priv, priv2); in iterate_ubuf() 32 iter->iov_offset += progress; in iterate_ubuf() 33 iter->count -= progress; in iterate_ubuf() 41 size_t iterate_iovec(struct iov_iter *iter, size_t len, void *priv, void *priv2, in iterate_iovec() argument 44 const struct iovec *p = iter->__iov; in iterate_iovec() 45 size_t progress = 0, skip = iter->iov_offset; in iterate_iovec() 64 iter->nr_segs -= p - iter->__iov; in iterate_iovec() 65 iter->__iov = p; in iterate_iovec() [all …]
|
| H A D | radix-tree.h | 233 struct radix_tree_iter *iter, void __rcu **slot); 249 const struct radix_tree_iter *iter, unsigned int tag); 264 struct radix_tree_iter *iter, gfp_t gfp, 281 radix_tree_iter_init(struct radix_tree_iter *iter, unsigned long start) in radix_tree_iter_init() argument 291 iter->index = 0; in radix_tree_iter_init() 292 iter->next_index = start; in radix_tree_iter_init() 310 struct radix_tree_iter *iter, unsigned flags); 324 struct radix_tree_iter *iter, unsigned long index) in radix_tree_iter_lookup() argument 326 radix_tree_iter_init(iter, index); in radix_tree_iter_lookup() 327 return radix_tree_next_chunk(root, iter, RADIX_TREE_ITER_CONTIG); in radix_tree_iter_lookup() [all …]
|
| /linux/net/netlabel/ |
| H A D | netlabel_addrlist.c | 49 struct netlbl_af4list *iter; in netlbl_af4list_search() local 51 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search() 52 if (iter->valid && (addr & iter->mask) == iter->addr) in netlbl_af4list_search() 53 return iter; in netlbl_af4list_search() 74 struct netlbl_af4list *iter; in netlbl_af4list_search_exact() local 76 list_for_each_entry_rcu(iter, head, list) in netlbl_af4list_search_exact() 77 if (iter->valid && iter->addr == addr && iter->mask == mask) in netlbl_af4list_search_exact() 78 return iter; in netlbl_af4list_search_exact() 99 struct netlbl_af6list *iter; in netlbl_af6list_search() local 101 list_for_each_entry_rcu(iter, head, list) in netlbl_af6list_search() [all …]
|
| /linux/kernel/ |
| H A D | kallsyms.c | 608 static int get_ksymbol_mod(struct kallsym_iter *iter) in get_ksymbol_mod() argument 610 int ret = module_get_kallsym(iter->pos - kallsyms_num_syms, in get_ksymbol_mod() 611 &iter->value, &iter->type, in get_ksymbol_mod() 612 iter->name, iter->module_name, in get_ksymbol_mod() 613 &iter->exported); in get_ksymbol_mod() 615 iter->pos_mod_end = iter->pos; in get_ksymbol_mod() 627 static int get_ksymbol_ftrace_mod(struct kallsym_iter *iter) in get_ksymbol_ftrace_mod() argument 629 int ret = ftrace_mod_get_kallsym(iter->pos - iter->pos_mod_end, in get_ksymbol_ftrace_mod() 630 &iter->value, &iter->type, in get_ksymbol_ftrace_mod() 631 iter->name, iter->module_name, in get_ksymbol_ftrace_mod() [all …]
|
| H A D | ucount.c | 217 struct ucounts *ucounts, *iter, *bad; in inc_ucount() local 220 for (iter = ucounts; iter; iter = tns->ucounts) { in inc_ucount() 222 tns = iter->ns; in inc_ucount() 224 if (!atomic_long_inc_below(&iter->ucount[type], max)) in inc_ucount() 229 bad = iter; in inc_ucount() 230 for (iter = ucounts; iter != bad; iter = iter->ns->ucounts) in inc_ucount() 231 atomic_long_dec(&iter->ucount[type]); in inc_ucount() 239 struct ucounts *iter; in dec_ucount() local 240 for (iter = ucounts; iter; iter = iter->ns->ucounts) { in dec_ucount() 241 long dec = atomic_long_dec_if_positive(&iter->ucount[type]); in dec_ucount() [all …]
|
| /linux/kernel/trace/ |
| H A D | trace_output.c | 30 enum print_line_t trace_print_bputs_msg_only(struct trace_iterator *iter) in trace_print_bputs_msg_only() argument 32 struct trace_seq *s = &iter->seq; in trace_print_bputs_msg_only() 33 struct trace_entry *entry = iter->ent; in trace_print_bputs_msg_only() 43 enum print_line_t trace_print_bprintk_msg_only(struct trace_iterator *iter) in trace_print_bprintk_msg_only() argument 45 struct trace_seq *s = &iter->seq; in trace_print_bprintk_msg_only() 46 struct trace_entry *entry = iter->ent; in trace_print_bprintk_msg_only() 56 enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter) in trace_print_printk_msg_only() argument 58 struct trace_seq *s = &iter->seq; in trace_print_printk_msg_only() 59 struct trace_entry *entry = iter->ent; in trace_print_printk_msg_only() 213 trace_print_bitmask_seq(struct trace_iterator *iter, void *bitmask_ptr, in trace_print_bitmask_seq() argument [all …]
|
| /linux/fs/afs/ |
| H A D | dir_search.c | 40 static bool afs_dir_reset_iter(struct afs_dir_iter *iter) in afs_dir_reset_iter() argument 42 unsigned long long i_size = i_size_read(&iter->dvnode->netfs.inode); in afs_dir_reset_iter() 49 iter->loop_check = nblocks * (AFS_DIR_SLOTS_PER_BLOCK - AFS_DIR_RESV_BLOCKS); in afs_dir_reset_iter() 50 iter->prev_entry = 0; /* Hash head is previous */ in afs_dir_reset_iter() 57 bool afs_dir_init_iter(struct afs_dir_iter *iter, const struct qstr *name) in afs_dir_init_iter() argument 59 iter->nr_slots = afs_dir_calc_slots(name->len); in afs_dir_init_iter() 60 iter->bucket = afs_dir_hash_name(name); in afs_dir_init_iter() 61 return afs_dir_reset_iter(iter); in afs_dir_init_iter() 67 union afs_xdr_dir_block *afs_dir_find_block(struct afs_dir_iter *iter, size_t block) in afs_dir_find_block() argument 69 struct folio_queue *fq = iter->fq; in afs_dir_find_block() [all …]
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_dump.c | 67 static void etnaviv_core_dump_header(struct core_dump_iterator *iter, in etnaviv_core_dump_header() argument 70 struct etnaviv_dump_object_header *hdr = iter->hdr; in etnaviv_core_dump_header() 74 hdr->file_offset = cpu_to_le32(iter->data - iter->start); in etnaviv_core_dump_header() 75 hdr->file_size = cpu_to_le32(data_end - iter->data); in etnaviv_core_dump_header() 77 iter->hdr++; in etnaviv_core_dump_header() 78 iter->data += le32_to_cpu(hdr->file_size); in etnaviv_core_dump_header() 81 static void etnaviv_core_dump_registers(struct core_dump_iterator *iter, in etnaviv_core_dump_registers() argument 84 struct etnaviv_dump_registers *reg = iter->data; in etnaviv_core_dump_registers() 97 etnaviv_core_dump_header(iter, ETDUMP_BUF_REG, reg); in etnaviv_core_dump_registers() 100 static void etnaviv_core_dump_mmu(struct core_dump_iterator *iter, in etnaviv_core_dump_mmu() argument [all …]
|
| /linux/drivers/gpu/drm/panfrost/ |
| H A D | panfrost_dump.c | 61 static void panfrost_core_dump_header(struct panfrost_dump_iterator *iter, in panfrost_core_dump_header() argument 64 struct panfrost_dump_object_header *hdr = iter->hdr; in panfrost_core_dump_header() 68 hdr->file_offset = iter->data - iter->start; in panfrost_core_dump_header() 69 hdr->file_size = data_end - iter->data; in panfrost_core_dump_header() 71 iter->hdr++; in panfrost_core_dump_header() 72 iter->data += hdr->file_size; in panfrost_core_dump_header() 76 panfrost_core_dump_registers(struct panfrost_dump_iterator *iter, in panfrost_core_dump_registers() argument 80 struct panfrost_dump_registers *dumpreg = iter->data; in panfrost_core_dump_registers() 100 panfrost_core_dump_header(iter, PANFROSTDUMP_BUF_REG, dumpreg); in panfrost_core_dump_registers() 106 struct panfrost_dump_iterator iter; in panfrost_core_dump() local [all …]
|
| /linux/drivers/video/fbdev/core/ |
| H A D | fb_imageblit.h | 39 struct fb_bitmap_iter *iter = iterator; in fb_bitmap_image() local 41 if (iter->i < iter->width) { in fb_bitmap_image() 42 int bit = ~iter->i & (BITS_PER_BYTE-1); in fb_bitmap_image() 43 int byte = iter->i++ / BITS_PER_BYTE; in fb_bitmap_image() 45 *pixels = iter->colors[(iter->data[byte] >> bit) & 1]; in fb_bitmap_image() 48 iter->data += BITS_TO_BYTES(iter->width); in fb_bitmap_image() 49 iter->i = 0; in fb_bitmap_image() 64 struct fb_color_iter *iter = iterator; in fb_color_image() local 66 if (iter->i < iter->width) { in fb_color_image() 67 unsigned long color = iter->data[iter->i++]; in fb_color_image() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | iters_state_safety.c | 36 struct bpf_iter_num iter; in create_and_destroy() local 48 : __imm_ptr(iter), ITER_HELPERS in create_and_destroy() 59 struct bpf_iter_num iter; in create_and_forget_to_destroy_fail() local 68 : __imm_ptr(iter), ITER_HELPERS in create_and_forget_to_destroy_fail() 80 struct bpf_iter_num iter; in destroy_without_creating_fail() local 86 : __imm_ptr(iter), ITER_HELPERS in destroy_without_creating_fail() 97 struct bpf_iter_num iter; in compromise_iter_w_direct_write_fail() local 113 : __imm_ptr(iter), ITER_HELPERS in compromise_iter_w_direct_write_fail() 124 struct bpf_iter_num iter; in compromise_iter_w_direct_write_and_skip_destroy_fail() local 138 : __imm_ptr(iter), ITER_HELPERS in compromise_iter_w_direct_write_and_skip_destroy_fail() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | icm.h | 97 struct mlx4_icm_iter *iter) in mlx4_icm_first() argument 99 iter->icm = icm; in mlx4_icm_first() 100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 103 iter->page_idx = 0; in mlx4_icm_first() 106 static inline int mlx4_icm_last(struct mlx4_icm_iter *iter) in mlx4_icm_last() argument 108 return !iter->chunk; in mlx4_icm_last() 111 static inline void mlx4_icm_next(struct mlx4_icm_iter *iter) in mlx4_icm_next() argument 113 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next() 115 iter->chunk = NULL; in mlx4_icm_next() [all …]
|