| /linux/net/netfilter/ |
| H A D | nft_last.c | 17 struct nft_last *last; member 29 struct nft_last *last; in nft_last_init() local 33 last = kzalloc(sizeof(*last), GFP_KERNEL_ACCOUNT); in nft_last_init() 34 if (!last) in nft_last_init() 38 last->set = ntohl(nla_get_be32(tb[NFTA_LAST_SET])); in nft_last_init() 40 if (last->set && tb[NFTA_LAST_MSECS]) { in nft_last_init() 45 last->jiffies = jiffies - (unsigned long)last_jiffies; in nft_last_init() 47 priv->last = last; in nft_last_init() 51 kfree(last); in nft_last_init() 60 struct nft_last *last = priv->last; in nft_last_eval() local [all …]
|
| H A D | nft_set_pipapo_avx2.c | 154 union nft_pipapo_map_bucket *mt, bool last) in nft_pipapo_avx2_refill() argument 164 if (last) \ in nft_pipapo_avx2_refill() 217 bool first, bool last) in nft_pipapo_avx2_lookup_4b_2() argument 243 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_2() 244 if (last) in nft_pipapo_avx2_lookup_4b_2() 280 bool first, bool last) in nft_pipapo_avx2_lookup_4b_4() argument 320 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_4() 321 if (last) in nft_pipapo_avx2_lookup_4b_4() 357 bool first, bool last) in nft_pipapo_avx2_lookup_4b_8() argument 415 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_8() [all …]
|
| /linux/scripts/ |
| H A D | decode_stacktrace.sh | 271 local last=$(( ${#words[@]} - 1 )) 286 if [[ ${words[$last]} =~ \([A-Z]*\) ]]; then 287 info_str=${words[$last]} 288 unset words[$last] spaces[$last] 289 last=$(( $last - 1 )) 294 if [[ ${words[$last]} =~ ^[0-9a-f]+\] ]]; then 295 words[$last-1]="${words[$last-1]} ${words[$last]}" 296 unset words[$last] spaces[$last] 297 last=$(( $last - 1 )) 300 if [[ ${words[$last]} =~ \[([^]]+)\] ]]; then [all …]
|
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_uiom_interval_tree.c | 43 #define LAST(node) ((node)->last) 70 usnic_uiom_interval_node_alloc(long int start, long int last, int ref_cnt, in usnic_uiom_interval_node_alloc() argument 79 interval->last = last; in usnic_uiom_interval_node_alloc() 105 unsigned long start, unsigned long last, in find_intervals_intersection_sorted() argument 112 for (node = usnic_uiom_interval_tree_iter_first(root, start, last); in find_intervals_intersection_sorted() 114 node = usnic_uiom_interval_tree_iter_next(node, start, last)) in find_intervals_intersection_sorted() 120 int usnic_uiom_get_intervals_diff(unsigned long start, unsigned long last, in usnic_uiom_get_intervals_diff() argument 132 find_intervals_intersection_sorted(root, start, last, in usnic_uiom_get_intervals_diff() 148 if (pivot > interval->last) { in usnic_uiom_get_intervals_diff() 150 } else if (pivot <= interval->last && in usnic_uiom_get_intervals_diff() [all …]
|
| H A D | usnic_uiom_interval_tree.h | 43 unsigned long last; member 58 unsigned long last); 61 unsigned long start, unsigned long last); 67 unsigned long start, unsigned long last, 75 unsigned long start, unsigned long last, 82 unsigned long last, int flags,
|
| /linux/fs/xfs/scrub/ |
| H A D | bitmap.c | 51 uint64_t last); 55 uint64_t last); 78 uint64_t last = start + len - 1; local 80 while ((bn = xbitmap64_tree_iter_first(&bitmap->xb_root, start, last))) { 81 if (bn->bn_start < start && bn->bn_last > last) { 94 new_bn->bn_start = last + 1; 102 } else if (bn->bn_last > last) { 105 bn->bn_start = last + 1; 127 uint64_t last = start + len - 1; in xbitmap64_set() local 131 left = xbitmap64_tree_iter_first(&bitmap->xb_root, start, last); in xbitmap64_set() [all …]
|
| /linux/arch/mips/sibyte/common/ |
| H A D | cfe_console.c | 16 int i, last, written; in cfe_console_write() local 18 for (i=0, last=0; i<count; i++) { in cfe_console_write() 24 written = cfe_write(cfe_cons_handle, &str[last], i-last); in cfe_console_write() 27 last += written; in cfe_console_write() 28 } while (last < i); in cfe_console_write() 33 if (last != count) { in cfe_console_write() 35 written = cfe_write(cfe_cons_handle, &str[last], count-last); in cfe_console_write() 38 last += written; in cfe_console_write() 39 } while (last < count); in cfe_console_write()
|
| /linux/drivers/vhost/ |
| H A D | iotlb.c | 17 #define LAST(map) ((map)->last) 51 u64 start, u64 last, in vhost_iotlb_add_range_ctx() argument 57 if (last < start) in vhost_iotlb_add_range_ctx() 63 if (start == 0 && last == ULONG_MAX) { in vhost_iotlb_add_range_ctx() 64 u64 mid = last / 2; in vhost_iotlb_add_range_ctx() 87 map->size = last - start + 1; in vhost_iotlb_add_range_ctx() 88 map->last = last; in vhost_iotlb_add_range_ctx() 104 u64 start, u64 last, in vhost_iotlb_add_range() argument 107 return vhost_iotlb_add_range_ctx(iotlb, start, last, in vhost_iotlb_add_range() 118 void vhost_iotlb_del_range(struct vhost_iotlb *iotlb, u64 start, u64 last) in vhost_iotlb_del_range() argument [all …]
|
| /linux/lib/ |
| H A D | interval_tree_test.c | 33 search(struct rb_root_cached *root, unsigned long start, unsigned long last) in search() argument 38 for (node = interval_tree_iter_first(root, start, last); node; in search() 39 node = interval_tree_iter_next(node, start, last)) in search() 53 nodes[i].last = b; in init() 111 unsigned long last = search_all ? max_endpoint : queries[j]; in search_check() local 113 results += search(&root, start, last); in search_check() 133 unsigned long start, last; in intersection_range_check() local 164 last = ULONG_MAX; in intersection_range_check() 166 last = (prandom_u32_state(&rnd) >> 4) % max_endpoint; in intersection_range_check() 167 start = (prandom_u32_state(&rnd) >> 4) % last; in intersection_range_check() [all …]
|
| /linux/io_uring/ |
| H A D | slist.h | 30 list->last = node; in wq_list_add_after() 38 list->last = node; in wq_list_add_tail() 41 list->last->next = node; in wq_list_add_tail() 42 list->last = node; in wq_list_add_tail() 51 list->last = node; in wq_list_add_head() 56 struct io_wq_work_node *last, in wq_list_cut() argument 61 WRITE_ONCE(list->first, last->next); in wq_list_cut() 63 prev->next = last->next; in wq_list_cut() 65 if (last == list->last) in wq_list_cut() 66 list->last = prev; in wq_list_cut() [all …]
|
| /linux/kernel/bpf/ |
| H A D | range_tree.c | 131 range_it_iter_first(struct range_tree *rt, u32 start, u32 last) in range_it_iter_first() argument 133 return __range_it_iter_first(&rt->it_root, start, last); in range_it_iter_first() 139 u32 last = start + len - 1; in range_tree_clear() local 143 while ((rn = range_it_iter_first(rt, start, last))) { in range_tree_clear() 144 if (rn->rn_start < start && rn->rn_last > last) { in range_tree_clear() 158 new_rn->rn_start = last + 1; in range_tree_clear() 166 } else if (rn->rn_last > last) { in range_tree_clear() 169 rn->rn_start = last + 1; in range_tree_clear() 186 u32 last = start + len - 1; in is_range_tree_set() local 190 left = range_it_iter_first(rt, start, last); in is_range_tree_set() [all …]
|
| /linux/drivers/video/fbdev/core/ |
| H A D | fb_copyarea.h | 54 unsigned long first, last; in fb_copy_aligned_fwd() local 57 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), reverse); in fb_copy_aligned_fwd() 62 last = last ? (last & first) : first; in fb_copy_aligned_fwd() 65 if (last == ~0UL) in fb_copy_aligned_fwd() 68 fb_copy_offset_masked(last, 0, dst, src); in fb_copy_aligned_fwd() 90 if (last) in fb_copy_aligned_fwd() 91 fb_copy_offset_masked(last, offset, dst, src); in fb_copy_aligned_fwd() 100 unsigned long first, last; in fb_copy_aligned_rev() local 103 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), reverse); in fb_copy_aligned_rev() 107 if (last) in fb_copy_aligned_rev() [all …]
|
| H A D | fb_fillrect.h | 111 unsigned long first, last; in bitfill() local 115 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), pattern->reverse); in bitfill() 118 last = last ? (last & first) : first; in bitfill() 120 if (last == ~0UL) in bitfill() 122 else if (last) in bitfill() 123 fb_modify_offset(first, last, 0, dst); in bitfill() 147 if (last) in bitfill() 148 fb_modify_offset(get(pattern), last, offset, dst); in bitfill() 159 unsigned long first, last; in bitinvert() local 164 last = ~fb_pixel_mask(end & (BITS_PER_LONG-1), pattern->reverse); in bitinvert() [all …]
|
| /linux/include/trace/events/ |
| H A D | maple_tree.h | 24 __field(unsigned long, last) 33 __entry->last = mas->last; 43 (unsigned long) __entry->last 57 __field(unsigned long, last) 66 __entry->last = mas->last; 76 (unsigned long) __entry->last 92 __field(unsigned long, last) 103 __entry->last = mas->last; 115 (unsigned long) __entry->last,
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_svm.c | 77 svm_range_check_vm(struct kfd_process *p, uint64_t start, uint64_t last, 95 prange, prange->start, prange->last); in svm_range_unlink() 104 if (prange->it_node.start != 0 && prange->it_node.last != 0) in svm_range_unlink() 112 prange, prange->start, prange->last); in svm_range_add_notifier_locked() 131 prange, prange->start, prange->last); in svm_range_add_to_svms() 135 prange->it_node.last = prange->last; in svm_range_add_to_svms() 144 prange->notifier.interval_tree.last >> PAGE_SHIFT); in svm_range_remove_notifier() 147 prange->notifier.interval_tree.last != 0) in svm_range_remove_notifier() 281 uint64_t size = (prange->last - prange->start + 1) << PAGE_SHIFT; in svm_range_free() 286 prange->start, prange->last); in svm_range_free() [all …]
|
| /linux/tools/perf/util/ |
| H A D | ordered-events.c | 20 struct ordered_event *last = oe->last; in queue_event() local 25 oe->last = new; in queue_event() 29 if (!last) { in queue_event() 40 if (last->timestamp <= timestamp) { in queue_event() 41 while (last->timestamp <= timestamp) { in queue_event() 42 p = last->list.next; in queue_event() 48 last = list_entry(p, struct ordered_event, list); in queue_event() 50 list_add_tail(&new->list, &last->list); in queue_event() 52 while (last->timestamp > timestamp) { in queue_event() 53 p = last->list.prev; in queue_event() [all …]
|
| /linux/include/linux/ |
| H A D | interval_tree_generic.h | 42 ITTYPE start = ITSTART(node), last = ITLAST(node); \ 49 if (parent->ITSUBTREE < last) \ 50 parent->ITSUBTREE = last; \ 59 node->ITSUBTREE = last; \ 81 ITPREFIX ## _subtree_search(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ 104 if (ITSTART(node) <= last) { /* Cond1 */ \ 116 ITTYPE start, ITTYPE last) \ 141 if (ITSTART(leftmost) > last) \ 144 return ITPREFIX ## _subtree_search(node, start, last); \ 148 ITPREFIX ## _iter_next(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ [all …]
|
| /linux/tools/include/linux/ |
| H A D | interval_tree_generic.h | 42 ITTYPE start = ITSTART(node), last = ITLAST(node); \ 49 if (parent->ITSUBTREE < last) \ 50 parent->ITSUBTREE = last; \ 59 node->ITSUBTREE = last; \ 81 ITPREFIX ## _subtree_search(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ 104 if (ITSTART(node) <= last) { /* Cond1 */ \ 120 ITTYPE start, ITTYPE last) \ 145 if (ITSTART(leftmost) > last) \ 148 return ITPREFIX ## _subtree_search(node, start, last); \ 152 ITPREFIX ## _iter_next(ITSTRUCT *node, ITTYPE start, ITTYPE last) \ [all …]
|
| /linux/mm/ |
| H A D | interval_tree.c | 34 unsigned long last = vma_last_pgoff(node); in vma_interval_tree_insert_after() local 44 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 45 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 49 if (parent->shared.rb_subtree_last < last) in vma_interval_tree_insert_after() 50 parent->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 55 node->shared.rb_subtree_last = last; in vma_interval_tree_insert_after() 93 unsigned long first, unsigned long last) in anon_vma_interval_tree_iter_first() argument 95 return __anon_vma_interval_tree_iter_first(root, first, last); in anon_vma_interval_tree_iter_first() 100 unsigned long first, unsigned long last) in anon_vma_interval_tree_iter_next() argument 102 return __anon_vma_interval_tree_iter_next(node, first, last); in anon_vma_interval_tree_iter_next()
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | frwr_ops.c | 499 struct ib_send_wr *first, **prev, *last; in frwr_unmap_sync() local 516 last = &mr->mr_invwr; in frwr_unmap_sync() 517 last->next = NULL; in frwr_unmap_sync() 518 last->wr_cqe = &mr->mr_cqe; in frwr_unmap_sync() 519 last->sg_list = NULL; in frwr_unmap_sync() 520 last->num_sge = 0; in frwr_unmap_sync() 521 last->opcode = IB_WR_LOCAL_INV; in frwr_unmap_sync() 522 last->send_flags = IB_SEND_SIGNALED; in frwr_unmap_sync() 523 last->ex.invalidate_rkey = mr->mr_handle; in frwr_unmap_sync() 525 last->wr_cqe->done = frwr_wc_localinv; in frwr_unmap_sync() [all …]
|
| /linux/drivers/base/regmap/ |
| H A D | regcache-maple.c | 44 unsigned long index, last; in regcache_maple_write() local 60 last = reg; in regcache_maple_write() 65 lower_sz = (mas.last - mas.index + 1) * sizeof(unsigned long); in regcache_maple_write() 70 last = mas.last; in regcache_maple_write() 71 upper_sz = (mas.last - mas.index + 1) * sizeof(unsigned long); in regcache_maple_write() 76 entry = kmalloc_array(last - index + 1, sizeof(*entry), map->alloc_flags); in regcache_maple_write() 93 mas_set_range(&mas, index, last); in regcache_maple_write() 144 if (mas.last > max) { in regcache_maple_drop() 146 upper_last = mas.last; in regcache_maple_drop() 149 mas.last - max, sizeof(*upper), in regcache_maple_drop() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| H A D | io_pagetable.h | 79 return area->pages_node.last; in iopt_area_last_index() 89 return area->node.last; in iopt_area_last_iova() 94 return (area->node.last - area->node.start) + 1; in iopt_area_length() 121 unsigned long last) \ 127 last); \ 134 unsigned long last) \ 138 node = interval_tree_iter_next(&last_node->node, start, last); \ 234 unsigned long last, struct page **out_pages); 236 unsigned long last, struct page **out_pages); 238 unsigned long last); [all …]
|
| /linux/drivers/crypto/marvell/cesa/ |
| H A D | tdma.c | 44 engine->chain_sw.last = NULL; in mv_cesa_dma_step() 47 engine->chain_hw.last = dreq->chain.last; in mv_cesa_dma_step() 85 dreq->chain.last = NULL; in mv_cesa_dma_cleanup() 108 struct mv_cesa_tdma_desc *last = engine->chain_sw.last; in mv_cesa_tdma_chain() local 114 if (!last || dreq->chain.first->flags & CESA_TDMA_SET_STATE) in mv_cesa_tdma_chain() 117 last->next = dreq->chain.first; in mv_cesa_tdma_chain() 118 last->next_dma = cpu_to_le32(dreq->chain.first->cur_dma); in mv_cesa_tdma_chain() 120 last = dreq->chain.last; in mv_cesa_tdma_chain() 121 engine->chain_sw.last = last; in mv_cesa_tdma_chain() 126 if (last->flags & CESA_TDMA_BREAK_CHAIN) { in mv_cesa_tdma_chain() [all …]
|
| /linux/crypto/asymmetric_keys/ |
| H A D | pkcs7_trust.c | 27 struct x509_certificate *x509, *last = NULL, *p; in pkcs7_validate_trust_one() local 74 last = x509; in pkcs7_validate_trust_one() 75 sig = last->sig; in pkcs7_validate_trust_one() 81 if (last && (last->sig->auth_ids[0] || last->sig->auth_ids[1])) { in pkcs7_validate_trust_one() 83 last->sig->auth_ids[0], in pkcs7_validate_trust_one() 84 last->sig->auth_ids[1], in pkcs7_validate_trust_one() 87 x509 = last; in pkcs7_validate_trust_one()
|
| /linux/drivers/gpu/drm/i915/gt/uc/ |
| H A D | selftest_guc.c | 54 struct i915_request *last[3] = {NULL, NULL, NULL}, *rq; in intel_guc_scrub_ctbs() local 95 last[i] = rq; in intel_guc_scrub_ctbs() 99 ret = i915_request_wait(last[i], 0, HZ); in intel_guc_scrub_ctbs() 104 i915_request_put(last[i]); in intel_guc_scrub_ctbs() 105 last[i] = NULL; in intel_guc_scrub_ctbs() 124 if (last[i]) in intel_guc_scrub_ctbs() 125 i915_request_put(last[i]); in intel_guc_scrub_ctbs() 153 struct i915_request *spin_rq = NULL, *rq, *last = NULL; in intel_guc_steal_guc_ids() local 207 if ((ret != -EAGAIN) || !last) { in intel_guc_steal_guc_ids() 209 last ? "" : "first ", context_index, ERR_PTR(ret)); in intel_guc_steal_guc_ids() [all …]
|