| /linux/drivers/md/ |
| H A D | dm-log-userspace-base.c | 465 static int flush_one_by_one(struct log_c *lc, struct list_head *flush_list) in flush_one_by_one() argument 470 list_for_each_entry(fe, flush_list, list) { in flush_one_by_one() 482 static int flush_by_group(struct log_c *lc, struct list_head *flush_list, in flush_by_group() argument 495 while (!list_empty(flush_list)) { in flush_by_group() 498 list_for_each_entry_safe(fe, tmp_fe, flush_list, list) { in flush_by_group() 528 list_splice_init(&tmp_list, flush_list); in flush_by_group() 529 r = flush_one_by_one(lc, flush_list); in flush_by_group() 539 list_splice_init(&tmp_list, flush_list); in flush_by_group()
|
| H A D | dm-zoned-target.c | 62 struct bio_list flush_list; member 518 bio = bio_list_pop(&dmz->flush_list); in dmz_flush_work() 661 bio_list_add(&dmz->flush_list, bio); in dmz_map() 914 bio_list_init(&dmz->flush_list); in dmz_ctr()
|
| H A D | dm-writecache.c | 197 struct bio_list flush_list; member 1291 bio = bio_list_pop(&wc->flush_list); in writecache_flush_thread() 1325 if (bio_list_empty(&wc->flush_list)) in writecache_offload_bio() 1327 bio_list_add(&wc->flush_list, bio); in writecache_offload_bio() 2524 bio_list_init(&wc->flush_list); in writecache_ctr()
|
| /linux/include/net/ |
| H A D | xdp_sock.h | 127 void __xsk_map_flush(struct list_head *flush_list); 217 static inline void __xsk_map_flush(struct list_head *flush_list) in __xsk_map_flush() argument
|
| /linux/kernel/bpf/ |
| H A D | cpumap.c | 768 struct list_head *flush_list = bpf_net_ctx_get_cpu_map_flush_list(); in bq_enqueue() local 770 list_add(&bq->flush_node, flush_list); in bq_enqueue() 803 void __cpu_map_flush(struct list_head *flush_list) in __cpu_map_flush() argument 807 list_for_each_entry_safe(bq, tmp, flush_list, flush_node) { in __cpu_map_flush()
|
| H A D | devmap.c | 423 void __dev_flush(struct list_head *flush_list) in __dev_flush() argument 427 list_for_each_entry_safe(bq, tmp, flush_list, flush_node) { in __dev_flush() 472 struct list_head *flush_list = bpf_net_ctx_get_dev_flush_list(); in bq_enqueue() local 476 list_add(&bq->flush_node, flush_list); in bq_enqueue()
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_admin.c | 635 LIST_HEAD(flush_list); in ionic_flush_qs() 644 list_add_tail(&qp->ibkill_flush_ent, &flush_list); in ionic_flush_qs() 648 list_for_each_entry_safe(qp, qp_tmp, &flush_list, ibkill_flush_ent) { in ionic_flush_qs() 658 list_add_tail(&cq->ibkill_flush_ent, &flush_list); in ionic_flush_qs() 662 list_for_each_entry_safe(cq, cq_tmp, &flush_list, ibkill_flush_ent) { in ionic_flush_qs()
|
| /linux/block/ |
| H A D | blk-flush.c | 331 list_add_tail(&flush_rq->queuelist, &q->flush_list); in blk_kick_flush()
|
| H A D | blk-mq.c | 1554 LIST_HEAD(flush_list); in blk_mq_requeue_work() 1559 list_splice_init(&q->flush_list, &flush_list); in blk_mq_requeue_work() 1577 while (!list_empty(&flush_list)) { in blk_mq_requeue_work() 1578 rq = list_entry(flush_list.next, struct request, queuelist); in blk_mq_requeue_work() 4603 INIT_LIST_HEAD(&q->flush_list); in blk_mq_init_allocated_queue()
|
| /linux/net/xdp/ |
| H A D | xsk.c | 402 struct list_head *flush_list = bpf_net_ctx_get_xskmap_flush_list(); in __xsk_map_redirect() local 404 list_add(&xs->flush_node, flush_list); in __xsk_map_redirect() 410 void __xsk_map_flush(struct list_head *flush_list) in __xsk_map_flush() argument 414 list_for_each_entry_safe(xs, tmp, flush_list, flush_node) { in __xsk_map_flush()
|
| /linux/include/linux/ |
| H A D | bpf.h | 2759 void __dev_flush(struct list_head *flush_list); 2772 void __cpu_map_flush(struct list_head *flush_list); 3015 static inline void __dev_flush(struct list_head *flush_list) in __dev_flush() argument 3061 static inline void __cpu_map_flush(struct list_head *flush_list) in __cpu_map_flush() argument
|
| H A D | blkdev.h | 579 struct list_head flush_list; member
|