Lines Matching full:if
20 if (!bio) in __blk_recalc_rq_segments()
35 if (high || highprv) in __blk_recalc_rq_segments()
37 if (cluster) { in __blk_recalc_rq_segments()
38 if (seg_size + bv->bv_len in __blk_recalc_rq_segments()
41 if (!BIOVEC_PHYS_MERGEABLE(bvprv, bv)) in __blk_recalc_rq_segments()
43 if (!BIOVEC_SEG_BOUNDARY(q, bvprv, bv)) in __blk_recalc_rq_segments()
51 if (nr_phys_segs == 1 && seg_size > in __blk_recalc_rq_segments()
63 if (nr_phys_segs == 1 && seg_size > fbio->bi_seg_front_size) in __blk_recalc_rq_segments()
65 if (seg_size > bbio->bi_seg_back_size) in __blk_recalc_rq_segments()
90 if (!blk_queue_cluster(q)) in blk_phys_contig_segment()
93 if (bio->bi_seg_back_size + nxt->bi_seg_front_size > in blk_phys_contig_segment()
97 if (!bio_has_data(bio)) in blk_phys_contig_segment()
100 if (!BIOVEC_PHYS_MERGEABLE(__BVEC_END(bio), __BVEC_START(nxt))) in blk_phys_contig_segment()
104 * bio and nxt are contiguous in memory; check if the queue allows in blk_phys_contig_segment()
107 if (BIO_SEG_BOUNDARY(q, bio, nxt)) in blk_phys_contig_segment()
136 if (bvprv && cluster) { in blk_rq_map_sg()
137 if (sg->length + nbytes > queue_max_segment_size(q)) in blk_rq_map_sg()
140 if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec)) in blk_rq_map_sg()
142 if (!BIOVEC_SEG_BOUNDARY(q, bvprv, bvec)) in blk_rq_map_sg()
148 if (!sg) in blk_rq_map_sg()
152 * If the driver previously mapped a shorter in blk_rq_map_sg()
172 if (unlikely(rq->cmd_flags & REQ_COPY_USER) && in blk_rq_map_sg()
181 if (q->dma_drain_size && q->dma_drain_needed(rq)) { in blk_rq_map_sg()
182 if (rq->cmd_flags & REQ_WRITE) in blk_rq_map_sg()
195 if (sg) in blk_rq_map_sg()
208 if (req->nr_phys_segments + nr_phys_segs > queue_max_segments(q)) in ll_new_hw_segment()
211 if (bio_integrity(bio) && blk_integrity_merge_bio(q, req, bio)) in ll_new_hw_segment()
223 if (req == q->last_merge) in ll_new_hw_segment()
233 if (unlikely(req->cmd_type == REQ_TYPE_BLOCK_PC)) in ll_back_merge_fn()
238 if (blk_rq_sectors(req) + bio_sectors(bio) > max_sectors) { in ll_back_merge_fn()
240 if (req == q->last_merge) in ll_back_merge_fn()
244 if (!bio_flagged(req->biotail, BIO_SEG_VALID)) in ll_back_merge_fn()
246 if (!bio_flagged(bio, BIO_SEG_VALID)) in ll_back_merge_fn()
257 if (unlikely(req->cmd_type == REQ_TYPE_BLOCK_PC)) in ll_front_merge_fn()
263 if (blk_rq_sectors(req) + bio_sectors(bio) > max_sectors) { in ll_front_merge_fn()
265 if (req == q->last_merge) in ll_front_merge_fn()
269 if (!bio_flagged(bio, BIO_SEG_VALID)) in ll_front_merge_fn()
271 if (!bio_flagged(req->bio, BIO_SEG_VALID)) in ll_front_merge_fn()
285 * First check if the either of the requests are re-queued in ll_merge_requests_fn()
286 * requests. Can't merge them if they are. in ll_merge_requests_fn()
288 if (req->special || next->special) in ll_merge_requests_fn()
294 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > queue_max_sectors(q)) in ll_merge_requests_fn()
298 if (blk_phys_contig_segment(q, req->biotail, next->bio)) { in ll_merge_requests_fn()
299 if (req->nr_phys_segments == 1) in ll_merge_requests_fn()
301 if (next->nr_phys_segments == 1) in ll_merge_requests_fn()
306 if (total_phys_segments > queue_max_segments(q)) in ll_merge_requests_fn()
309 if (blk_integrity_rq(req) && blk_integrity_merge_rq(q, req, next)) in ll_merge_requests_fn()
331 if (rq->cmd_flags & REQ_MIXED_MERGE) in blk_rq_set_mixed_merge()
349 if (blk_do_io_stat(req)) { in blk_account_io_merge()
370 if (!rq_mergeable(req) || !rq_mergeable(next)) in attempt_merge()
376 if ((req->cmd_flags & REQ_DISCARD) != (next->cmd_flags & REQ_DISCARD)) in attempt_merge()
382 if ((req->cmd_flags & REQ_SECURE) != (next->cmd_flags & REQ_SECURE)) in attempt_merge()
388 if (blk_rq_pos(req) + blk_rq_sectors(req) != blk_rq_pos(next)) in attempt_merge()
391 if (rq_data_dir(req) != rq_data_dir(next) in attempt_merge()
397 * If we are allowed to merge, then append bio list in attempt_merge()
402 if (!ll_merge_requests_fn(q, req, next)) in attempt_merge()
406 * If failfast settings disagree or any of the two is already in attempt_merge()
411 if ((req->cmd_flags | next->cmd_flags) & REQ_MIXED_MERGE || in attempt_merge()
424 if (time_after(req->start_time, next->start_time)) in attempt_merge()
440 if (blk_rq_cpu_valid(next)) in attempt_merge()
453 if (next) in attempt_back_merge()
463 if (prev) in attempt_front_merge()
477 if (!rq_mergeable(rq)) in blk_rq_merge_ok()
481 if ((bio->bi_rw & REQ_DISCARD) != (rq->bio->bi_rw & REQ_DISCARD)) in blk_rq_merge_ok()
485 if ((bio->bi_rw & REQ_SECURE) != (rq->bio->bi_rw & REQ_SECURE)) in blk_rq_merge_ok()
489 if (bio_data_dir(bio) != rq_data_dir(rq)) in blk_rq_merge_ok()
493 if (rq->rq_disk != bio->bi_bdev->bd_disk || rq->special) in blk_rq_merge_ok()
497 if (bio_integrity(bio) != blk_integrity_rq(rq)) in blk_rq_merge_ok()
505 if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_sector) in blk_try_merge()
507 else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_sector) in blk_try_merge()