blk.h 12.2 KB
Newer Older
1
/* SPDX-License-Identifier: GPL-2.0 */
2 3 4
#ifndef BLK_INTERNAL_H
#define BLK_INTERNAL_H

5
#include <linux/idr.h>
6
#include <linux/blk-mq.h>
7
#include <linux/part_stat.h>
8
#include <linux/blk-crypto.h>
9
#include <linux/memblock.h>	/* for max_pfn/max_low_pfn */
10
#include <xen/xen.h>
11
#include "blk-crypto-internal.h"
12
#include "blk-mq.h"
13
#include "blk-mq-sched.h"
14

15 16
struct elevator_type;

17 18 19
/* Max future timer expiry for timeouts */
#define BLK_MAX_TIMEOUT		(5 * HZ)

20 21
extern struct dentry *blk_debugfs_root;

22 23 24
struct blk_flush_queue {
	unsigned int		flush_pending_idx:1;
	unsigned int		flush_running_idx:1;
25
	blk_status_t 		rq_status;
26 27 28 29
	unsigned long		flush_pending_since;
	struct list_head	flush_queue[2];
	struct list_head	flush_data_in_flight;
	struct request		*flush_rq;
30

31 32 33
	spinlock_t		mq_flush_lock;
};

34 35
extern struct kmem_cache *blk_requestq_cachep;
extern struct kobj_type blk_queue_ktype;
36
extern struct ida blk_queue_ida;
37

38 39
static inline struct blk_flush_queue *
blk_get_flush_queue(struct request_queue *q, struct blk_mq_ctx *ctx)
40
{
41
	return blk_mq_map_queue(q, REQ_OP_FLUSH, ctx)->fq;
42 43
}

T
Tejun Heo 已提交
44 45 46 47 48
static inline void __blk_get_queue(struct request_queue *q)
{
	kobject_get(&q->kobj);
}

M
Ming Lei 已提交
49
bool is_flush_rq(struct request *req);
50

51 52
struct blk_flush_queue *blk_alloc_flush_queue(int node, int cmd_size,
					      gfp_t flags);
53
void blk_free_flush_queue(struct blk_flush_queue *q);
54

55
void blk_freeze_queue(struct request_queue *q);
56
void __blk_mq_unfreeze_queue(struct request_queue *q, bool force_atomic);
57
void blk_queue_start_drain(struct request_queue *q);
58

59
#define BIO_INLINE_VECS 4
60 61 62
struct bio_vec *bvec_alloc(mempool_t *pool, unsigned short *nr_vecs,
		gfp_t gfp_mask);
void bvec_free(mempool_t *pool, struct bio_vec *bv, unsigned short nr_vecs);
63

64 65
static inline bool biovec_phys_mergeable(struct request_queue *q,
		struct bio_vec *vec1, struct bio_vec *vec2)
66
{
67
	unsigned long mask = queue_segment_boundary(q);
C
Christoph Hellwig 已提交
68 69
	phys_addr_t addr1 = page_to_phys(vec1->bv_page) + vec1->bv_offset;
	phys_addr_t addr2 = page_to_phys(vec2->bv_page) + vec2->bv_offset;
70 71

	if (addr1 + vec1->bv_len != addr2)
72
		return false;
73
	if (xen_domain() && !xen_biovec_phys_mergeable(vec1, vec2->bv_page))
74
		return false;
75 76
	if ((addr1 | mask) != ((addr2 + vec2->bv_len - 1) | mask))
		return false;
77 78 79
	return true;
}

80 81 82
static inline bool __bvec_gap_to_prev(struct request_queue *q,
		struct bio_vec *bprv, unsigned int offset)
{
83
	return (offset & queue_virt_boundary(q)) ||
84 85 86 87 88 89 90 91 92 93 94 95 96 97 98
		((bprv->bv_offset + bprv->bv_len) & queue_virt_boundary(q));
}

/*
 * Check if adding a bio_vec after bprv with offset would create a gap in
 * the SG list. Most drivers don't care about this, but some do.
 */
static inline bool bvec_gap_to_prev(struct request_queue *q,
		struct bio_vec *bprv, unsigned int offset)
{
	if (!queue_virt_boundary(q))
		return false;
	return __bvec_gap_to_prev(q, bprv, offset);
}

99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136
static inline bool rq_mergeable(struct request *rq)
{
	if (blk_rq_is_passthrough(rq))
		return false;

	if (req_op(rq) == REQ_OP_FLUSH)
		return false;

	if (req_op(rq) == REQ_OP_WRITE_ZEROES)
		return false;

	if (req_op(rq) == REQ_OP_ZONE_APPEND)
		return false;

	if (rq->cmd_flags & REQ_NOMERGE_FLAGS)
		return false;
	if (rq->rq_flags & RQF_NOMERGE_FLAGS)
		return false;

	return true;
}

/*
 * There are two different ways to handle DISCARD merges:
 *  1) If max_discard_segments > 1, the driver treats every bio as a range and
 *     send the bios to controller together. The ranges don't need to be
 *     contiguous.
 *  2) Otherwise, the request will be normal read/write requests.  The ranges
 *     need to be contiguous.
 */
static inline bool blk_discard_mergable(struct request *req)
{
	if (req_op(req) == REQ_OP_DISCARD &&
	    queue_max_discard_segments(req->q) > 1)
		return true;
	return false;
}

137 138
#ifdef CONFIG_BLK_DEV_INTEGRITY
void blk_flush_integrity(void);
139
bool __bio_integrity_endio(struct bio *);
140
void bio_integrity_free(struct bio *bio);
141 142 143 144 145 146
static inline bool bio_integrity_endio(struct bio *bio)
{
	if (bio_integrity(bio))
		return __bio_integrity_endio(bio);
	return true;
}
147

148 149
bool blk_integrity_merge_rq(struct request_queue *, struct request *,
		struct request *);
150 151
bool blk_integrity_merge_bio(struct request_queue *, struct request *,
		struct bio *);
152

153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171
static inline bool integrity_req_gap_back_merge(struct request *req,
		struct bio *next)
{
	struct bio_integrity_payload *bip = bio_integrity(req->bio);
	struct bio_integrity_payload *bip_next = bio_integrity(next);

	return bvec_gap_to_prev(req->q, &bip->bip_vec[bip->bip_vcnt - 1],
				bip_next->bip_vec[0].bv_offset);
}

static inline bool integrity_req_gap_front_merge(struct request *req,
		struct bio *bio)
{
	struct bio_integrity_payload *bip = bio_integrity(bio);
	struct bio_integrity_payload *bip_next = bio_integrity(req->bio);

	return bvec_gap_to_prev(req->q, &bip->bip_vec[bip->bip_vcnt - 1],
				bip_next->bip_vec[0].bv_offset);
}
172

173
int blk_integrity_add(struct gendisk *disk);
174
void blk_integrity_del(struct gendisk *);
175
#else /* CONFIG_BLK_DEV_INTEGRITY */
176 177 178 179 180
static inline bool blk_integrity_merge_rq(struct request_queue *rq,
		struct request *r1, struct request *r2)
{
	return true;
}
181 182 183 184 185
static inline bool blk_integrity_merge_bio(struct request_queue *rq,
		struct request *r, struct bio *b)
{
	return true;
}
186 187 188 189 190 191 192 193 194 195 196
static inline bool integrity_req_gap_back_merge(struct request *req,
		struct bio *next)
{
	return false;
}
static inline bool integrity_req_gap_front_merge(struct request *req,
		struct bio *bio)
{
	return false;
}

197 198 199
static inline void blk_flush_integrity(void)
{
}
200 201 202 203
static inline bool bio_integrity_endio(struct bio *bio)
{
	return true;
}
204 205 206
static inline void bio_integrity_free(struct bio *bio)
{
}
207
static inline int blk_integrity_add(struct gendisk *disk)
208
{
209
	return 0;
210 211 212 213
}
static inline void blk_integrity_del(struct gendisk *disk)
{
}
214
#endif /* CONFIG_BLK_DEV_INTEGRITY */
215

216
unsigned long blk_rq_timeout(unsigned long timeout);
217
void blk_add_timer(struct request *req);
218 219

bool blk_attempt_plug_merge(struct request_queue *q, struct bio *bio,
220
		unsigned int nr_segs, struct request **same_queue_rq);
221 222
bool blk_bio_list_merge(struct request_queue *q, struct list_head *list,
			struct bio *bio, unsigned int nr_segs);
223

224
void blk_account_io_start(struct request *req);
225
void blk_account_io_done(struct request *req, u64 now);
226

227 228 229 230 231 232
/*
 * Plug flush limits
 */
#define BLK_MAX_REQUEST_COUNT	32
#define BLK_PLUG_FLUSH_SIZE	(128 * 1024)

233 234 235
/*
 * Internal elevator interface
 */
236
#define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
237

238
void blk_insert_flush(struct request *rq);
T
Tejun Heo 已提交
239

240 241
int elevator_switch_mq(struct request_queue *q,
			      struct elevator_type *new_e);
242
void __elevator_exit(struct request_queue *, struct elevator_queue *);
243
int elv_register_queue(struct request_queue *q, bool uevent);
244 245
void elv_unregister_queue(struct request_queue *q);

246 247 248
static inline void elevator_exit(struct request_queue *q,
		struct elevator_queue *e)
{
249 250
	lockdep_assert_held(&q->sysfs_lock);

251
	blk_mq_sched_free_rqs(q);
252 253 254
	__elevator_exit(q, e);
}

255 256 257 258 259 260 261 262 263 264
ssize_t part_size_show(struct device *dev, struct device_attribute *attr,
		char *buf);
ssize_t part_stat_show(struct device *dev, struct device_attribute *attr,
		char *buf);
ssize_t part_inflight_show(struct device *dev, struct device_attribute *attr,
		char *buf);
ssize_t part_fail_show(struct device *dev, struct device_attribute *attr,
		char *buf);
ssize_t part_fail_store(struct device *dev, struct device_attribute *attr,
		const char *buf, size_t count);
265 266 267 268
ssize_t part_timeout_show(struct device *, struct device_attribute *, char *);
ssize_t part_timeout_store(struct device *, struct device_attribute *,
				const char *, size_t);

269
void __blk_queue_split(struct bio **bio, unsigned int *nr_segs);
270 271
int ll_back_merge_fn(struct request *req, struct bio *bio,
		unsigned int nr_segs);
272
bool blk_attempt_req_merge(struct request_queue *q, struct request *rq,
273
				struct request *next);
274
unsigned int blk_recalc_rq_segments(struct request *rq);
275
void blk_rq_set_mixed_merge(struct request *rq);
276
bool blk_rq_merge_ok(struct request *rq, struct bio *bio);
277
enum elv_merge blk_try_merge(struct request *rq, struct bio *bio);
278

279 280
int blk_dev_init(void);

281 282 283 284
/*
 * Contribute to IO statistics IFF:
 *
 *	a) it's attached to a gendisk, and
285
 *	b) the queue had IO stats enabled when this request was started
286
 */
287
static inline bool blk_do_io_stat(struct request *rq)
288
{
289
	return rq->rq_disk && (rq->rq_flags & RQF_IO_STAT);
290 291
}

292 293 294 295 296 297 298
static inline void req_set_nomerge(struct request_queue *q, struct request *req)
{
	req->cmd_flags |= REQ_NOMERGE;
	if (req == q->last_merge)
		q->last_merge = NULL;
}

299 300 301 302 303 304 305 306 307 308
/*
 * The max size one bio can handle is UINT_MAX becasue bvec_iter.bi_size
 * is defined as 'unsigned int', meantime it has to aligned to with logical
 * block size which is the minimum accepted unit by hardware.
 */
static inline unsigned int bio_allowed_max_sectors(struct request_queue *q)
{
	return round_down(UINT_MAX, queue_logical_block_size(q)) >> 9;
}

309 310 311 312 313 314 315 316 317 318 319 320 321 322
/*
 * The max bio size which is aligned to q->limits.discard_granularity. This
 * is a hint to split large discard bio in generic block layer, then if device
 * driver needs to split the discard bio into smaller ones, their bi_size can
 * be very probably and easily aligned to discard_granularity of the device's
 * queue.
 */
static inline unsigned int bio_aligned_discard_max_sectors(
					struct request_queue *q)
{
	return round_down(UINT_MAX, q->limits.discard_granularity) >>
			SECTOR_SHIFT;
}

323 324 325 326
/*
 * Internal io_context interface
 */
void get_io_context(struct io_context *ioc);
327
struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
328 329
struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
			     gfp_t gfp_mask);
330
void ioc_clear_queue(struct request_queue *q);
331

332
int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node);
333

334 335 336 337
#ifdef CONFIG_BLK_DEV_THROTTLING_LOW
extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page);
extern ssize_t blk_throtl_sample_time_store(struct request_queue *q,
	const char *page, size_t count);
338
extern void blk_throtl_bio_endio(struct bio *bio);
339
extern void blk_throtl_stat_add(struct request *rq, u64 time);
340 341
#else
static inline void blk_throtl_bio_endio(struct bio *bio) { }
342
static inline void blk_throtl_stat_add(struct request *rq, u64 time) { }
343
#endif
344

345 346 347 348 349 350 351 352 353
void __blk_queue_bounce(struct request_queue *q, struct bio **bio);

static inline bool blk_queue_may_bounce(struct request_queue *q)
{
	return IS_ENABLED(CONFIG_BOUNCE) &&
		q->limits.bounce == BLK_BOUNCE_HIGH &&
		max_low_pfn >= max_pfn;
}

354 355
static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio)
{
356 357
	if (unlikely(blk_queue_may_bounce(q) && bio_has_data(*bio)))
		__blk_queue_bounce(q, bio);	
358 359
}

360 361 362 363 364 365
#ifdef CONFIG_BLK_CGROUP_IOLATENCY
extern int blk_iolatency_init(struct request_queue *q);
#else
static inline int blk_iolatency_init(struct request_queue *q) { return 0; }
#endif

366 367
struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp);

368 369
#ifdef CONFIG_BLK_DEV_ZONED
void blk_queue_free_zone_bitmaps(struct request_queue *q);
370
void blk_queue_clear_zone_settings(struct request_queue *q);
371 372
#else
static inline void blk_queue_free_zone_bitmaps(struct request_queue *q) {}
373
static inline void blk_queue_clear_zone_settings(struct request_queue *q) {}
374 375
#endif

376 377
int blk_alloc_ext_minor(void);
void blk_free_ext_minor(unsigned int minor);
378 379 380
#define ADDPART_FLAG_NONE	0
#define ADDPART_FLAG_RAID	1
#define ADDPART_FLAG_WHOLEDISK	2
381 382
int bdev_add_partition(struct gendisk *disk, int partno, sector_t start,
		sector_t length);
383
int bdev_del_partition(struct gendisk *disk, int partno);
384 385
int bdev_resize_partition(struct gendisk *disk, int partno, sector_t start,
		sector_t length);
386

387
int bio_add_hw_page(struct request_queue *q, struct bio *bio,
388
		struct page *page, unsigned int len, unsigned int offset,
389
		unsigned int max_sectors, bool *same_page);
390

391 392
struct request_queue *blk_alloc_queue(int node_id);

393
int disk_alloc_events(struct gendisk *disk);
394 395 396
void disk_add_events(struct gendisk *disk);
void disk_del_events(struct gendisk *disk);
void disk_release_events(struct gendisk *disk);
397 398 399
extern struct device_attribute dev_attr_events;
extern struct device_attribute dev_attr_events_async;
extern struct device_attribute dev_attr_events_poll_msecs;
400

401 402 403 404 405 406 407
static inline void bio_clear_hipri(struct bio *bio)
{
	/* can't support alloc cache if we turn off polling */
	bio_clear_flag(bio, BIO_PERCPU_CACHE);
	bio->bi_opf &= ~REQ_HIPRI;
}

408 409
extern const struct address_space_operations def_blk_aops;

410
#endif /* BLK_INTERNAL_H */