Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v5.17-rc3 170 lines 5.4 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef __BLK_NULL_BLK_H 3#define __BLK_NULL_BLK_H 4 5#undef pr_fmt 6#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 7 8#include <linux/blkdev.h> 9#include <linux/slab.h> 10#include <linux/blk-mq.h> 11#include <linux/hrtimer.h> 12#include <linux/configfs.h> 13#include <linux/badblocks.h> 14#include <linux/fault-inject.h> 15#include <linux/spinlock.h> 16#include <linux/mutex.h> 17 18struct nullb_cmd { 19 struct request *rq; 20 struct bio *bio; 21 unsigned int tag; 22 blk_status_t error; 23 struct nullb_queue *nq; 24 struct hrtimer timer; 25 bool fake_timeout; 26}; 27 28struct nullb_queue { 29 unsigned long *tag_map; 30 wait_queue_head_t wait; 31 unsigned int queue_depth; 32 struct nullb_device *dev; 33 unsigned int requeue_selection; 34 35 struct list_head poll_list; 36 spinlock_t poll_lock; 37 38 struct nullb_cmd *cmds; 39}; 40 41struct nullb_zone { 42 /* 43 * Zone lock to prevent concurrent modification of a zone write 44 * pointer position and condition: with memory backing, a write 45 * command execution may sleep on memory allocation. For this case, 46 * use mutex as the zone lock. Otherwise, use the spinlock for 47 * locking the zone. 48 */ 49 union { 50 spinlock_t spinlock; 51 struct mutex mutex; 52 }; 53 enum blk_zone_type type; 54 enum blk_zone_cond cond; 55 sector_t start; 56 sector_t wp; 57 unsigned int len; 58 unsigned int capacity; 59}; 60 61struct nullb_device { 62 struct nullb *nullb; 63 struct config_item item; 64 struct radix_tree_root data; /* data stored in the disk */ 65 struct radix_tree_root cache; /* disk cache data */ 66 unsigned long flags; /* device flags */ 67 unsigned int curr_cache; 68 struct badblocks badblocks; 69 70 unsigned int nr_zones; 71 unsigned int nr_zones_imp_open; 72 unsigned int nr_zones_exp_open; 73 unsigned int nr_zones_closed; 74 unsigned int imp_close_zone_no; 75 struct nullb_zone *zones; 76 sector_t zone_size_sects; 77 bool need_zone_res_mgmt; 78 spinlock_t zone_res_lock; 79 80 unsigned long size; /* device size in MB */ 81 unsigned long completion_nsec; /* time in ns to complete a request */ 82 unsigned long cache_size; /* disk cache size in MB */ 83 unsigned long zone_size; /* zone size in MB if device is zoned */ 84 unsigned long zone_capacity; /* zone capacity in MB if device is zoned */ 85 unsigned int zone_nr_conv; /* number of conventional zones */ 86 unsigned int zone_max_open; /* max number of open zones */ 87 unsigned int zone_max_active; /* max number of active zones */ 88 unsigned int submit_queues; /* number of submission queues */ 89 unsigned int prev_submit_queues; /* number of submission queues before change */ 90 unsigned int poll_queues; /* number of IOPOLL submission queues */ 91 unsigned int prev_poll_queues; /* number of IOPOLL submission queues before change */ 92 unsigned int home_node; /* home node for the device */ 93 unsigned int queue_mode; /* block interface */ 94 unsigned int blocksize; /* block size */ 95 unsigned int max_sectors; /* Max sectors per command */ 96 unsigned int irqmode; /* IRQ completion handler */ 97 unsigned int hw_queue_depth; /* queue depth */ 98 unsigned int index; /* index of the disk, only valid with a disk */ 99 unsigned int mbps; /* Bandwidth throttle cap (in MB/s) */ 100 bool blocking; /* blocking blk-mq device */ 101 bool use_per_node_hctx; /* use per-node allocation for hardware context */ 102 bool power; /* power on/off the device */ 103 bool memory_backed; /* if data is stored in memory */ 104 bool discard; /* if support discard */ 105 bool zoned; /* if device is zoned */ 106 bool virt_boundary; /* virtual boundary on/off for the device */ 107}; 108 109struct nullb { 110 struct nullb_device *dev; 111 struct list_head list; 112 unsigned int index; 113 struct request_queue *q; 114 struct gendisk *disk; 115 struct blk_mq_tag_set *tag_set; 116 struct blk_mq_tag_set __tag_set; 117 unsigned int queue_depth; 118 atomic_long_t cur_bytes; 119 struct hrtimer bw_timer; 120 unsigned long cache_flush_pos; 121 spinlock_t lock; 122 123 struct nullb_queue *queues; 124 unsigned int nr_queues; 125 char disk_name[DISK_NAME_LEN]; 126}; 127 128blk_status_t null_handle_discard(struct nullb_device *dev, sector_t sector, 129 sector_t nr_sectors); 130blk_status_t null_process_cmd(struct nullb_cmd *cmd, 131 enum req_opf op, sector_t sector, 132 unsigned int nr_sectors); 133 134#ifdef CONFIG_BLK_DEV_ZONED 135int null_init_zoned_dev(struct nullb_device *dev, struct request_queue *q); 136int null_register_zoned_dev(struct nullb *nullb); 137void null_free_zoned_dev(struct nullb_device *dev); 138int null_report_zones(struct gendisk *disk, sector_t sector, 139 unsigned int nr_zones, report_zones_cb cb, void *data); 140blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, 141 enum req_opf op, sector_t sector, 142 sector_t nr_sectors); 143size_t null_zone_valid_read_len(struct nullb *nullb, 144 sector_t sector, unsigned int len); 145#else 146static inline int null_init_zoned_dev(struct nullb_device *dev, 147 struct request_queue *q) 148{ 149 pr_err("CONFIG_BLK_DEV_ZONED not enabled\n"); 150 return -EINVAL; 151} 152static inline int null_register_zoned_dev(struct nullb *nullb) 153{ 154 return -ENODEV; 155} 156static inline void null_free_zoned_dev(struct nullb_device *dev) {} 157static inline blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, 158 enum req_opf op, sector_t sector, sector_t nr_sectors) 159{ 160 return BLK_STS_NOTSUPP; 161} 162static inline size_t null_zone_valid_read_len(struct nullb *nullb, 163 sector_t sector, 164 unsigned int len) 165{ 166 return len; 167} 168#define null_report_zones NULL 169#endif /* CONFIG_BLK_DEV_ZONED */ 170#endif /* __NULL_BLK_H */