Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * blk-integrity.c - Block layer data integrity extensions
4 *
5 * Copyright (C) 2007, 2008 Oracle Corporation
6 * Written by: Martin K. Petersen <martin.petersen@oracle.com>
7 */
8
9#include <linux/blk-integrity.h>
10#include <linux/backing-dev.h>
11#include <linux/mempool.h>
12#include <linux/bio.h>
13#include <linux/scatterlist.h>
14#include <linux/export.h>
15#include <linux/slab.h>
16#include <linux/t10-pi.h>
17
18#include "blk.h"
19
20/**
21 * blk_rq_count_integrity_sg - Count number of integrity scatterlist elements
22 * @q: request queue
23 * @bio: bio with integrity metadata attached
24 *
25 * Description: Returns the number of elements required in a
26 * scatterlist corresponding to the integrity metadata in a bio.
27 */
28int blk_rq_count_integrity_sg(struct request_queue *q, struct bio *bio)
29{
30 struct bio_vec iv, ivprv = { NULL };
31 unsigned int segments = 0;
32 unsigned int seg_size = 0;
33 struct bvec_iter iter;
34 int prev = 0;
35
36 bio_for_each_integrity_vec(iv, bio, iter) {
37
38 if (prev) {
39 if (!biovec_phys_mergeable(q, &ivprv, &iv))
40 goto new_segment;
41 if (seg_size + iv.bv_len > queue_max_segment_size(q))
42 goto new_segment;
43
44 seg_size += iv.bv_len;
45 } else {
46new_segment:
47 segments++;
48 seg_size = iv.bv_len;
49 }
50
51 prev = 1;
52 ivprv = iv;
53 }
54
55 return segments;
56}
57
58int blk_get_meta_cap(struct block_device *bdev, unsigned int cmd,
59 struct logical_block_metadata_cap __user *argp)
60{
61 struct blk_integrity *bi;
62 struct logical_block_metadata_cap meta_cap = {};
63 size_t usize = _IOC_SIZE(cmd);
64
65 if (!extensible_ioctl_valid(cmd, FS_IOC_GETLBMD_CAP, LBMD_SIZE_VER0))
66 return -ENOIOCTLCMD;
67
68 bi = blk_get_integrity(bdev->bd_disk);
69 if (!bi)
70 goto out;
71
72 if (bi->flags & BLK_INTEGRITY_DEVICE_CAPABLE)
73 meta_cap.lbmd_flags |= LBMD_PI_CAP_INTEGRITY;
74 if (bi->flags & BLK_INTEGRITY_REF_TAG)
75 meta_cap.lbmd_flags |= LBMD_PI_CAP_REFTAG;
76 meta_cap.lbmd_interval = 1 << bi->interval_exp;
77 meta_cap.lbmd_size = bi->metadata_size;
78 meta_cap.lbmd_pi_size = bi->pi_tuple_size;
79 meta_cap.lbmd_pi_offset = bi->pi_offset;
80 meta_cap.lbmd_opaque_size = bi->metadata_size - bi->pi_tuple_size;
81 if (meta_cap.lbmd_opaque_size && !bi->pi_offset)
82 meta_cap.lbmd_opaque_offset = bi->pi_tuple_size;
83
84 switch (bi->csum_type) {
85 case BLK_INTEGRITY_CSUM_NONE:
86 meta_cap.lbmd_guard_tag_type = LBMD_PI_CSUM_NONE;
87 break;
88 case BLK_INTEGRITY_CSUM_IP:
89 meta_cap.lbmd_guard_tag_type = LBMD_PI_CSUM_IP;
90 break;
91 case BLK_INTEGRITY_CSUM_CRC:
92 meta_cap.lbmd_guard_tag_type = LBMD_PI_CSUM_CRC16_T10DIF;
93 break;
94 case BLK_INTEGRITY_CSUM_CRC64:
95 meta_cap.lbmd_guard_tag_type = LBMD_PI_CSUM_CRC64_NVME;
96 break;
97 }
98
99 if (bi->csum_type != BLK_INTEGRITY_CSUM_NONE)
100 meta_cap.lbmd_app_tag_size = 2;
101
102 if (bi->flags & BLK_INTEGRITY_REF_TAG) {
103 switch (bi->csum_type) {
104 case BLK_INTEGRITY_CSUM_CRC64:
105 meta_cap.lbmd_ref_tag_size =
106 sizeof_field(struct crc64_pi_tuple, ref_tag);
107 break;
108 case BLK_INTEGRITY_CSUM_CRC:
109 case BLK_INTEGRITY_CSUM_IP:
110 meta_cap.lbmd_ref_tag_size =
111 sizeof_field(struct t10_pi_tuple, ref_tag);
112 break;
113 default:
114 break;
115 }
116 }
117
118out:
119 return copy_struct_to_user(argp, usize, &meta_cap, sizeof(meta_cap),
120 NULL);
121}
122
123int blk_rq_integrity_map_user(struct request *rq, void __user *ubuf,
124 ssize_t bytes)
125{
126 int ret;
127 struct iov_iter iter;
128
129 iov_iter_ubuf(&iter, rq_data_dir(rq), ubuf, bytes);
130 ret = bio_integrity_map_user(rq->bio, &iter);
131 if (ret)
132 return ret;
133
134 rq->nr_integrity_segments = blk_rq_count_integrity_sg(rq->q, rq->bio);
135 rq->cmd_flags |= REQ_INTEGRITY;
136 return 0;
137}
138EXPORT_SYMBOL_GPL(blk_rq_integrity_map_user);
139
140bool blk_integrity_merge_rq(struct request_queue *q, struct request *req,
141 struct request *next)
142{
143 struct bio_integrity_payload *bip, *bip_next;
144
145 if (blk_integrity_rq(req) == 0 && blk_integrity_rq(next) == 0)
146 return true;
147
148 if (blk_integrity_rq(req) == 0 || blk_integrity_rq(next) == 0)
149 return false;
150
151 bip = bio_integrity(req->bio);
152 bip_next = bio_integrity(next->bio);
153 if (bip->bip_flags != bip_next->bip_flags)
154 return false;
155
156 if (bip->bip_flags & BIP_CHECK_APPTAG &&
157 bip->app_tag != bip_next->app_tag)
158 return false;
159
160 if (req->nr_integrity_segments + next->nr_integrity_segments >
161 q->limits.max_integrity_segments)
162 return false;
163
164 if (integrity_req_gap_back_merge(req, next->bio))
165 return false;
166
167 return true;
168}
169
170bool blk_integrity_merge_bio(struct request_queue *q, struct request *req,
171 struct bio *bio)
172{
173 struct bio_integrity_payload *bip, *bip_bio = bio_integrity(bio);
174 int nr_integrity_segs;
175
176 if (blk_integrity_rq(req) == 0 && bip_bio == NULL)
177 return true;
178
179 if (blk_integrity_rq(req) == 0 || bip_bio == NULL)
180 return false;
181
182 bip = bio_integrity(req->bio);
183 if (bip->bip_flags != bip_bio->bip_flags)
184 return false;
185
186 if (bip->bip_flags & BIP_CHECK_APPTAG &&
187 bip->app_tag != bip_bio->app_tag)
188 return false;
189
190 nr_integrity_segs = blk_rq_count_integrity_sg(q, bio);
191 if (req->nr_integrity_segments + nr_integrity_segs >
192 q->limits.max_integrity_segments)
193 return false;
194
195 return true;
196}
197
198static inline struct blk_integrity *dev_to_bi(struct device *dev)
199{
200 return &dev_to_disk(dev)->queue->limits.integrity;
201}
202
203const char *blk_integrity_profile_name(struct blk_integrity *bi)
204{
205 switch (bi->csum_type) {
206 case BLK_INTEGRITY_CSUM_IP:
207 if (bi->flags & BLK_INTEGRITY_REF_TAG)
208 return "T10-DIF-TYPE1-IP";
209 return "T10-DIF-TYPE3-IP";
210 case BLK_INTEGRITY_CSUM_CRC:
211 if (bi->flags & BLK_INTEGRITY_REF_TAG)
212 return "T10-DIF-TYPE1-CRC";
213 return "T10-DIF-TYPE3-CRC";
214 case BLK_INTEGRITY_CSUM_CRC64:
215 if (bi->flags & BLK_INTEGRITY_REF_TAG)
216 return "EXT-DIF-TYPE1-CRC64";
217 return "EXT-DIF-TYPE3-CRC64";
218 case BLK_INTEGRITY_CSUM_NONE:
219 break;
220 }
221
222 return "nop";
223}
224EXPORT_SYMBOL_GPL(blk_integrity_profile_name);
225
226static ssize_t flag_store(struct device *dev, const char *page, size_t count,
227 unsigned char flag)
228{
229 struct request_queue *q = dev_to_disk(dev)->queue;
230 struct queue_limits lim;
231 unsigned long val;
232 int err;
233
234 err = kstrtoul(page, 10, &val);
235 if (err)
236 return err;
237
238 /* note that the flags are inverted vs the values in the sysfs files */
239 lim = queue_limits_start_update(q);
240 if (val)
241 lim.integrity.flags &= ~flag;
242 else
243 lim.integrity.flags |= flag;
244
245 err = queue_limits_commit_update_frozen(q, &lim);
246 if (err)
247 return err;
248 return count;
249}
250
251static ssize_t flag_show(struct device *dev, char *page, unsigned char flag)
252{
253 struct blk_integrity *bi = dev_to_bi(dev);
254
255 return sysfs_emit(page, "%d\n", !(bi->flags & flag));
256}
257
258static ssize_t format_show(struct device *dev, struct device_attribute *attr,
259 char *page)
260{
261 struct blk_integrity *bi = dev_to_bi(dev);
262
263 if (!bi->metadata_size)
264 return sysfs_emit(page, "none\n");
265 return sysfs_emit(page, "%s\n", blk_integrity_profile_name(bi));
266}
267
268static ssize_t tag_size_show(struct device *dev, struct device_attribute *attr,
269 char *page)
270{
271 struct blk_integrity *bi = dev_to_bi(dev);
272
273 return sysfs_emit(page, "%u\n", bi->tag_size);
274}
275
276static ssize_t protection_interval_bytes_show(struct device *dev,
277 struct device_attribute *attr,
278 char *page)
279{
280 struct blk_integrity *bi = dev_to_bi(dev);
281
282 return sysfs_emit(page, "%u\n",
283 bi->interval_exp ? 1 << bi->interval_exp : 0);
284}
285
286static ssize_t read_verify_store(struct device *dev,
287 struct device_attribute *attr,
288 const char *page, size_t count)
289{
290 return flag_store(dev, page, count, BLK_INTEGRITY_NOVERIFY);
291}
292
293static ssize_t read_verify_show(struct device *dev,
294 struct device_attribute *attr, char *page)
295{
296 return flag_show(dev, page, BLK_INTEGRITY_NOVERIFY);
297}
298
299static ssize_t write_generate_store(struct device *dev,
300 struct device_attribute *attr,
301 const char *page, size_t count)
302{
303 return flag_store(dev, page, count, BLK_INTEGRITY_NOGENERATE);
304}
305
306static ssize_t write_generate_show(struct device *dev,
307 struct device_attribute *attr, char *page)
308{
309 return flag_show(dev, page, BLK_INTEGRITY_NOGENERATE);
310}
311
312static ssize_t device_is_integrity_capable_show(struct device *dev,
313 struct device_attribute *attr,
314 char *page)
315{
316 struct blk_integrity *bi = dev_to_bi(dev);
317
318 return sysfs_emit(page, "%u\n",
319 !!(bi->flags & BLK_INTEGRITY_DEVICE_CAPABLE));
320}
321
322static DEVICE_ATTR_RO(format);
323static DEVICE_ATTR_RO(tag_size);
324static DEVICE_ATTR_RO(protection_interval_bytes);
325static DEVICE_ATTR_RW(read_verify);
326static DEVICE_ATTR_RW(write_generate);
327static DEVICE_ATTR_RO(device_is_integrity_capable);
328
329static struct attribute *integrity_attrs[] = {
330 &dev_attr_format.attr,
331 &dev_attr_tag_size.attr,
332 &dev_attr_protection_interval_bytes.attr,
333 &dev_attr_read_verify.attr,
334 &dev_attr_write_generate.attr,
335 &dev_attr_device_is_integrity_capable.attr,
336 NULL
337};
338
339const struct attribute_group blk_integrity_attr_group = {
340 .name = "integrity",
341 .attrs = integrity_attrs,
342};