Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

dm: rename request variables to bios

Use 'bio' in the name of variables and functions that deal with
bios rather than 'request' to avoid confusion with the normal
block layer use of 'request'.

No functional changes.

Signed-off-by: Alasdair G Kergon <agk@redhat.com>

+81 -81
+3 -3
drivers/md/dm-crypt.c
··· 1637 1637 1638 1638 if (opt_params == 1 && opt_string && 1639 1639 !strcasecmp(opt_string, "allow_discards")) 1640 - ti->num_discard_requests = 1; 1640 + ti->num_discard_bios = 1; 1641 1641 else if (opt_params) { 1642 1642 ret = -EINVAL; 1643 1643 ti->error = "Invalid feature arguments"; ··· 1665 1665 goto bad; 1666 1666 } 1667 1667 1668 - ti->num_flush_requests = 1; 1668 + ti->num_flush_bios = 1; 1669 1669 ti->discard_zeroes_data_unsupported = true; 1670 1670 1671 1671 return 0; ··· 1726 1726 DMEMIT(" %llu %s %llu", (unsigned long long)cc->iv_offset, 1727 1727 cc->dev->name, (unsigned long long)cc->start); 1728 1728 1729 - if (ti->num_discard_requests) 1729 + if (ti->num_discard_bios) 1730 1730 DMEMIT(" 1 allow_discards"); 1731 1731 1732 1732 break;
+2 -2
drivers/md/dm-delay.c
··· 198 198 mutex_init(&dc->timer_lock); 199 199 atomic_set(&dc->may_delay, 1); 200 200 201 - ti->num_flush_requests = 1; 202 - ti->num_discard_requests = 1; 201 + ti->num_flush_bios = 1; 202 + ti->num_discard_bios = 1; 203 203 ti->private = dc; 204 204 return 0; 205 205
+2 -2
drivers/md/dm-flakey.c
··· 216 216 goto bad; 217 217 } 218 218 219 - ti->num_flush_requests = 1; 220 - ti->num_discard_requests = 1; 219 + ti->num_flush_bios = 1; 220 + ti->num_discard_bios = 1; 221 221 ti->per_bio_data_size = sizeof(struct per_bio_data); 222 222 ti->private = fc; 223 223 return 0;
+3 -3
drivers/md/dm-linear.c
··· 53 53 goto bad; 54 54 } 55 55 56 - ti->num_flush_requests = 1; 57 - ti->num_discard_requests = 1; 58 - ti->num_write_same_requests = 1; 56 + ti->num_flush_bios = 1; 57 + ti->num_discard_bios = 1; 58 + ti->num_write_same_bios = 1; 59 59 ti->private = lc; 60 60 return 0; 61 61
+2 -2
drivers/md/dm-mpath.c
··· 905 905 goto bad; 906 906 } 907 907 908 - ti->num_flush_requests = 1; 909 - ti->num_discard_requests = 1; 908 + ti->num_flush_bios = 1; 909 + ti->num_discard_bios = 1; 910 910 911 911 return 0; 912 912
+1 -1
drivers/md/dm-raid.c
··· 1151 1151 1152 1152 INIT_WORK(&rs->md.event_work, do_table_event); 1153 1153 ti->private = rs; 1154 - ti->num_flush_requests = 1; 1154 + ti->num_flush_bios = 1; 1155 1155 1156 1156 mutex_lock(&rs->md.reconfig_mutex); 1157 1157 ret = md_run(&rs->md);
+2 -2
drivers/md/dm-raid1.c
··· 1072 1072 if (r) 1073 1073 goto err_free_context; 1074 1074 1075 - ti->num_flush_requests = 1; 1076 - ti->num_discard_requests = 1; 1075 + ti->num_flush_bios = 1; 1076 + ti->num_discard_bios = 1; 1077 1077 ti->per_bio_data_size = sizeof(struct dm_raid1_bio_record); 1078 1078 ti->discard_zeroes_data_unsupported = true; 1079 1079
+5 -5
drivers/md/dm-snap.c
··· 1037 1037 int i; 1038 1038 int r = -EINVAL; 1039 1039 char *origin_path, *cow_path; 1040 - unsigned args_used, num_flush_requests = 1; 1040 + unsigned args_used, num_flush_bios = 1; 1041 1041 fmode_t origin_mode = FMODE_READ; 1042 1042 1043 1043 if (argc != 4) { ··· 1047 1047 } 1048 1048 1049 1049 if (dm_target_is_snapshot_merge(ti)) { 1050 - num_flush_requests = 2; 1050 + num_flush_bios = 2; 1051 1051 origin_mode = FMODE_WRITE; 1052 1052 } 1053 1053 ··· 1127 1127 spin_lock_init(&s->tracked_chunk_lock); 1128 1128 1129 1129 ti->private = s; 1130 - ti->num_flush_requests = num_flush_requests; 1130 + ti->num_flush_bios = num_flush_bios; 1131 1131 ti->per_bio_data_size = sizeof(struct dm_snap_tracked_chunk); 1132 1132 1133 1133 /* Add snapshot to the list of snapshots for this origin */ ··· 1691 1691 init_tracked_chunk(bio); 1692 1692 1693 1693 if (bio->bi_rw & REQ_FLUSH) { 1694 - if (!dm_bio_get_target_request_nr(bio)) 1694 + if (!dm_bio_get_target_bio_nr(bio)) 1695 1695 bio->bi_bdev = s->origin->bdev; 1696 1696 else 1697 1697 bio->bi_bdev = s->cow->bdev; ··· 2102 2102 } 2103 2103 2104 2104 ti->private = dev; 2105 - ti->num_flush_requests = 1; 2105 + ti->num_flush_bios = 1; 2106 2106 2107 2107 return 0; 2108 2108 }
+10 -10
drivers/md/dm-stripe.c
··· 160 160 if (r) 161 161 return r; 162 162 163 - ti->num_flush_requests = stripes; 164 - ti->num_discard_requests = stripes; 165 - ti->num_write_same_requests = stripes; 163 + ti->num_flush_bios = stripes; 164 + ti->num_discard_bios = stripes; 165 + ti->num_write_same_bios = stripes; 166 166 167 167 sc->chunk_size = chunk_size; 168 168 if (chunk_size & (chunk_size - 1)) ··· 276 276 { 277 277 struct stripe_c *sc = ti->private; 278 278 uint32_t stripe; 279 - unsigned target_request_nr; 279 + unsigned target_bio_nr; 280 280 281 281 if (bio->bi_rw & REQ_FLUSH) { 282 - target_request_nr = dm_bio_get_target_request_nr(bio); 283 - BUG_ON(target_request_nr >= sc->stripes); 284 - bio->bi_bdev = sc->stripe[target_request_nr].dev->bdev; 282 + target_bio_nr = dm_bio_get_target_bio_nr(bio); 283 + BUG_ON(target_bio_nr >= sc->stripes); 284 + bio->bi_bdev = sc->stripe[target_bio_nr].dev->bdev; 285 285 return DM_MAPIO_REMAPPED; 286 286 } 287 287 if (unlikely(bio->bi_rw & REQ_DISCARD) || 288 288 unlikely(bio->bi_rw & REQ_WRITE_SAME)) { 289 - target_request_nr = dm_bio_get_target_request_nr(bio); 290 - BUG_ON(target_request_nr >= sc->stripes); 291 - return stripe_map_range(sc, bio, target_request_nr); 289 + target_bio_nr = dm_bio_get_target_bio_nr(bio); 290 + BUG_ON(target_bio_nr >= sc->stripes); 291 + return stripe_map_range(sc, bio, target_bio_nr); 292 292 } 293 293 294 294 stripe_map_sector(sc, bio->bi_sector, &stripe, &bio->bi_sector);
+5 -5
drivers/md/dm-table.c
··· 822 822 823 823 t->highs[t->num_targets++] = tgt->begin + tgt->len - 1; 824 824 825 - if (!tgt->num_discard_requests && tgt->discards_supported) 826 - DMWARN("%s: %s: ignoring discards_supported because num_discard_requests is zero.", 825 + if (!tgt->num_discard_bios && tgt->discards_supported) 826 + DMWARN("%s: %s: ignoring discards_supported because num_discard_bios is zero.", 827 827 dm_device_name(t->md), type); 828 828 829 829 return 0; ··· 1359 1359 while (i < dm_table_get_num_targets(t)) { 1360 1360 ti = dm_table_get_target(t, i++); 1361 1361 1362 - if (!ti->num_flush_requests) 1362 + if (!ti->num_flush_bios) 1363 1363 continue; 1364 1364 1365 1365 if (ti->flush_supported) ··· 1438 1438 while (i < dm_table_get_num_targets(t)) { 1439 1439 ti = dm_table_get_target(t, i++); 1440 1440 1441 - if (!ti->num_write_same_requests) 1441 + if (!ti->num_write_same_bios) 1442 1442 return false; 1443 1443 1444 1444 if (!ti->type->iterate_devices || ··· 1656 1656 while (i < dm_table_get_num_targets(t)) { 1657 1657 ti = dm_table_get_target(t, i++); 1658 1658 1659 - if (!ti->num_discard_requests) 1659 + if (!ti->num_discard_bios) 1660 1660 continue; 1661 1661 1662 1662 if (ti->discards_supported)
+1 -1
drivers/md/dm-target.c
··· 116 116 /* 117 117 * Return error for discards instead of -EOPNOTSUPP 118 118 */ 119 - tt->num_discard_requests = 1; 119 + tt->num_discard_bios = 1; 120 120 121 121 return 0; 122 122 }
+6 -6
drivers/md/dm-thin.c
··· 1944 1944 pt->data_dev = data_dev; 1945 1945 pt->low_water_blocks = low_water_blocks; 1946 1946 pt->adjusted_pf = pt->requested_pf = pf; 1947 - ti->num_flush_requests = 1; 1947 + ti->num_flush_bios = 1; 1948 1948 1949 1949 /* 1950 1950 * Only need to enable discards if the pool should pass ··· 1952 1952 * processing will cause mappings to be removed from the btree. 1953 1953 */ 1954 1954 if (pf.discard_enabled && pf.discard_passdown) { 1955 - ti->num_discard_requests = 1; 1955 + ti->num_discard_bios = 1; 1956 1956 1957 1957 /* 1958 1958 * Setting 'discards_supported' circumvents the normal ··· 2593 2593 if (r) 2594 2594 goto bad_thin_open; 2595 2595 2596 - ti->num_flush_requests = 1; 2596 + ti->num_flush_bios = 1; 2597 2597 ti->flush_supported = true; 2598 2598 ti->per_bio_data_size = sizeof(struct dm_thin_endio_hook); 2599 2599 2600 2600 /* In case the pool supports discards, pass them on. */ 2601 2601 if (tc->pool->pf.discard_enabled) { 2602 2602 ti->discards_supported = true; 2603 - ti->num_discard_requests = 1; 2603 + ti->num_discard_bios = 1; 2604 2604 ti->discard_zeroes_data_unsupported = true; 2605 - /* Discard requests must be split on a block boundary */ 2606 - ti->split_discard_requests = true; 2605 + /* Discard bios must be split on a block boundary */ 2606 + ti->split_discard_bios = true; 2607 2607 } 2608 2608 2609 2609 dm_put(pool_md);
+1 -1
drivers/md/dm-zero.c
··· 25 25 /* 26 26 * Silently drop discards, avoiding -EOPNOTSUPP. 27 27 */ 28 - ti->num_discard_requests = 1; 28 + ti->num_discard_bios = 1; 29 29 30 30 return 0; 31 31 }
+23 -23
drivers/md/dm.c
··· 1103 1103 1104 1104 static struct dm_target_io *alloc_tio(struct clone_info *ci, 1105 1105 struct dm_target *ti, int nr_iovecs, 1106 - unsigned target_request_nr) 1106 + unsigned target_bio_nr) 1107 1107 { 1108 1108 struct dm_target_io *tio; 1109 1109 struct bio *clone; ··· 1114 1114 tio->io = ci->io; 1115 1115 tio->ti = ti; 1116 1116 memset(&tio->info, 0, sizeof(tio->info)); 1117 - tio->target_request_nr = target_request_nr; 1117 + tio->target_bio_nr = target_bio_nr; 1118 1118 1119 1119 return tio; 1120 1120 } 1121 1121 1122 1122 static void __issue_target_request(struct clone_info *ci, struct dm_target *ti, 1123 - unsigned request_nr, sector_t len) 1123 + unsigned target_bio_nr, sector_t len) 1124 1124 { 1125 - struct dm_target_io *tio = alloc_tio(ci, ti, ci->bio->bi_max_vecs, request_nr); 1125 + struct dm_target_io *tio = alloc_tio(ci, ti, ci->bio->bi_max_vecs, target_bio_nr); 1126 1126 struct bio *clone = &tio->clone; 1127 1127 1128 1128 /* ··· 1137 1137 __map_bio(tio); 1138 1138 } 1139 1139 1140 - static void __issue_target_requests(struct clone_info *ci, struct dm_target *ti, 1141 - unsigned num_requests, sector_t len) 1140 + static void __issue_target_bios(struct clone_info *ci, struct dm_target *ti, 1141 + unsigned num_bios, sector_t len) 1142 1142 { 1143 - unsigned request_nr; 1143 + unsigned target_bio_nr; 1144 1144 1145 - for (request_nr = 0; request_nr < num_requests; request_nr++) 1146 - __issue_target_request(ci, ti, request_nr, len); 1145 + for (target_bio_nr = 0; target_bio_nr < num_bios; target_bio_nr++) 1146 + __issue_target_request(ci, ti, target_bio_nr, len); 1147 1147 } 1148 1148 1149 1149 static int __clone_and_map_empty_flush(struct clone_info *ci) ··· 1153 1153 1154 1154 BUG_ON(bio_has_data(ci->bio)); 1155 1155 while ((ti = dm_table_get_target(ci->map, target_nr++))) 1156 - __issue_target_requests(ci, ti, ti->num_flush_requests, 0); 1156 + __issue_target_bios(ci, ti, ti->num_flush_bios, 0); 1157 1157 1158 1158 return 0; 1159 1159 } ··· 1173 1173 ci->sector_count = 0; 1174 1174 } 1175 1175 1176 - typedef unsigned (*get_num_requests_fn)(struct dm_target *ti); 1176 + typedef unsigned (*get_num_bios_fn)(struct dm_target *ti); 1177 1177 1178 - static unsigned get_num_discard_requests(struct dm_target *ti) 1178 + static unsigned get_num_discard_bios(struct dm_target *ti) 1179 1179 { 1180 - return ti->num_discard_requests; 1180 + return ti->num_discard_bios; 1181 1181 } 1182 1182 1183 - static unsigned get_num_write_same_requests(struct dm_target *ti) 1183 + static unsigned get_num_write_same_bios(struct dm_target *ti) 1184 1184 { 1185 - return ti->num_write_same_requests; 1185 + return ti->num_write_same_bios; 1186 1186 } 1187 1187 1188 1188 typedef bool (*is_split_required_fn)(struct dm_target *ti); 1189 1189 1190 1190 static bool is_split_required_for_discard(struct dm_target *ti) 1191 1191 { 1192 - return ti->split_discard_requests; 1192 + return ti->split_discard_bios; 1193 1193 } 1194 1194 1195 1195 static int __clone_and_map_changing_extent_only(struct clone_info *ci, 1196 - get_num_requests_fn get_num_requests, 1196 + get_num_bios_fn get_num_bios, 1197 1197 is_split_required_fn is_split_required) 1198 1198 { 1199 1199 struct dm_target *ti; 1200 1200 sector_t len; 1201 - unsigned num_requests; 1201 + unsigned num_bios; 1202 1202 1203 1203 do { 1204 1204 ti = dm_table_find_target(ci->map, ci->sector); ··· 1211 1211 * reconfiguration might also have changed that since the 1212 1212 * check was performed. 1213 1213 */ 1214 - num_requests = get_num_requests ? get_num_requests(ti) : 0; 1215 - if (!num_requests) 1214 + num_bios = get_num_bios ? get_num_bios(ti) : 0; 1215 + if (!num_bios) 1216 1216 return -EOPNOTSUPP; 1217 1217 1218 1218 if (is_split_required && !is_split_required(ti)) ··· 1220 1220 else 1221 1221 len = min(ci->sector_count, max_io_len(ci->sector, ti)); 1222 1222 1223 - __issue_target_requests(ci, ti, num_requests, len); 1223 + __issue_target_bios(ci, ti, num_bios, len); 1224 1224 1225 1225 ci->sector += len; 1226 1226 } while (ci->sector_count -= len); ··· 1230 1230 1231 1231 static int __clone_and_map_discard(struct clone_info *ci) 1232 1232 { 1233 - return __clone_and_map_changing_extent_only(ci, get_num_discard_requests, 1233 + return __clone_and_map_changing_extent_only(ci, get_num_discard_bios, 1234 1234 is_split_required_for_discard); 1235 1235 } 1236 1236 1237 1237 static int __clone_and_map_write_same(struct clone_info *ci) 1238 1238 { 1239 - return __clone_and_map_changing_extent_only(ci, get_num_write_same_requests, NULL); 1239 + return __clone_and_map_changing_extent_only(ci, get_num_write_same_bios, NULL); 1240 1240 } 1241 1241 1242 1242 static int __clone_and_map(struct clone_info *ci)
+15 -15
include/linux/device-mapper.h
··· 187 187 uint32_t max_io_len; 188 188 189 189 /* 190 - * A number of zero-length barrier requests that will be submitted 190 + * A number of zero-length barrier bios that will be submitted 191 191 * to the target for the purpose of flushing cache. 192 192 * 193 - * The request number can be accessed with dm_bio_get_target_request_nr. 194 - * It is a responsibility of the target driver to remap these requests 193 + * The bio number can be accessed with dm_bio_get_target_bio_nr. 194 + * It is a responsibility of the target driver to remap these bios 195 195 * to the real underlying devices. 196 196 */ 197 - unsigned num_flush_requests; 197 + unsigned num_flush_bios; 198 198 199 199 /* 200 - * The number of discard requests that will be submitted to the target. 201 - * The request number can be accessed with dm_bio_get_target_request_nr. 200 + * The number of discard bios that will be submitted to the target. 201 + * The bio number can be accessed with dm_bio_get_target_bio_nr. 202 202 */ 203 - unsigned num_discard_requests; 203 + unsigned num_discard_bios; 204 204 205 205 /* 206 - * The number of WRITE SAME requests that will be submitted to the target. 207 - * The request number can be accessed with dm_bio_get_target_request_nr. 206 + * The number of WRITE SAME bios that will be submitted to the target. 207 + * The bio number can be accessed with dm_bio_get_target_bio_nr. 208 208 */ 209 - unsigned num_write_same_requests; 209 + unsigned num_write_same_bios; 210 210 211 211 /* 212 212 * The minimum number of extra bytes allocated in each bio for the ··· 233 233 bool discards_supported:1; 234 234 235 235 /* 236 - * Set if the target required discard request to be split 236 + * Set if the target required discard bios to be split 237 237 * on max_io_len boundary. 238 238 */ 239 - bool split_discard_requests:1; 239 + bool split_discard_bios:1; 240 240 241 241 /* 242 242 * Set if this target does not return zeroes on discarded blocks. ··· 261 261 struct dm_io *io; 262 262 struct dm_target *ti; 263 263 union map_info info; 264 - unsigned target_request_nr; 264 + unsigned target_bio_nr; 265 265 struct bio clone; 266 266 }; 267 267 ··· 275 275 return (struct bio *)((char *)data + data_size + offsetof(struct dm_target_io, clone)); 276 276 } 277 277 278 - static inline unsigned dm_bio_get_target_request_nr(const struct bio *bio) 278 + static inline unsigned dm_bio_get_target_bio_nr(const struct bio *bio) 279 279 { 280 - return container_of(bio, struct dm_target_io, clone)->target_request_nr; 280 + return container_of(bio, struct dm_target_io, clone)->target_bio_nr; 281 281 } 282 282 283 283 int dm_register_target(struct target_type *t);