1 /*
2 * Copyright (C) 2001-2003 Sistina Software (UK) Limited.
3 *
4 * This file is released under the GPL.
5 */
6
7 #include "dm.h"
8 #include <linux/module.h>
9 #include <linux/init.h>
10 #include <linux/blkdev.h>
11 #include <linux/bio.h>
12 #include <linux/dax.h>
13 #include <linux/slab.h>
14 #include <linux/device-mapper.h>
15
16 #define DM_MSG_PREFIX "linear"
17
18 /*
19 * Linear: maps a linear range of a device.
20 */
21 struct linear_c {
22 struct dm_dev *dev;
23 sector_t start;
24 };
25
26 /*
27 * Construct a linear mapping: <dev_path> <offset>
28 */
linear_ctr(struct dm_target * ti,unsigned int argc,char ** argv)29 static int linear_ctr(struct dm_target *ti, unsigned int argc, char **argv)
30 {
31 struct linear_c *lc;
32 unsigned long long tmp;
33 char dummy;
34 int ret;
35
36 if (argc != 2) {
37 ti->error = "Invalid argument count";
38 return -EINVAL;
39 }
40
41 lc = kmalloc(sizeof(*lc), GFP_KERNEL);
42 if (lc == NULL) {
43 ti->error = "Cannot allocate linear context";
44 return -ENOMEM;
45 }
46
47 ret = -EINVAL;
48 if (sscanf(argv[1], "%llu%c", &tmp, &dummy) != 1 || tmp != (sector_t)tmp) {
49 ti->error = "Invalid device sector";
50 goto bad;
51 }
52 lc->start = tmp;
53
54 ret = dm_get_device(ti, argv[0], dm_table_get_mode(ti->table), &lc->dev);
55 if (ret) {
56 ti->error = "Device lookup failed";
57 goto bad;
58 }
59
60 ti->num_flush_bios = 1;
61 ti->num_discard_bios = 1;
62 ti->num_secure_erase_bios = 1;
63 ti->num_write_same_bios = 1;
64 ti->num_write_zeroes_bios = 1;
65 ti->may_passthrough_inline_crypto = true;
66 ti->private = lc;
67 return 0;
68
69 bad:
70 kfree(lc);
71 return ret;
72 }
73
linear_dtr(struct dm_target * ti)74 static void linear_dtr(struct dm_target *ti)
75 {
76 struct linear_c *lc = (struct linear_c *) ti->private;
77
78 dm_put_device(ti, lc->dev);
79 kfree(lc);
80 }
81
linear_map_sector(struct dm_target * ti,sector_t bi_sector)82 static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector)
83 {
84 struct linear_c *lc = ti->private;
85
86 return lc->start + dm_target_offset(ti, bi_sector);
87 }
88
linear_map_bio(struct dm_target * ti,struct bio * bio)89 static void linear_map_bio(struct dm_target *ti, struct bio *bio)
90 {
91 struct linear_c *lc = ti->private;
92
93 bio_set_dev(bio, lc->dev->bdev);
94 if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET)
95 bio->bi_iter.bi_sector =
96 linear_map_sector(ti, bio->bi_iter.bi_sector);
97 }
98
linear_map(struct dm_target * ti,struct bio * bio)99 static int linear_map(struct dm_target *ti, struct bio *bio)
100 {
101 linear_map_bio(ti, bio);
102
103 return DM_MAPIO_REMAPPED;
104 }
105
linear_status(struct dm_target * ti,status_type_t type,unsigned status_flags,char * result,unsigned maxlen)106 static void linear_status(struct dm_target *ti, status_type_t type,
107 unsigned status_flags, char *result, unsigned maxlen)
108 {
109 struct linear_c *lc = (struct linear_c *) ti->private;
110
111 switch (type) {
112 case STATUSTYPE_INFO:
113 result[0] = '\0';
114 break;
115
116 case STATUSTYPE_TABLE:
117 snprintf(result, maxlen, "%s %llu", lc->dev->name,
118 (unsigned long long)lc->start);
119 break;
120 }
121 }
122
linear_prepare_ioctl(struct dm_target * ti,struct block_device ** bdev)123 static int linear_prepare_ioctl(struct dm_target *ti, struct block_device **bdev)
124 {
125 struct linear_c *lc = (struct linear_c *) ti->private;
126 struct dm_dev *dev = lc->dev;
127
128 *bdev = dev->bdev;
129
130 /*
131 * Only pass ioctls through if the device sizes match exactly.
132 */
133 if (lc->start ||
134 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT)
135 return 1;
136 return 0;
137 }
138
139 #ifdef CONFIG_BLK_DEV_ZONED
linear_report_zones(struct dm_target * ti,struct dm_report_zones_args * args,unsigned int nr_zones)140 static int linear_report_zones(struct dm_target *ti,
141 struct dm_report_zones_args *args, unsigned int nr_zones)
142 {
143 struct linear_c *lc = ti->private;
144 sector_t sector = linear_map_sector(ti, args->next_sector);
145
146 args->start = lc->start;
147 return blkdev_report_zones(lc->dev->bdev, sector, nr_zones,
148 dm_report_zones_cb, args);
149 }
150 #endif
151
linear_iterate_devices(struct dm_target * ti,iterate_devices_callout_fn fn,void * data)152 static int linear_iterate_devices(struct dm_target *ti,
153 iterate_devices_callout_fn fn, void *data)
154 {
155 struct linear_c *lc = ti->private;
156
157 return fn(ti, lc->dev, lc->start, ti->len, data);
158 }
159
160 #if IS_ENABLED(CONFIG_DAX_DRIVER)
linear_dax_direct_access(struct dm_target * ti,pgoff_t pgoff,long nr_pages,void ** kaddr,pfn_t * pfn)161 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
162 long nr_pages, void **kaddr, pfn_t *pfn)
163 {
164 long ret;
165 struct linear_c *lc = ti->private;
166 struct block_device *bdev = lc->dev->bdev;
167 struct dax_device *dax_dev = lc->dev->dax_dev;
168 sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
169
170 dev_sector = linear_map_sector(ti, sector);
171 ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
172 if (ret)
173 return ret;
174 return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
175 }
176
linear_dax_copy_from_iter(struct dm_target * ti,pgoff_t pgoff,void * addr,size_t bytes,struct iov_iter * i)177 static size_t linear_dax_copy_from_iter(struct dm_target *ti, pgoff_t pgoff,
178 void *addr, size_t bytes, struct iov_iter *i)
179 {
180 struct linear_c *lc = ti->private;
181 struct block_device *bdev = lc->dev->bdev;
182 struct dax_device *dax_dev = lc->dev->dax_dev;
183 sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
184
185 dev_sector = linear_map_sector(ti, sector);
186 if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
187 return 0;
188 return dax_copy_from_iter(dax_dev, pgoff, addr, bytes, i);
189 }
190
linear_dax_copy_to_iter(struct dm_target * ti,pgoff_t pgoff,void * addr,size_t bytes,struct iov_iter * i)191 static size_t linear_dax_copy_to_iter(struct dm_target *ti, pgoff_t pgoff,
192 void *addr, size_t bytes, struct iov_iter *i)
193 {
194 struct linear_c *lc = ti->private;
195 struct block_device *bdev = lc->dev->bdev;
196 struct dax_device *dax_dev = lc->dev->dax_dev;
197 sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
198
199 dev_sector = linear_map_sector(ti, sector);
200 if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
201 return 0;
202 return dax_copy_to_iter(dax_dev, pgoff, addr, bytes, i);
203 }
204
205 #else
206 #define linear_dax_direct_access NULL
207 #define linear_dax_copy_from_iter NULL
208 #define linear_dax_copy_to_iter NULL
209 #endif
210
211 static struct target_type linear_target = {
212 .name = "linear",
213 .version = {1, 4, 0},
214 #ifdef CONFIG_BLK_DEV_ZONED
215 .features = DM_TARGET_PASSES_INTEGRITY | DM_TARGET_ZONED_HM,
216 .report_zones = linear_report_zones,
217 #else
218 .features = DM_TARGET_PASSES_INTEGRITY,
219 #endif
220 .module = THIS_MODULE,
221 .ctr = linear_ctr,
222 .dtr = linear_dtr,
223 .map = linear_map,
224 .status = linear_status,
225 .prepare_ioctl = linear_prepare_ioctl,
226 .iterate_devices = linear_iterate_devices,
227 .direct_access = linear_dax_direct_access,
228 .dax_copy_from_iter = linear_dax_copy_from_iter,
229 .dax_copy_to_iter = linear_dax_copy_to_iter,
230 };
231
dm_linear_init(void)232 int __init dm_linear_init(void)
233 {
234 int r = dm_register_target(&linear_target);
235
236 if (r < 0)
237 DMERR("register failed %d", r);
238
239 return r;
240 }
241
dm_linear_exit(void)242 void dm_linear_exit(void)
243 {
244 dm_unregister_target(&linear_target);
245 }
246