• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright (C) 2001-2003 Sistina Software (UK) Limited.
3  *
4  * This file is released under the GPL.
5  */
6 
7 #include "dm.h"
8 #include <linux/module.h>
9 #include <linux/init.h>
10 #include <linux/blkdev.h>
11 #include <linux/bio.h>
12 #include <linux/dax.h>
13 #include <linux/slab.h>
14 #include <linux/device-mapper.h>
15 
16 #define DM_MSG_PREFIX "linear"
17 
18 /*
19  * Linear: maps a linear range of a device.
20  */
21 struct linear_c {
22 	struct dm_dev *dev;
23 	sector_t start;
24 };
25 
26 /*
27  * Construct a linear mapping: <dev_path> <offset>
28  */
linear_ctr(struct dm_target * ti,unsigned int argc,char ** argv)29 static int linear_ctr(struct dm_target *ti, unsigned int argc, char **argv)
30 {
31 	struct linear_c *lc;
32 	unsigned long long tmp;
33 	char dummy;
34 	int ret;
35 
36 	if (argc != 2) {
37 		ti->error = "Invalid argument count";
38 		return -EINVAL;
39 	}
40 
41 	lc = kmalloc(sizeof(*lc), GFP_KERNEL);
42 	if (lc == NULL) {
43 		ti->error = "Cannot allocate linear context";
44 		return -ENOMEM;
45 	}
46 
47 	ret = -EINVAL;
48 	if (sscanf(argv[1], "%llu%c", &tmp, &dummy) != 1 || tmp != (sector_t)tmp) {
49 		ti->error = "Invalid device sector";
50 		goto bad;
51 	}
52 	lc->start = tmp;
53 
54 	ret = dm_get_device(ti, argv[0], dm_table_get_mode(ti->table), &lc->dev);
55 	if (ret) {
56 		ti->error = "Device lookup failed";
57 		goto bad;
58 	}
59 
60 	ti->num_flush_bios = 1;
61 	ti->num_discard_bios = 1;
62 	ti->num_secure_erase_bios = 1;
63 	ti->num_write_same_bios = 1;
64 	ti->num_write_zeroes_bios = 1;
65 	ti->may_passthrough_inline_crypto = true;
66 	ti->private = lc;
67 	return 0;
68 
69       bad:
70 	kfree(lc);
71 	return ret;
72 }
73 
linear_dtr(struct dm_target * ti)74 static void linear_dtr(struct dm_target *ti)
75 {
76 	struct linear_c *lc = (struct linear_c *) ti->private;
77 
78 	dm_put_device(ti, lc->dev);
79 	kfree(lc);
80 }
81 
linear_map_sector(struct dm_target * ti,sector_t bi_sector)82 static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector)
83 {
84 	struct linear_c *lc = ti->private;
85 
86 	return lc->start + dm_target_offset(ti, bi_sector);
87 }
88 
linear_map_bio(struct dm_target * ti,struct bio * bio)89 static void linear_map_bio(struct dm_target *ti, struct bio *bio)
90 {
91 	struct linear_c *lc = ti->private;
92 
93 	bio_set_dev(bio, lc->dev->bdev);
94 	if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET)
95 		bio->bi_iter.bi_sector =
96 			linear_map_sector(ti, bio->bi_iter.bi_sector);
97 }
98 
linear_map(struct dm_target * ti,struct bio * bio)99 static int linear_map(struct dm_target *ti, struct bio *bio)
100 {
101 	linear_map_bio(ti, bio);
102 
103 	return DM_MAPIO_REMAPPED;
104 }
105 
linear_status(struct dm_target * ti,status_type_t type,unsigned status_flags,char * result,unsigned maxlen)106 static void linear_status(struct dm_target *ti, status_type_t type,
107 			  unsigned status_flags, char *result, unsigned maxlen)
108 {
109 	struct linear_c *lc = (struct linear_c *) ti->private;
110 
111 	switch (type) {
112 	case STATUSTYPE_INFO:
113 		result[0] = '\0';
114 		break;
115 
116 	case STATUSTYPE_TABLE:
117 		snprintf(result, maxlen, "%s %llu", lc->dev->name,
118 				(unsigned long long)lc->start);
119 		break;
120 	}
121 }
122 
linear_prepare_ioctl(struct dm_target * ti,struct block_device ** bdev)123 static int linear_prepare_ioctl(struct dm_target *ti, struct block_device **bdev)
124 {
125 	struct linear_c *lc = (struct linear_c *) ti->private;
126 	struct dm_dev *dev = lc->dev;
127 
128 	*bdev = dev->bdev;
129 
130 	/*
131 	 * Only pass ioctls through if the device sizes match exactly.
132 	 */
133 	if (lc->start ||
134 	    ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT)
135 		return 1;
136 	return 0;
137 }
138 
139 #ifdef CONFIG_BLK_DEV_ZONED
linear_report_zones(struct dm_target * ti,struct dm_report_zones_args * args,unsigned int nr_zones)140 static int linear_report_zones(struct dm_target *ti,
141 		struct dm_report_zones_args *args, unsigned int nr_zones)
142 {
143 	struct linear_c *lc = ti->private;
144 	sector_t sector = linear_map_sector(ti, args->next_sector);
145 
146 	args->start = lc->start;
147 	return blkdev_report_zones(lc->dev->bdev, sector, nr_zones,
148 				   dm_report_zones_cb, args);
149 }
150 #endif
151 
linear_iterate_devices(struct dm_target * ti,iterate_devices_callout_fn fn,void * data)152 static int linear_iterate_devices(struct dm_target *ti,
153 				  iterate_devices_callout_fn fn, void *data)
154 {
155 	struct linear_c *lc = ti->private;
156 
157 	return fn(ti, lc->dev, lc->start, ti->len, data);
158 }
159 
160 #if IS_ENABLED(CONFIG_DAX_DRIVER)
linear_dax_direct_access(struct dm_target * ti,pgoff_t pgoff,long nr_pages,void ** kaddr,pfn_t * pfn)161 static long linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
162 		long nr_pages, void **kaddr, pfn_t *pfn)
163 {
164 	long ret;
165 	struct linear_c *lc = ti->private;
166 	struct block_device *bdev = lc->dev->bdev;
167 	struct dax_device *dax_dev = lc->dev->dax_dev;
168 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
169 
170 	dev_sector = linear_map_sector(ti, sector);
171 	ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
172 	if (ret)
173 		return ret;
174 	return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
175 }
176 
linear_dax_copy_from_iter(struct dm_target * ti,pgoff_t pgoff,void * addr,size_t bytes,struct iov_iter * i)177 static size_t linear_dax_copy_from_iter(struct dm_target *ti, pgoff_t pgoff,
178 		void *addr, size_t bytes, struct iov_iter *i)
179 {
180 	struct linear_c *lc = ti->private;
181 	struct block_device *bdev = lc->dev->bdev;
182 	struct dax_device *dax_dev = lc->dev->dax_dev;
183 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
184 
185 	dev_sector = linear_map_sector(ti, sector);
186 	if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
187 		return 0;
188 	return dax_copy_from_iter(dax_dev, pgoff, addr, bytes, i);
189 }
190 
linear_dax_copy_to_iter(struct dm_target * ti,pgoff_t pgoff,void * addr,size_t bytes,struct iov_iter * i)191 static size_t linear_dax_copy_to_iter(struct dm_target *ti, pgoff_t pgoff,
192 		void *addr, size_t bytes, struct iov_iter *i)
193 {
194 	struct linear_c *lc = ti->private;
195 	struct block_device *bdev = lc->dev->bdev;
196 	struct dax_device *dax_dev = lc->dev->dax_dev;
197 	sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
198 
199 	dev_sector = linear_map_sector(ti, sector);
200 	if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
201 		return 0;
202 	return dax_copy_to_iter(dax_dev, pgoff, addr, bytes, i);
203 }
204 
205 #else
206 #define linear_dax_direct_access NULL
207 #define linear_dax_copy_from_iter NULL
208 #define linear_dax_copy_to_iter NULL
209 #endif
210 
211 static struct target_type linear_target = {
212 	.name   = "linear",
213 	.version = {1, 4, 0},
214 #ifdef CONFIG_BLK_DEV_ZONED
215 	.features = DM_TARGET_PASSES_INTEGRITY | DM_TARGET_ZONED_HM,
216 	.report_zones = linear_report_zones,
217 #else
218 	.features = DM_TARGET_PASSES_INTEGRITY,
219 #endif
220 	.module = THIS_MODULE,
221 	.ctr    = linear_ctr,
222 	.dtr    = linear_dtr,
223 	.map    = linear_map,
224 	.status = linear_status,
225 	.prepare_ioctl = linear_prepare_ioctl,
226 	.iterate_devices = linear_iterate_devices,
227 	.direct_access = linear_dax_direct_access,
228 	.dax_copy_from_iter = linear_dax_copy_from_iter,
229 	.dax_copy_to_iter = linear_dax_copy_to_iter,
230 };
231 
dm_linear_init(void)232 int __init dm_linear_init(void)
233 {
234 	int r = dm_register_target(&linear_target);
235 
236 	if (r < 0)
237 		DMERR("register failed %d", r);
238 
239 	return r;
240 }
241 
dm_linear_exit(void)242 void dm_linear_exit(void)
243 {
244 	dm_unregister_target(&linear_target);
245 }
246