blob: e6fd31b03c38a4d4f8684d03f95c36820ae0cebb [file] [log] [blame]
rjw1f884582022-01-06 17:20:42 +08001/*
2 * Copyright (C) 2001-2003 Sistina Software (UK) Limited.
3 *
4 * This file is released under the GPL.
5 */
6
7#include "dm.h"
8#include <linux/module.h>
9#include <linux/init.h>
10#include <linux/blkdev.h>
11#include <linux/bio.h>
12#include <linux/dax.h>
13#include <linux/slab.h>
14#include <linux/device-mapper.h>
15
16#define DM_MSG_PREFIX "linear"
17
18/*
19 * Linear: maps a linear range of a device.
20 */
21struct linear_c {
22 struct dm_dev *dev;
23 sector_t start;
24};
25
26/*
27 * Construct a linear mapping: <dev_path> <offset>
28 */
29int dm_linear_ctr(struct dm_target *ti, unsigned int argc, char **argv)
30{
31 struct linear_c *lc;
32 unsigned long long tmp;
33 char dummy;
34 int ret;
35
36 if (argc != 2) {
37 ti->error = "Invalid argument count";
38 return -EINVAL;
39 }
40
41 lc = kmalloc(sizeof(*lc), GFP_KERNEL);
42 if (lc == NULL) {
43 ti->error = "Cannot allocate linear context";
44 return -ENOMEM;
45 }
46
47 ret = -EINVAL;
48 if (sscanf(argv[1], "%llu%c", &tmp, &dummy) != 1) {
49 ti->error = "Invalid device sector";
50 goto bad;
51 }
52 lc->start = tmp;
53
54 ret = dm_get_device(ti, argv[0], dm_table_get_mode(ti->table), &lc->dev);
55 if (ret) {
56 ti->error = "Device lookup failed";
57 goto bad;
58 }
59
60 ti->num_flush_bios = 1;
61 ti->num_discard_bios = 1;
62 ti->num_write_same_bios = 1;
63 ti->num_write_zeroes_bios = 1;
64 ti->private = lc;
65 return 0;
66
67 bad:
68 kfree(lc);
69 return ret;
70}
71
72void dm_linear_dtr(struct dm_target *ti)
73{
74 struct linear_c *lc = (struct linear_c *) ti->private;
75
76 dm_put_device(ti, lc->dev);
77 kfree(lc);
78}
79
80static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector)
81{
82 struct linear_c *lc = ti->private;
83
84 return lc->start + dm_target_offset(ti, bi_sector);
85}
86
87static void linear_map_bio(struct dm_target *ti, struct bio *bio)
88{
89 struct linear_c *lc = ti->private;
90
91 bio_set_dev(bio, lc->dev->bdev);
92 if (bio_sectors(bio) || bio_op(bio) == REQ_OP_ZONE_RESET)
93 bio->bi_iter.bi_sector =
94 linear_map_sector(ti, bio->bi_iter.bi_sector);
95}
96
97int dm_linear_map(struct dm_target *ti, struct bio *bio)
98{
99 linear_map_bio(ti, bio);
100
101 return DM_MAPIO_REMAPPED;
102}
103
104int dm_linear_end_io(struct dm_target *ti, struct bio *bio,
105 blk_status_t *error)
106{
107 struct linear_c *lc = ti->private;
108
109 if (!*error && bio_op(bio) == REQ_OP_ZONE_REPORT)
110 dm_remap_zone_report(ti, bio, lc->start);
111
112 return DM_ENDIO_DONE;
113}
114EXPORT_SYMBOL_GPL(dm_linear_end_io);
115
116void dm_linear_status(struct dm_target *ti, status_type_t type,
117 unsigned status_flags, char *result, unsigned maxlen)
118{
119 struct linear_c *lc = (struct linear_c *) ti->private;
120
121 switch (type) {
122 case STATUSTYPE_INFO:
123 result[0] = '\0';
124 break;
125
126 case STATUSTYPE_TABLE:
127 snprintf(result, maxlen, "%s %llu", lc->dev->name,
128 (unsigned long long)lc->start);
129 break;
130 }
131}
132
133int dm_linear_prepare_ioctl(struct dm_target *ti,
134 struct block_device **bdev, fmode_t *mode)
135{
136 struct linear_c *lc = (struct linear_c *) ti->private;
137 struct dm_dev *dev = lc->dev;
138
139 *bdev = dev->bdev;
140
141 /*
142 * Only pass ioctls through if the device sizes match exactly.
143 */
144 if (lc->start ||
145 ti->len != i_size_read(dev->bdev->bd_inode) >> SECTOR_SHIFT)
146 return 1;
147 return 0;
148}
149
150int dm_linear_iterate_devices(struct dm_target *ti,
151 iterate_devices_callout_fn fn, void *data)
152{
153 struct linear_c *lc = ti->private;
154
155 return fn(ti, lc->dev, lc->start, ti->len, data);
156}
157
158long dm_linear_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
159 long nr_pages, void **kaddr, pfn_t *pfn)
160{
161 long ret;
162 struct linear_c *lc = ti->private;
163 struct block_device *bdev = lc->dev->bdev;
164 struct dax_device *dax_dev = lc->dev->dax_dev;
165 sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
166
167 dev_sector = linear_map_sector(ti, sector);
168 ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff);
169 if (ret)
170 return ret;
171 return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn);
172}
173EXPORT_SYMBOL_GPL(dm_linear_dax_direct_access);
174
175size_t dm_linear_dax_copy_from_iter(struct dm_target *ti, pgoff_t pgoff,
176 void *addr, size_t bytes, struct iov_iter *i)
177{
178 struct linear_c *lc = ti->private;
179 struct block_device *bdev = lc->dev->bdev;
180 struct dax_device *dax_dev = lc->dev->dax_dev;
181 sector_t dev_sector, sector = pgoff * PAGE_SECTORS;
182
183 dev_sector = linear_map_sector(ti, sector);
184 if (bdev_dax_pgoff(bdev, dev_sector, ALIGN(bytes, PAGE_SIZE), &pgoff))
185 return 0;
186 return dax_copy_from_iter(dax_dev, pgoff, addr, bytes, i);
187}
188EXPORT_SYMBOL_GPL(dm_linear_dax_copy_from_iter);
189
190static struct target_type linear_target = {
191 .name = "linear",
192 .version = {1, 4, 0},
193 .features = DM_TARGET_PASSES_INTEGRITY | DM_TARGET_ZONED_HM,
194 .module = THIS_MODULE,
195 .ctr = dm_linear_ctr,
196 .dtr = dm_linear_dtr,
197 .map = dm_linear_map,
198 .status = dm_linear_status,
199 .end_io = dm_linear_end_io,
200 .prepare_ioctl = dm_linear_prepare_ioctl,
201 .iterate_devices = dm_linear_iterate_devices,
202 .direct_access = dm_linear_dax_direct_access,
203 .dax_copy_from_iter = dm_linear_dax_copy_from_iter,
204};
205
206int __init dm_linear_init(void)
207{
208 int r = dm_register_target(&linear_target);
209
210 if (r < 0)
211 DMERR("register failed %d", r);
212
213 return r;
214}
215
216void dm_linear_exit(void)
217{
218 dm_unregister_target(&linear_target);
219}