blob: 9db80c92096a8bb56bf88535eb93105f1e973822 [file] [log] [blame]
yuezonghe824eb0c2024-06-27 02:32:26 -07001/*
2 * Internal header file for device mapper
3 *
4 * Copyright (C) 2001, 2002 Sistina Software
5 * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved.
6 *
7 * This file is released under the LGPL.
8 */
9
10#ifndef DM_INTERNAL_H
11#define DM_INTERNAL_H
12
13#include <linux/fs.h>
14#include <linux/device-mapper.h>
15#include <linux/list.h>
16#include <linux/blkdev.h>
17#include <linux/hdreg.h>
18#include <linux/completion.h>
19#include <linux/kobject.h>
20
21/*
22 * Suspend feature flags
23 */
24#define DM_SUSPEND_LOCKFS_FLAG (1 << 0)
25#define DM_SUSPEND_NOFLUSH_FLAG (1 << 1)
26
27/*
28 * Type of table and mapped_device's mempool
29 */
30#define DM_TYPE_NONE 0
31#define DM_TYPE_BIO_BASED 1
32#define DM_TYPE_REQUEST_BASED 2
33
34/*
35 * List of devices that a metadevice uses and should open/close.
36 */
37struct dm_dev_internal {
38 struct list_head list;
39 atomic_t count;
40 struct dm_dev dm_dev;
41};
42
43struct dm_table;
44struct dm_md_mempools;
45
46/*-----------------------------------------------------------------
47 * Internal table functions.
48 *---------------------------------------------------------------*/
49void dm_table_destroy(struct dm_table *t);
50void dm_table_event_callback(struct dm_table *t,
51 void (*fn)(void *), void *context);
52struct dm_target *dm_table_get_target(struct dm_table *t, unsigned int index);
53struct dm_target *dm_table_find_target(struct dm_table *t, sector_t sector);
54int dm_calculate_queue_limits(struct dm_table *table,
55 struct queue_limits *limits);
56void dm_table_set_restrictions(struct dm_table *t, struct request_queue *q,
57 struct queue_limits *limits);
58struct list_head *dm_table_get_devices(struct dm_table *t);
59void dm_table_presuspend_targets(struct dm_table *t);
60void dm_table_postsuspend_targets(struct dm_table *t);
61int dm_table_resume_targets(struct dm_table *t);
62int dm_table_any_congested(struct dm_table *t, int bdi_bits);
63int dm_table_any_busy_target(struct dm_table *t);
64unsigned dm_table_get_type(struct dm_table *t);
65struct target_type *dm_table_get_immutable_target_type(struct dm_table *t);
66bool dm_table_request_based(struct dm_table *t);
67bool dm_table_supports_discards(struct dm_table *t);
68int dm_table_alloc_md_mempools(struct dm_table *t);
69void dm_table_free_md_mempools(struct dm_table *t);
70struct dm_md_mempools *dm_table_get_md_mempools(struct dm_table *t);
71
72int dm_queue_merge_is_compulsory(struct request_queue *q);
73
74void dm_lock_md_type(struct mapped_device *md);
75void dm_unlock_md_type(struct mapped_device *md);
76void dm_set_md_type(struct mapped_device *md, unsigned type);
77unsigned dm_get_md_type(struct mapped_device *md);
78struct target_type *dm_get_immutable_target_type(struct mapped_device *md);
79
80int dm_setup_md_queue(struct mapped_device *md);
81
82/*
83 * To check the return value from dm_table_find_target().
84 */
85#define dm_target_is_valid(t) ((t)->table)
86
87/*
88 * To check whether the target type is request-based or not (bio-based).
89 */
90#define dm_target_request_based(t) ((t)->type->map_rq != NULL)
91
92/*-----------------------------------------------------------------
93 * A registry of target types.
94 *---------------------------------------------------------------*/
95int dm_target_init(void);
96void dm_target_exit(void);
97struct target_type *dm_get_target_type(const char *name);
98void dm_put_target_type(struct target_type *tt);
99int dm_target_iterate(void (*iter_func)(struct target_type *tt,
100 void *param), void *param);
101
102int dm_split_args(int *argc, char ***argvp, char *input);
103
104/*
105 * Is this mapped_device being deleted?
106 */
107int dm_deleting_md(struct mapped_device *md);
108
109/*
110 * Is this mapped_device suspended?
111 */
112int dm_suspended_md(struct mapped_device *md);
113
114/*
115 * The device-mapper can be driven through one of two interfaces;
116 * ioctl or filesystem, depending which patch you have applied.
117 */
118int dm_interface_init(void);
119void dm_interface_exit(void);
120
121/*
122 * sysfs interface
123 */
124struct dm_kobject_holder {
125 struct kobject kobj;
126 struct completion completion;
127};
128
129static inline struct completion *dm_get_completion_from_kobject(struct kobject *kobj)
130{
131 return &container_of(kobj, struct dm_kobject_holder, kobj)->completion;
132}
133
134int dm_sysfs_init(struct mapped_device *md);
135void dm_sysfs_exit(struct mapped_device *md);
136struct kobject *dm_kobject(struct mapped_device *md);
137struct mapped_device *dm_get_from_kobject(struct kobject *kobj);
138
139/*
140 * The kobject helper
141 */
142void dm_kobject_release(struct kobject *kobj);
143
144/*
145 * Targets for linear and striped mappings
146 */
147int dm_linear_init(void);
148void dm_linear_exit(void);
149
150int dm_stripe_init(void);
151void dm_stripe_exit(void);
152
153/*
154 * mapped_device operations
155 */
156void dm_destroy(struct mapped_device *md);
157void dm_destroy_immediate(struct mapped_device *md);
158int dm_open_count(struct mapped_device *md);
159int dm_lock_for_deletion(struct mapped_device *md);
160
161int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action,
162 unsigned cookie);
163
164int dm_io_init(void);
165void dm_io_exit(void);
166
167int dm_kcopyd_init(void);
168void dm_kcopyd_exit(void);
169
170/*
171 * Mempool operations
172 */
173struct dm_md_mempools *dm_alloc_md_mempools(unsigned type, unsigned integrity);
174void dm_free_md_mempools(struct dm_md_mempools *pools);
175
176#endif