blob: a308f7a7e57704787c9fe816aae5440b7ea37289 [file] [log] [blame]
rjw1f884582022-01-06 17:20:42 +08001/*
2 * Copyright (C) 2008 Red Hat, Inc., Eric Paris <eparis@redhat.com>
3 *
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2, or (at your option)
7 * any later version.
8 *
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 *
14 * You should have received a copy of the GNU General Public License
15 * along with this program; see the file COPYING. If not, write to
16 * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
17 */
18
19#include <linux/dcache.h>
20#include <linux/fs.h>
21#include <linux/gfp.h>
22#include <linux/init.h>
23#include <linux/module.h>
24#include <linux/mount.h>
25#include <linux/srcu.h>
26
27#include <linux/fsnotify_backend.h>
28#include "fsnotify.h"
29
30/*
31 * Clear all of the marks on an inode when it is being evicted from core
32 */
33void __fsnotify_inode_delete(struct inode *inode)
34{
35 fsnotify_clear_marks_by_inode(inode);
36}
37EXPORT_SYMBOL_GPL(__fsnotify_inode_delete);
38
39void __fsnotify_vfsmount_delete(struct vfsmount *mnt)
40{
41 fsnotify_clear_marks_by_mount(mnt);
42}
43
44/**
45 * fsnotify_unmount_inodes - an sb is unmounting. handle any watched inodes.
46 * @sb: superblock being unmounted.
47 *
48 * Called during unmount with no locks held, so needs to be safe against
49 * concurrent modifiers. We temporarily drop sb->s_inode_list_lock and CAN block.
50 */
51void fsnotify_unmount_inodes(struct super_block *sb)
52{
53 struct inode *inode, *iput_inode = NULL;
54
55 spin_lock(&sb->s_inode_list_lock);
56 list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
57 /*
58 * We cannot __iget() an inode in state I_FREEING,
59 * I_WILL_FREE, or I_NEW which is fine because by that point
60 * the inode cannot have any associated watches.
61 */
62 spin_lock(&inode->i_lock);
63 if (inode->i_state & (I_FREEING|I_WILL_FREE|I_NEW)) {
64 spin_unlock(&inode->i_lock);
65 continue;
66 }
67
68 /*
69 * If i_count is zero, the inode cannot have any watches and
70 * doing an __iget/iput with MS_ACTIVE clear would actually
71 * evict all inodes with zero i_count from icache which is
72 * unnecessarily violent and may in fact be illegal to do.
73 */
74 if (!atomic_read(&inode->i_count)) {
75 spin_unlock(&inode->i_lock);
76 continue;
77 }
78
79 __iget(inode);
80 spin_unlock(&inode->i_lock);
81 spin_unlock(&sb->s_inode_list_lock);
82
83 if (iput_inode)
84 iput(iput_inode);
85
86 /* for each watch, send FS_UNMOUNT and then remove it */
87 fsnotify(inode, FS_UNMOUNT, inode, FSNOTIFY_EVENT_INODE, NULL, 0);
88
89 fsnotify_inode_delete(inode);
90
91 iput_inode = inode;
92
93 cond_resched();
94 spin_lock(&sb->s_inode_list_lock);
95 }
96 spin_unlock(&sb->s_inode_list_lock);
97
98 if (iput_inode)
99 iput(iput_inode);
100}
101
102/*
103 * Given an inode, first check if we care what happens to our children. Inotify
104 * and dnotify both tell their parents about events. If we care about any event
105 * on a child we run all of our children and set a dentry flag saying that the
106 * parent cares. Thus when an event happens on a child it can quickly tell if
107 * if there is a need to find a parent and send the event to the parent.
108 */
109void __fsnotify_update_child_dentry_flags(struct inode *inode)
110{
111 struct dentry *alias;
112 int watched;
113
114 if (!S_ISDIR(inode->i_mode))
115 return;
116
117 /* determine if the children should tell inode about their events */
118 watched = fsnotify_inode_watches_children(inode);
119
120 spin_lock(&inode->i_lock);
121 /* run all of the dentries associated with this inode. Since this is a
122 * directory, there damn well better only be one item on this list */
123 hlist_for_each_entry(alias, &inode->i_dentry, d_u.d_alias) {
124 struct dentry *child;
125
126 /* run all of the children of the original inode and fix their
127 * d_flags to indicate parental interest (their parent is the
128 * original inode) */
129 spin_lock(&alias->d_lock);
130 list_for_each_entry(child, &alias->d_subdirs, d_child) {
131 if (!child->d_inode)
132 continue;
133
134 spin_lock_nested(&child->d_lock, DENTRY_D_LOCK_NESTED);
135 if (watched)
136 child->d_flags |= DCACHE_FSNOTIFY_PARENT_WATCHED;
137 else
138 child->d_flags &= ~DCACHE_FSNOTIFY_PARENT_WATCHED;
139 spin_unlock(&child->d_lock);
140 }
141 spin_unlock(&alias->d_lock);
142 }
143 spin_unlock(&inode->i_lock);
144}
145
146/* Notify this dentry's parent about a child's events. */
147int __fsnotify_parent(const struct path *path, struct dentry *dentry, __u32 mask)
148{
149 struct dentry *parent;
150 struct inode *p_inode;
151 int ret = 0;
152
153 if (!dentry)
154 dentry = path->dentry;
155
156 if (!(dentry->d_flags & DCACHE_FSNOTIFY_PARENT_WATCHED))
157 return 0;
158
159 parent = dget_parent(dentry);
160 p_inode = parent->d_inode;
161
162 if (unlikely(!fsnotify_inode_watches_children(p_inode))) {
163 __fsnotify_update_child_dentry_flags(p_inode);
164 } else if (p_inode->i_fsnotify_mask & mask & ~FS_EVENT_ON_CHILD) {
165 struct name_snapshot name;
166
167 /* we are notifying a parent so come up with the new mask which
168 * specifies these are events which came from a child. */
169 mask |= FS_EVENT_ON_CHILD;
170
171 take_dentry_name_snapshot(&name, dentry);
172 if (path)
173 ret = fsnotify(p_inode, mask, path, FSNOTIFY_EVENT_PATH,
174 name.name, 0);
175 else
176 ret = fsnotify(p_inode, mask, dentry->d_inode, FSNOTIFY_EVENT_INODE,
177 name.name, 0);
178 release_dentry_name_snapshot(&name);
179 }
180
181 dput(parent);
182
183 return ret;
184}
185EXPORT_SYMBOL_GPL(__fsnotify_parent);
186
187static int send_to_group(struct inode *to_tell,
188 struct fsnotify_mark *inode_mark,
189 struct fsnotify_mark *vfsmount_mark,
190 __u32 mask, const void *data,
191 int data_is, u32 cookie,
192 const unsigned char *file_name,
193 struct fsnotify_iter_info *iter_info)
194{
195 struct fsnotify_group *group = NULL;
196 __u32 test_mask = (mask & ~FS_EVENT_ON_CHILD);
197 __u32 marks_mask = 0;
198 __u32 marks_ignored_mask = 0;
199
200 if (unlikely(!inode_mark && !vfsmount_mark)) {
201 BUG();
202 return 0;
203 }
204
205 /* clear ignored on inode modification */
206 if (mask & FS_MODIFY) {
207 if (inode_mark &&
208 !(inode_mark->flags & FSNOTIFY_MARK_FLAG_IGNORED_SURV_MODIFY))
209 inode_mark->ignored_mask = 0;
210 if (vfsmount_mark &&
211 !(vfsmount_mark->flags & FSNOTIFY_MARK_FLAG_IGNORED_SURV_MODIFY))
212 vfsmount_mark->ignored_mask = 0;
213 }
214
215 /* does the inode mark tell us to do something? */
216 if (inode_mark) {
217 group = inode_mark->group;
218 marks_mask |= inode_mark->mask;
219 marks_ignored_mask |= inode_mark->ignored_mask;
220 }
221
222 /* does the vfsmount_mark tell us to do something? */
223 if (vfsmount_mark) {
224 group = vfsmount_mark->group;
225 marks_mask |= vfsmount_mark->mask;
226 marks_ignored_mask |= vfsmount_mark->ignored_mask;
227 }
228
229 pr_debug("%s: group=%p to_tell=%p mask=%x inode_mark=%p"
230 " vfsmount_mark=%p marks_mask=%x marks_ignored_mask=%x"
231 " data=%p data_is=%d cookie=%d\n",
232 __func__, group, to_tell, mask, inode_mark, vfsmount_mark,
233 marks_mask, marks_ignored_mask, data,
234 data_is, cookie);
235
236 if (!(test_mask & marks_mask & ~marks_ignored_mask))
237 return 0;
238
239 return group->ops->handle_event(group, to_tell, inode_mark,
240 vfsmount_mark, mask, data, data_is,
241 file_name, cookie, iter_info);
242}
243
244/*
245 * This is the main call to fsnotify. The VFS calls into hook specific functions
246 * in linux/fsnotify.h. Those functions then in turn call here. Here will call
247 * out to all of the registered fsnotify_group. Those groups can then use the
248 * notification event in whatever means they feel necessary.
249 */
250int fsnotify(struct inode *to_tell, __u32 mask, const void *data, int data_is,
251 const unsigned char *file_name, u32 cookie)
252{
253 struct hlist_node *inode_node = NULL, *vfsmount_node = NULL;
254 struct fsnotify_mark *inode_mark = NULL, *vfsmount_mark = NULL;
255 struct fsnotify_group *inode_group, *vfsmount_group;
256 struct fsnotify_mark_connector *inode_conn, *vfsmount_conn;
257 struct fsnotify_iter_info iter_info;
258 struct mount *mnt;
259 int ret = 0;
260 /* global tests shouldn't care about events on child only the specific event */
261 __u32 test_mask = (mask & ~FS_EVENT_ON_CHILD);
262
263 if (data_is == FSNOTIFY_EVENT_PATH)
264 mnt = real_mount(((const struct path *)data)->mnt);
265 else
266 mnt = NULL;
267
268 /* An event "on child" is not intended for a mount mark */
269 if (mask & FS_EVENT_ON_CHILD)
270 mnt = NULL;
271
272 /*
273 * Optimization: srcu_read_lock() has a memory barrier which can
274 * be expensive. It protects walking the *_fsnotify_marks lists.
275 * However, if we do not walk the lists, we do not have to do
276 * SRCU because we have no references to any objects and do not
277 * need SRCU to keep them "alive".
278 */
279 if (!to_tell->i_fsnotify_marks &&
280 (!mnt || !mnt->mnt_fsnotify_marks))
281 return 0;
282 /*
283 * if this is a modify event we may need to clear the ignored masks
284 * otherwise return if neither the inode nor the vfsmount care about
285 * this type of event.
286 */
287 if (!(mask & FS_MODIFY) &&
288 !(test_mask & to_tell->i_fsnotify_mask) &&
289 !(mnt && test_mask & mnt->mnt_fsnotify_mask))
290 return 0;
291
292 iter_info.srcu_idx = srcu_read_lock(&fsnotify_mark_srcu);
293
294 inode_conn = srcu_dereference(to_tell->i_fsnotify_marks,
295 &fsnotify_mark_srcu);
296 if (inode_conn)
297 inode_node = srcu_dereference(inode_conn->list.first,
298 &fsnotify_mark_srcu);
299
300 if (mnt) {
301 inode_conn = srcu_dereference(to_tell->i_fsnotify_marks,
302 &fsnotify_mark_srcu);
303 if (inode_conn)
304 inode_node = srcu_dereference(inode_conn->list.first,
305 &fsnotify_mark_srcu);
306 vfsmount_conn = srcu_dereference(mnt->mnt_fsnotify_marks,
307 &fsnotify_mark_srcu);
308 if (vfsmount_conn)
309 vfsmount_node = srcu_dereference(
310 vfsmount_conn->list.first,
311 &fsnotify_mark_srcu);
312 }
313
314 /*
315 * We need to merge inode & vfsmount mark lists so that inode mark
316 * ignore masks are properly reflected for mount mark notifications.
317 * That's why this traversal is so complicated...
318 */
319 while (inode_node || vfsmount_node) {
320 inode_group = NULL;
321 inode_mark = NULL;
322 vfsmount_group = NULL;
323 vfsmount_mark = NULL;
324
325 if (inode_node) {
326 inode_mark = hlist_entry(srcu_dereference(inode_node, &fsnotify_mark_srcu),
327 struct fsnotify_mark, obj_list);
328 inode_group = inode_mark->group;
329 }
330
331 if (vfsmount_node) {
332 vfsmount_mark = hlist_entry(srcu_dereference(vfsmount_node, &fsnotify_mark_srcu),
333 struct fsnotify_mark, obj_list);
334 vfsmount_group = vfsmount_mark->group;
335 }
336 /*
337 * Need to protect both marks against freeing so that we can
338 * continue iteration from this place, regardless of which mark
339 * we actually happen to send an event for.
340 */
341 iter_info.inode_mark = inode_mark;
342 iter_info.vfsmount_mark = vfsmount_mark;
343
344 if (inode_group && vfsmount_group) {
345 int cmp = fsnotify_compare_groups(inode_group,
346 vfsmount_group);
347 if (cmp > 0) {
348 inode_group = NULL;
349 inode_mark = NULL;
350 } else if (cmp < 0) {
351 vfsmount_group = NULL;
352 vfsmount_mark = NULL;
353 }
354 }
355
356 ret = send_to_group(to_tell, inode_mark, vfsmount_mark, mask,
357 data, data_is, cookie, file_name,
358 &iter_info);
359
360 if (ret && (mask & ALL_FSNOTIFY_PERM_EVENTS))
361 goto out;
362
363 if (inode_group)
364 inode_node = srcu_dereference(inode_node->next,
365 &fsnotify_mark_srcu);
366 if (vfsmount_group)
367 vfsmount_node = srcu_dereference(vfsmount_node->next,
368 &fsnotify_mark_srcu);
369 }
370 ret = 0;
371out:
372 srcu_read_unlock(&fsnotify_mark_srcu, iter_info.srcu_idx);
373
374 return ret;
375}
376EXPORT_SYMBOL_GPL(fsnotify);
377
378extern struct kmem_cache *fsnotify_mark_connector_cachep;
379
380static __init int fsnotify_init(void)
381{
382 int ret;
383
384 BUG_ON(hweight32(ALL_FSNOTIFY_EVENTS) != 23);
385
386 ret = init_srcu_struct(&fsnotify_mark_srcu);
387 if (ret)
388 panic("initializing fsnotify_mark_srcu");
389
390 fsnotify_mark_connector_cachep = KMEM_CACHE(fsnotify_mark_connector,
391 SLAB_PANIC);
392
393 return 0;
394}
395core_initcall(fsnotify_init);