blob: d21a7ecec6ee0641e020e975244ebac0df3f19d6 [file] [log] [blame]
b.liue9582032025-04-17 19:18:16 +08001/*
2 * JFFS2 -- Journalling Flash File System, Version 2.
3 *
4 * Copyright © 2001-2007 Red Hat, Inc.
5 * Copyright © 2004-2010 David Woodhouse <dwmw2@infradead.org>
6 *
7 * Created by David Woodhouse <dwmw2@infradead.org>
8 *
9 * For licensing information, see the file 'LICENCE' in this directory.
10 *
11 */
12
13#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15#include <linux/kernel.h>
16#include <linux/sched.h>
17#include <linux/slab.h>
18#include <linux/vmalloc.h>
19#include <linux/mtd/mtd.h>
20#include <linux/mm.h> /* kvfree() */
21#include "nodelist.h"
22
23static void jffs2_build_remove_unlinked_inode(struct jffs2_sb_info *,
24 struct jffs2_inode_cache *, struct jffs2_full_dirent **);
25
26static inline struct jffs2_inode_cache *
27first_inode_chain(int *i, struct jffs2_sb_info *c)
28{
29 for (; *i < c->inocache_hashsize; (*i)++) {
30 if (c->inocache_list[*i])
31 return c->inocache_list[*i];
32 }
33 return NULL;
34}
35
36static inline struct jffs2_inode_cache *
37next_inode(int *i, struct jffs2_inode_cache *ic, struct jffs2_sb_info *c)
38{
39 /* More in this chain? */
40 if (ic->next)
41 return ic->next;
42 (*i)++;
43 return first_inode_chain(i, c);
44}
45
46#define for_each_inode(i, c, ic) \
47 for (i = 0, ic = first_inode_chain(&i, (c)); \
48 ic; \
49 ic = next_inode(&i, ic, (c)))
50
51
52static void jffs2_build_inode_pass1(struct jffs2_sb_info *c,
53 struct jffs2_inode_cache *ic,
54 int *dir_hardlinks)
55{
56 struct jffs2_full_dirent *fd;
57
58 dbg_fsbuild("building directory inode #%u\n", ic->ino);
59
60 /* For each child, increase nlink */
61 for(fd = ic->scan_dents; fd; fd = fd->next) {
62 struct jffs2_inode_cache *child_ic;
63 if (!fd->ino)
64 continue;
65
66 /* we can get high latency here with huge directories */
67
68 child_ic = jffs2_get_ino_cache(c, fd->ino);
69 if (!child_ic) {
70 dbg_fsbuild("child \"%s\" (ino #%u) of dir ino #%u doesn't exist!\n",
71 fd->name, fd->ino, ic->ino);
72 jffs2_mark_node_obsolete(c, fd->raw);
73 /* Clear the ic/raw union so it doesn't cause problems later. */
74 fd->ic = NULL;
75 continue;
76 }
77
78 /* From this point, fd->raw is no longer used so we can set fd->ic */
79 fd->ic = child_ic;
80 child_ic->pino_nlink++;
81 /* If we appear (at this stage) to have hard-linked directories,
82 * set a flag to trigger a scan later */
83 if (fd->type == DT_DIR) {
84 child_ic->flags |= INO_FLAGS_IS_DIR;
85 if (child_ic->pino_nlink > 1)
86 *dir_hardlinks = 1;
87 }
88
89 dbg_fsbuild("increased nlink for child \"%s\" (ino #%u)\n", fd->name, fd->ino);
90 /* Can't free scan_dents so far. We might need them in pass 2 */
91 }
92}
93
94/* Scan plan:
95 - Scan physical nodes. Build map of inodes/dirents. Allocate inocaches as we go
96 - Scan directory tree from top down, setting nlink in inocaches
97 - Scan inocaches for inodes with nlink==0
98*/
99static int jffs2_build_filesystem(struct jffs2_sb_info *c)
100{
101 int ret, i, dir_hardlinks = 0;
102 struct jffs2_inode_cache *ic;
103 struct jffs2_full_dirent *fd;
104 struct jffs2_full_dirent *dead_fds = NULL;
105
106 dbg_fsbuild("build FS data structures\n");
107
108 /* First, scan the medium and build all the inode caches with
109 lists of physical nodes */
110
111 c->flags |= JFFS2_SB_FLAG_SCANNING;
112 ret = jffs2_scan_medium(c);
113 c->flags &= ~JFFS2_SB_FLAG_SCANNING;
114 if (ret)
115 goto exit;
116
117 dbg_fsbuild("scanned flash completely\n");
118 jffs2_dbg_dump_block_lists_nolock(c);
119
120 if (c->flags & (1 << 7)) {
121 printk("%s(): unlocking the mtd device... ", __func__);
122 mtd_unlock(c->mtd, 0, c->mtd->size);
123 printk("done.\n");
124
125 printk("%s(): erasing all blocks after the end marker... ", __func__);
126 jffs2_erase_pending_blocks(c, -1);
127 printk("done.\n");
128 }
129
130 dbg_fsbuild("pass 1 starting\n");
131 c->flags |= JFFS2_SB_FLAG_BUILDING;
132 /* Now scan the directory tree, increasing nlink according to every dirent found. */
133 for_each_inode(i, c, ic) {
134 if (ic->scan_dents) {
135 jffs2_build_inode_pass1(c, ic, &dir_hardlinks);
136 cond_resched();
137 }
138 }
139
140 dbg_fsbuild("pass 1 complete\n");
141
142 /* Next, scan for inodes with nlink == 0 and remove them. If
143 they were directories, then decrement the nlink of their
144 children too, and repeat the scan. As that's going to be
145 a fairly uncommon occurrence, it's not so evil to do it this
146 way. Recursion bad. */
147 dbg_fsbuild("pass 2 starting\n");
148
149 for_each_inode(i, c, ic) {
150 if (ic->pino_nlink)
151 continue;
152
153 jffs2_build_remove_unlinked_inode(c, ic, &dead_fds);
154 cond_resched();
155 }
156
157 dbg_fsbuild("pass 2a starting\n");
158
159 while (dead_fds) {
160 fd = dead_fds;
161 dead_fds = fd->next;
162
163 ic = jffs2_get_ino_cache(c, fd->ino);
164
165 if (ic)
166 jffs2_build_remove_unlinked_inode(c, ic, &dead_fds);
167 jffs2_free_full_dirent(fd);
168 }
169
170 dbg_fsbuild("pass 2a complete\n");
171
172 if (dir_hardlinks) {
173 /* If we detected directory hardlinks earlier, *hopefully*
174 * they are gone now because some of the links were from
175 * dead directories which still had some old dirents lying
176 * around and not yet garbage-collected, but which have
177 * been discarded above. So clear the pino_nlink field
178 * in each directory, so that the final scan below can
179 * print appropriate warnings. */
180 for_each_inode(i, c, ic) {
181 if (ic->flags & INO_FLAGS_IS_DIR)
182 ic->pino_nlink = 0;
183 }
184 }
185 dbg_fsbuild("freeing temporary data structures\n");
186
187 /* Finally, we can scan again and free the dirent structs */
188 for_each_inode(i, c, ic) {
189 while(ic->scan_dents) {
190 fd = ic->scan_dents;
191 ic->scan_dents = fd->next;
192 /* We do use the pino_nlink field to count nlink of
193 * directories during fs build, so set it to the
194 * parent ino# now. Now that there's hopefully only
195 * one. */
196 if (fd->type == DT_DIR) {
197 if (!fd->ic) {
198 /* We'll have complained about it and marked the coresponding
199 raw node obsolete already. Just skip it. */
200 continue;
201 }
202
203 /* We *have* to have set this in jffs2_build_inode_pass1() */
204 BUG_ON(!(fd->ic->flags & INO_FLAGS_IS_DIR));
205
206 /* We clear ic->pino_nlink ∀ directories' ic *only* if dir_hardlinks
207 * is set. Otherwise, we know this should never trigger anyway, so
208 * we don't do the check. And ic->pino_nlink still contains the nlink
209 * value (which is 1). */
210 if (dir_hardlinks && fd->ic->pino_nlink) {
211 JFFS2_ERROR("child dir \"%s\" (ino #%u) of dir ino #%u is also hard linked from dir ino #%u\n",
212 fd->name, fd->ino, ic->ino, fd->ic->pino_nlink);
213 /* Should we unlink it from its previous parent? */
214 }
215
216 /* For directories, ic->pino_nlink holds that parent inode # */
217 fd->ic->pino_nlink = ic->ino;
218 }
219 jffs2_free_full_dirent(fd);
220 }
221 ic->scan_dents = NULL;
222 cond_resched();
223 }
224 ret = jffs2_build_xattr_subsystem(c);
225 if (ret)
226 goto exit;
227
228 c->flags &= ~JFFS2_SB_FLAG_BUILDING;
229
230 dbg_fsbuild("FS build complete\n");
231
232 /* Rotate the lists by some number to ensure wear levelling */
233 jffs2_rotate_lists(c);
234
235 ret = 0;
236
237exit:
238 if (ret) {
239 for_each_inode(i, c, ic) {
240 while(ic->scan_dents) {
241 fd = ic->scan_dents;
242 ic->scan_dents = fd->next;
243 jffs2_free_full_dirent(fd);
244 }
245 }
246 jffs2_clear_xattr_subsystem(c);
247 }
248
249 return ret;
250}
251
252static void jffs2_build_remove_unlinked_inode(struct jffs2_sb_info *c,
253 struct jffs2_inode_cache *ic,
254 struct jffs2_full_dirent **dead_fds)
255{
256 struct jffs2_raw_node_ref *raw;
257 struct jffs2_full_dirent *fd;
258
259 dbg_fsbuild("removing ino #%u with nlink == zero.\n", ic->ino);
260
261 raw = ic->nodes;
262 while (raw != (void *)ic) {
263 struct jffs2_raw_node_ref *next = raw->next_in_ino;
264 dbg_fsbuild("obsoleting node at 0x%08x\n", ref_offset(raw));
265 jffs2_mark_node_obsolete(c, raw);
266 raw = next;
267 }
268
269 if (ic->scan_dents) {
270 int whinged = 0;
271 dbg_fsbuild("inode #%u was a directory which may have children...\n", ic->ino);
272
273 while(ic->scan_dents) {
274 struct jffs2_inode_cache *child_ic;
275
276 fd = ic->scan_dents;
277 ic->scan_dents = fd->next;
278
279 if (!fd->ino) {
280 /* It's a deletion dirent. Ignore it */
281 dbg_fsbuild("child \"%s\" is a deletion dirent, skipping...\n", fd->name);
282 jffs2_free_full_dirent(fd);
283 continue;
284 }
285 if (!whinged)
286 whinged = 1;
287
288 dbg_fsbuild("removing child \"%s\", ino #%u\n", fd->name, fd->ino);
289
290 child_ic = jffs2_get_ino_cache(c, fd->ino);
291 if (!child_ic) {
292 dbg_fsbuild("cannot remove child \"%s\", ino #%u, because it doesn't exist\n",
293 fd->name, fd->ino);
294 jffs2_free_full_dirent(fd);
295 continue;
296 }
297
298 /* Reduce nlink of the child. If it's now zero, stick it on the
299 dead_fds list to be cleaned up later. Else just free the fd */
300 child_ic->pino_nlink--;
301
302 if (!child_ic->pino_nlink) {
303 dbg_fsbuild("inode #%u (\"%s\") now has no links; adding to dead_fds list.\n",
304 fd->ino, fd->name);
305 fd->next = *dead_fds;
306 *dead_fds = fd;
307 } else {
308 dbg_fsbuild("inode #%u (\"%s\") has now got nlink %d. Ignoring.\n",
309 fd->ino, fd->name, child_ic->pino_nlink);
310 jffs2_free_full_dirent(fd);
311 }
312 }
313 }
314
315 /*
316 We don't delete the inocache from the hash list and free it yet.
317 The erase code will do that, when all the nodes are completely gone.
318 */
319}
320
321static void jffs2_calc_trigger_levels(struct jffs2_sb_info *c)
322{
323 uint32_t size;
324
325 /* Deletion should almost _always_ be allowed. We're fairly
326 buggered once we stop allowing people to delete stuff
327 because there's not enough free space... */
328 c->resv_blocks_deletion = 2;
329
330 /* Be conservative about how much space we need before we allow writes.
331 On top of that which is required for deletia, require an extra 2%
332 of the medium to be available, for overhead caused by nodes being
333 split across blocks, etc. */
334
335 size = c->flash_size / 50; /* 2% of flash size */
336 size += c->nr_blocks * 100; /* And 100 bytes per eraseblock */
337 size += c->sector_size - 1; /* ... and round up */
338
339 c->resv_blocks_write = c->resv_blocks_deletion + (size / c->sector_size);
340
341 /* When do we let the GC thread run in the background */
342
343 c->resv_blocks_gctrigger = c->resv_blocks_write + 1;
344
345 /* When do we allow garbage collection to merge nodes to make
346 long-term progress at the expense of short-term space exhaustion? */
347 c->resv_blocks_gcmerge = c->resv_blocks_deletion + 1;
348
349 /* When do we allow garbage collection to eat from bad blocks rather
350 than actually making progress? */
351 c->resv_blocks_gcbad = 0;//c->resv_blocks_deletion + 2;
352
353 /* What number of 'very dirty' eraseblocks do we allow before we
354 trigger the GC thread even if we don't _need_ the space. When we
355 can't mark nodes obsolete on the medium, the old dirty nodes cause
356 performance problems because we have to inspect and discard them. */
357 c->vdirty_blocks_gctrigger = c->resv_blocks_gctrigger;
358 if (jffs2_can_mark_obsolete(c))
359 c->vdirty_blocks_gctrigger *= 10;
360
361 /* If there's less than this amount of dirty space, don't bother
362 trying to GC to make more space. It'll be a fruitless task */
363 c->nospc_dirty_size = c->sector_size + (c->flash_size / 100);
364
365 dbg_fsbuild("trigger levels (size %d KiB, block size %d KiB, %d blocks)\n",
366 c->flash_size / 1024, c->sector_size / 1024, c->nr_blocks);
367 dbg_fsbuild("Blocks required to allow deletion: %d (%d KiB)\n",
368 c->resv_blocks_deletion, c->resv_blocks_deletion*c->sector_size/1024);
369 dbg_fsbuild("Blocks required to allow writes: %d (%d KiB)\n",
370 c->resv_blocks_write, c->resv_blocks_write*c->sector_size/1024);
371 dbg_fsbuild("Blocks required to quiesce GC thread: %d (%d KiB)\n",
372 c->resv_blocks_gctrigger, c->resv_blocks_gctrigger*c->sector_size/1024);
373 dbg_fsbuild("Blocks required to allow GC merges: %d (%d KiB)\n",
374 c->resv_blocks_gcmerge, c->resv_blocks_gcmerge*c->sector_size/1024);
375 dbg_fsbuild("Blocks required to GC bad blocks: %d (%d KiB)\n",
376 c->resv_blocks_gcbad, c->resv_blocks_gcbad*c->sector_size/1024);
377 dbg_fsbuild("Amount of dirty space required to GC: %d bytes\n",
378 c->nospc_dirty_size);
379 dbg_fsbuild("Very dirty blocks before GC triggered: %d\n",
380 c->vdirty_blocks_gctrigger);
381}
382
383int jffs2_do_mount_fs(struct jffs2_sb_info *c)
384{
385 int ret;
386 int i;
387 int size;
388
389 c->free_size = c->flash_size;
390 c->nr_blocks = c->flash_size / c->sector_size;
391 size = sizeof(struct jffs2_eraseblock) * c->nr_blocks;
392#ifndef __ECOS
393 if (jffs2_blocks_use_vmalloc(c))
394 c->blocks = vzalloc(size);
395 else
396#endif
397 c->blocks = kzalloc(size, GFP_KERNEL);
398 if (!c->blocks)
399 return -ENOMEM;
400
401 for (i=0; i<c->nr_blocks; i++) {
402 INIT_LIST_HEAD(&c->blocks[i].list);
403 c->blocks[i].offset = i * c->sector_size;
404 c->blocks[i].free_size = c->sector_size;
405 }
406
407 INIT_LIST_HEAD(&c->clean_list);
408 INIT_LIST_HEAD(&c->very_dirty_list);
409 INIT_LIST_HEAD(&c->dirty_list);
410 INIT_LIST_HEAD(&c->erasable_list);
411 INIT_LIST_HEAD(&c->erasing_list);
412 INIT_LIST_HEAD(&c->erase_checking_list);
413 INIT_LIST_HEAD(&c->erase_pending_list);
414 INIT_LIST_HEAD(&c->erasable_pending_wbuf_list);
415 INIT_LIST_HEAD(&c->erase_complete_list);
416 INIT_LIST_HEAD(&c->free_list);
417 INIT_LIST_HEAD(&c->bad_list);
418 INIT_LIST_HEAD(&c->bad_used_list);
419 c->highest_ino = 1;
420 c->summary = NULL;
421
422 ret = jffs2_sum_init(c);
423 if (ret)
424 goto out_free;
425
426 if (jffs2_build_filesystem(c)) {
427 dbg_fsbuild("build_fs failed\n");
428 jffs2_free_ino_caches(c);
429 jffs2_free_raw_node_refs(c);
430 ret = -EIO;
431 goto out_sum_exit;
432 }
433
434 jffs2_calc_trigger_levels(c);
435
436 return 0;
437
438 out_sum_exit:
439 jffs2_sum_exit(c);
440 out_free:
441 kvfree(c->blocks);
442
443 return ret;
444}