README.Locking 7.1 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173
  1. $Id: README.Locking,v 1.1.1.1 2007/06/12 07:27:13 eyryu Exp $
  2. JFFS2 LOCKING DOCUMENTATION
  3. ---------------------------
  4. At least theoretically, JFFS2 does not require the Big Kernel Lock
  5. (BKL), which was always helpfully obtained for it by Linux 2.4 VFS
  6. code. It has its own locking, as described below.
  7. This document attempts to describe the existing locking rules for
  8. JFFS2. It is not expected to remain perfectly up to date, but ought to
  9. be fairly close.
  10. alloc_sem
  11. ---------
  12. The alloc_sem is a per-filesystem semaphore, used primarily to ensure
  13. contiguous allocation of space on the medium. It is automatically
  14. obtained during space allocations (jffs2_reserve_space()) and freed
  15. upon write completion (jffs2_complete_reservation()). Note that
  16. the garbage collector will obtain this right at the beginning of
  17. jffs2_garbage_collect_pass() and release it at the end, thereby
  18. preventing any other write activity on the file system during a
  19. garbage collect pass.
  20. When writing new nodes, the alloc_sem must be held until the new nodes
  21. have been properly linked into the data structures for the inode to
  22. which they belong. This is for the benefit of NAND flash - adding new
  23. nodes to an inode may obsolete old ones, and by holding the alloc_sem
  24. until this happens we ensure that any data in the write-buffer at the
  25. time this happens are part of the new node, not just something that
  26. was written afterwards. Hence, we can ensure the newly-obsoleted nodes
  27. don't actually get erased until the write-buffer has been flushed to
  28. the medium.
  29. With the introduction of NAND flash support and the write-buffer,
  30. the alloc_sem is also used to protect the wbuf-related members of the
  31. jffs2_sb_info structure. Atomically reading the wbuf_len member to see
  32. if the wbuf is currently holding any data is permitted, though.
  33. Ordering constraints: See f->sem.
  34. File Semaphore f->sem
  35. ---------------------
  36. This is the JFFS2-internal equivalent of the inode semaphore i->i_sem.
  37. It protects the contents of the jffs2_inode_info private inode data,
  38. including the linked list of node fragments (but see the notes below on
  39. erase_completion_lock), etc.
  40. The reason that the i_sem itself isn't used for this purpose is to
  41. avoid deadlocks with garbage collection -- the VFS will lock the i_sem
  42. before calling a function which may need to allocate space. The
  43. allocation may trigger garbage-collection, which may need to move a
  44. node belonging to the inode which was locked in the first place by the
  45. VFS. If the garbage collection code were to attempt to lock the i_sem
  46. of the inode from which it's garbage-collecting a physical node, this
  47. lead to deadlock, unless we played games with unlocking the i_sem
  48. before calling the space allocation functions.
  49. Instead of playing such games, we just have an extra internal
  50. semaphore, which is obtained by the garbage collection code and also
  51. by the normal file system code _after_ allocation of space.
  52. Ordering constraints:
  53. 1. Never attempt to allocate space or lock alloc_sem with
  54. any f->sem held.
  55. 2. Never attempt to lock two file semaphores in one thread.
  56. No ordering rules have been made for doing so.
  57. erase_completion_lock spinlock
  58. ------------------------------
  59. This is used to serialise access to the eraseblock lists, to the
  60. per-eraseblock lists of physical jffs2_raw_node_ref structures, and
  61. (NB) the per-inode list of physical nodes. The latter is a special
  62. case - see below.
  63. As the MTD API no longer permits erase-completion callback functions
  64. to be called from bottom-half (timer) context (on the basis that nobody
  65. ever actually implemented such a thing), it's now sufficient to use
  66. a simple spin_lock() rather than spin_lock_bh().
  67. Note that the per-inode list of physical nodes (f->nodes) is a special
  68. case. Any changes to _valid_ nodes (i.e. ->flash_offset & 1 == 0) in
  69. the list are protected by the file semaphore f->sem. But the erase
  70. code may remove _obsolete_ nodes from the list while holding only the
  71. erase_completion_lock. So you can walk the list only while holding the
  72. erase_completion_lock, and can drop the lock temporarily mid-walk as
  73. long as the pointer you're holding is to a _valid_ node, not an
  74. obsolete one.
  75. The erase_completion_lock is also used to protect the c->gc_task
  76. pointer when the garbage collection thread exits. The code to kill the
  77. GC thread locks it, sends the signal, then unlocks it - while the GC
  78. thread itself locks it, zeroes c->gc_task, then unlocks on the exit path.
  79. inocache_lock spinlock
  80. ----------------------
  81. This spinlock protects the hashed list (c->inocache_list) of the
  82. in-core jffs2_inode_cache objects (each inode in JFFS2 has the
  83. correspondent jffs2_inode_cache object). So, the inocache_lock
  84. has to be locked while walking the c->inocache_list hash buckets.
  85. This spinlock also covers allocation of new inode numbers, which is
  86. currently just '++->highest_ino++', but might one day get more complicated
  87. if we need to deal with wrapping after 4 milliard inode numbers are used.
  88. Note, the f->sem guarantees that the correspondent jffs2_inode_cache
  89. will not be removed. So, it is allowed to access it without locking
  90. the inocache_lock spinlock.
  91. Ordering constraints:
  92. If both erase_completion_lock and inocache_lock are needed, the
  93. c->erase_completion has to be acquired first.
  94. erase_free_sem
  95. --------------
  96. This semaphore is only used by the erase code which frees obsolete
  97. node references and the jffs2_garbage_collect_deletion_dirent()
  98. function. The latter function on NAND flash must read _obsolete_ nodes
  99. to determine whether the 'deletion dirent' under consideration can be
  100. discarded or whether it is still required to show that an inode has
  101. been unlinked. Because reading from the flash may sleep, the
  102. erase_completion_lock cannot be held, so an alternative, more
  103. heavyweight lock was required to prevent the erase code from freeing
  104. the jffs2_raw_node_ref structures in question while the garbage
  105. collection code is looking at them.
  106. Suggestions for alternative solutions to this problem would be welcomed.
  107. wbuf_sem
  108. --------
  109. This read/write semaphore protects against concurrent access to the
  110. write-behind buffer ('wbuf') used for flash chips where we must write
  111. in blocks. It protects both the contents of the wbuf and the metadata
  112. which indicates which flash region (if any) is currently covered by
  113. the buffer.
  114. Ordering constraints:
  115. Lock wbuf_sem last, after the alloc_sem or and f->sem.
  116. c->xattr_sem
  117. ------------
  118. This read/write semaphore protects against concurrent access to the
  119. xattr related objects which include stuff in superblock and ic->xref.
  120. In read-only path, write-semaphore is too much exclusion. It's enough
  121. by read-semaphore. But you must hold write-semaphore when updating,
  122. creating or deleting any xattr related object.
  123. Once xattr_sem released, there would be no assurance for the existence
  124. of those objects. Thus, a series of processes is often required to retry,
  125. when updating such a object is necessary under holding read semaphore.
  126. For example, do_jffs2_getxattr() holds read-semaphore to scan xref and
  127. xdatum at first. But it retries this process with holding write-semaphore
  128. after release read-semaphore, if it's necessary to load name/value pair
  129. from medium.
  130. Ordering constraints:
  131. Lock xattr_sem last, after the alloc_sem.