dm-zoned.h 8.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. /*
  3. * Copyright (C) 2017 Western Digital Corporation or its affiliates.
  4. *
  5. * This file is released under the GPL.
  6. */
  7. #ifndef DM_ZONED_H
  8. #define DM_ZONED_H
  9. #include <linux/types.h>
  10. #include <linux/blkdev.h>
  11. #include <linux/device-mapper.h>
  12. #include <linux/dm-kcopyd.h>
  13. #include <linux/list.h>
  14. #include <linux/spinlock.h>
  15. #include <linux/mutex.h>
  16. #include <linux/workqueue.h>
  17. #include <linux/rwsem.h>
  18. #include <linux/rbtree.h>
  19. #include <linux/radix-tree.h>
  20. #include <linux/shrinker.h>
  21. /*
  22. * dm-zoned creates block devices with 4KB blocks, always.
  23. */
  24. #define DMZ_BLOCK_SHIFT 12
  25. #define DMZ_BLOCK_SIZE (1 << DMZ_BLOCK_SHIFT)
  26. #define DMZ_BLOCK_MASK (DMZ_BLOCK_SIZE - 1)
  27. #define DMZ_BLOCK_SHIFT_BITS (DMZ_BLOCK_SHIFT + 3)
  28. #define DMZ_BLOCK_SIZE_BITS (1 << DMZ_BLOCK_SHIFT_BITS)
  29. #define DMZ_BLOCK_MASK_BITS (DMZ_BLOCK_SIZE_BITS - 1)
  30. #define DMZ_BLOCK_SECTORS_SHIFT (DMZ_BLOCK_SHIFT - SECTOR_SHIFT)
  31. #define DMZ_BLOCK_SECTORS (DMZ_BLOCK_SIZE >> SECTOR_SHIFT)
  32. #define DMZ_BLOCK_SECTORS_MASK (DMZ_BLOCK_SECTORS - 1)
  33. /*
  34. * 4KB block <-> 512B sector conversion.
  35. */
  36. #define dmz_blk2sect(b) ((sector_t)(b) << DMZ_BLOCK_SECTORS_SHIFT)
  37. #define dmz_sect2blk(s) ((sector_t)(s) >> DMZ_BLOCK_SECTORS_SHIFT)
  38. #define dmz_bio_block(bio) dmz_sect2blk((bio)->bi_iter.bi_sector)
  39. #define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
  40. struct dmz_metadata;
  41. struct dmz_reclaim;
  42. /*
  43. * Zoned block device information.
  44. */
  45. struct dmz_dev {
  46. struct block_device *bdev;
  47. struct dmz_metadata *metadata;
  48. struct dmz_reclaim *reclaim;
  49. char name[BDEVNAME_SIZE];
  50. uuid_t uuid;
  51. sector_t capacity;
  52. unsigned int dev_idx;
  53. unsigned int nr_zones;
  54. unsigned int zone_offset;
  55. unsigned int flags;
  56. sector_t zone_nr_sectors;
  57. unsigned int nr_rnd;
  58. atomic_t unmap_nr_rnd;
  59. struct list_head unmap_rnd_list;
  60. struct list_head map_rnd_list;
  61. unsigned int nr_seq;
  62. atomic_t unmap_nr_seq;
  63. struct list_head unmap_seq_list;
  64. struct list_head map_seq_list;
  65. };
  66. #define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \
  67. dmz_zone_nr_sectors_shift(zmd))
  68. #define dmz_chunk_block(zmd, b) ((b) & (dmz_zone_nr_blocks(zmd) - 1))
  69. /* Device flags. */
  70. #define DMZ_BDEV_DYING (1 << 0)
  71. #define DMZ_CHECK_BDEV (2 << 0)
  72. #define DMZ_BDEV_REGULAR (4 << 0)
  73. /*
  74. * Zone descriptor.
  75. */
  76. struct dm_zone {
  77. /* For listing the zone depending on its state */
  78. struct list_head link;
  79. /* Device containing this zone */
  80. struct dmz_dev *dev;
  81. /* Zone type and state */
  82. unsigned long flags;
  83. /* Zone activation reference count */
  84. atomic_t refcount;
  85. /* Zone id */
  86. unsigned int id;
  87. /* Zone write pointer block (relative to the zone start block) */
  88. unsigned int wp_block;
  89. /* Zone weight (number of valid blocks in the zone) */
  90. unsigned int weight;
  91. /* The chunk that the zone maps */
  92. unsigned int chunk;
  93. /*
  94. * For a sequential data zone, pointer to the random zone
  95. * used as a buffer for processing unaligned writes.
  96. * For a buffer zone, this points back to the data zone.
  97. */
  98. struct dm_zone *bzone;
  99. };
  100. /*
  101. * Zone flags.
  102. */
  103. enum {
  104. /* Zone write type */
  105. DMZ_CACHE,
  106. DMZ_RND,
  107. DMZ_SEQ,
  108. /* Zone critical condition */
  109. DMZ_OFFLINE,
  110. DMZ_READ_ONLY,
  111. /* How the zone is being used */
  112. DMZ_META,
  113. DMZ_DATA,
  114. DMZ_BUF,
  115. DMZ_RESERVED,
  116. /* Zone internal state */
  117. DMZ_RECLAIM,
  118. DMZ_SEQ_WRITE_ERR,
  119. DMZ_RECLAIM_TERMINATE,
  120. };
  121. /*
  122. * Zone data accessors.
  123. */
  124. #define dmz_is_cache(z) test_bit(DMZ_CACHE, &(z)->flags)
  125. #define dmz_is_rnd(z) test_bit(DMZ_RND, &(z)->flags)
  126. #define dmz_is_seq(z) test_bit(DMZ_SEQ, &(z)->flags)
  127. #define dmz_is_empty(z) ((z)->wp_block == 0)
  128. #define dmz_is_offline(z) test_bit(DMZ_OFFLINE, &(z)->flags)
  129. #define dmz_is_readonly(z) test_bit(DMZ_READ_ONLY, &(z)->flags)
  130. #define dmz_in_reclaim(z) test_bit(DMZ_RECLAIM, &(z)->flags)
  131. #define dmz_is_reserved(z) test_bit(DMZ_RESERVED, &(z)->flags)
  132. #define dmz_seq_write_err(z) test_bit(DMZ_SEQ_WRITE_ERR, &(z)->flags)
  133. #define dmz_reclaim_should_terminate(z) \
  134. test_bit(DMZ_RECLAIM_TERMINATE, &(z)->flags)
  135. #define dmz_is_meta(z) test_bit(DMZ_META, &(z)->flags)
  136. #define dmz_is_buf(z) test_bit(DMZ_BUF, &(z)->flags)
  137. #define dmz_is_data(z) test_bit(DMZ_DATA, &(z)->flags)
  138. #define dmz_weight(z) ((z)->weight)
  139. /*
  140. * Message functions.
  141. */
  142. #define dmz_dev_info(dev, format, args...) \
  143. DMINFO("(%s): " format, (dev)->name, ## args)
  144. #define dmz_dev_err(dev, format, args...) \
  145. DMERR("(%s): " format, (dev)->name, ## args)
  146. #define dmz_dev_warn(dev, format, args...) \
  147. DMWARN("(%s): " format, (dev)->name, ## args)
  148. #define dmz_dev_debug(dev, format, args...) \
  149. DMDEBUG("(%s): " format, (dev)->name, ## args)
  150. /*
  151. * Functions defined in dm-zoned-metadata.c
  152. */
  153. int dmz_ctr_metadata(struct dmz_dev *dev, int num_dev,
  154. struct dmz_metadata **zmd, const char *devname);
  155. void dmz_dtr_metadata(struct dmz_metadata *zmd);
  156. int dmz_resume_metadata(struct dmz_metadata *zmd);
  157. void dmz_lock_map(struct dmz_metadata *zmd);
  158. void dmz_unlock_map(struct dmz_metadata *zmd);
  159. void dmz_lock_metadata(struct dmz_metadata *zmd);
  160. void dmz_unlock_metadata(struct dmz_metadata *zmd);
  161. void dmz_lock_flush(struct dmz_metadata *zmd);
  162. void dmz_unlock_flush(struct dmz_metadata *zmd);
  163. int dmz_flush_metadata(struct dmz_metadata *zmd);
  164. const char *dmz_metadata_label(struct dmz_metadata *zmd);
  165. sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone);
  166. sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone);
  167. unsigned int dmz_nr_chunks(struct dmz_metadata *zmd);
  168. bool dmz_check_dev(struct dmz_metadata *zmd);
  169. bool dmz_dev_is_dying(struct dmz_metadata *zmd);
  170. #define DMZ_ALLOC_RND 0x01
  171. #define DMZ_ALLOC_CACHE 0x02
  172. #define DMZ_ALLOC_SEQ 0x04
  173. #define DMZ_ALLOC_RECLAIM 0x10
  174. struct dm_zone *dmz_alloc_zone(struct dmz_metadata *zmd,
  175. unsigned int dev_idx, unsigned long flags);
  176. void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
  177. void dmz_map_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
  178. unsigned int chunk);
  179. void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
  180. unsigned int dmz_nr_zones(struct dmz_metadata *zmd);
  181. unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd);
  182. unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd);
  183. unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd, int idx);
  184. unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd, int idx);
  185. unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd, int idx);
  186. unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd, int idx);
  187. unsigned int dmz_zone_nr_blocks(struct dmz_metadata *zmd);
  188. unsigned int dmz_zone_nr_blocks_shift(struct dmz_metadata *zmd);
  189. unsigned int dmz_zone_nr_sectors(struct dmz_metadata *zmd);
  190. unsigned int dmz_zone_nr_sectors_shift(struct dmz_metadata *zmd);
  191. /*
  192. * Activate a zone (increment its reference count).
  193. */
  194. static inline void dmz_activate_zone(struct dm_zone *zone)
  195. {
  196. atomic_inc(&zone->refcount);
  197. }
  198. int dmz_lock_zone_reclaim(struct dm_zone *zone);
  199. void dmz_unlock_zone_reclaim(struct dm_zone *zone);
  200. struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd,
  201. unsigned int dev_idx, bool idle);
  202. struct dm_zone *dmz_get_chunk_mapping(struct dmz_metadata *zmd,
  203. unsigned int chunk, int op);
  204. void dmz_put_chunk_mapping(struct dmz_metadata *zmd, struct dm_zone *zone);
  205. struct dm_zone *dmz_get_chunk_buffer(struct dmz_metadata *zmd,
  206. struct dm_zone *dzone);
  207. int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
  208. sector_t chunk_block, unsigned int nr_blocks);
  209. int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
  210. sector_t chunk_block, unsigned int nr_blocks);
  211. int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone,
  212. sector_t chunk_block);
  213. int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone,
  214. sector_t *chunk_block);
  215. int dmz_copy_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
  216. struct dm_zone *to_zone);
  217. int dmz_merge_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
  218. struct dm_zone *to_zone, sector_t chunk_block);
  219. /*
  220. * Functions defined in dm-zoned-reclaim.c
  221. */
  222. int dmz_ctr_reclaim(struct dmz_metadata *zmd, struct dmz_reclaim **zrc, int idx);
  223. void dmz_dtr_reclaim(struct dmz_reclaim *zrc);
  224. void dmz_suspend_reclaim(struct dmz_reclaim *zrc);
  225. void dmz_resume_reclaim(struct dmz_reclaim *zrc);
  226. void dmz_reclaim_bio_acc(struct dmz_reclaim *zrc);
  227. void dmz_schedule_reclaim(struct dmz_reclaim *zrc);
  228. /*
  229. * Functions defined in dm-zoned-target.c
  230. */
  231. bool dmz_bdev_is_dying(struct dmz_dev *dmz_dev);
  232. bool dmz_check_bdev(struct dmz_dev *dmz_dev);
  233. /*
  234. * Deactivate a zone. This decrement the zone reference counter
  235. * indicating that all BIOs to the zone have completed when the count is 0.
  236. */
  237. static inline void dmz_deactivate_zone(struct dm_zone *zone)
  238. {
  239. dmz_reclaim_bio_acc(zone->dev->reclaim);
  240. atomic_dec(&zone->refcount);
  241. }
  242. /*
  243. * Test if a zone is active, that is, has a refcount > 0.
  244. */
  245. static inline bool dmz_is_active(struct dm_zone *zone)
  246. {
  247. return atomic_read(&zone->refcount);
  248. }
  249. #endif /* DM_ZONED_H */