at v4.7 14 kB view raw
1#ifndef NVM_H 2#define NVM_H 3 4#include <linux/types.h> 5 6enum { 7 NVM_IO_OK = 0, 8 NVM_IO_REQUEUE = 1, 9 NVM_IO_DONE = 2, 10 NVM_IO_ERR = 3, 11 12 NVM_IOTYPE_NONE = 0, 13 NVM_IOTYPE_GC = 1, 14}; 15 16#define NVM_BLK_BITS (16) 17#define NVM_PG_BITS (16) 18#define NVM_SEC_BITS (8) 19#define NVM_PL_BITS (8) 20#define NVM_LUN_BITS (8) 21#define NVM_CH_BITS (7) 22 23struct ppa_addr { 24 /* Generic structure for all addresses */ 25 union { 26 struct { 27 u64 blk : NVM_BLK_BITS; 28 u64 pg : NVM_PG_BITS; 29 u64 sec : NVM_SEC_BITS; 30 u64 pl : NVM_PL_BITS; 31 u64 lun : NVM_LUN_BITS; 32 u64 ch : NVM_CH_BITS; 33 u64 reserved : 1; 34 } g; 35 36 struct { 37 u64 line : 63; 38 u64 is_cached : 1; 39 } c; 40 41 u64 ppa; 42 }; 43}; 44 45struct nvm_rq; 46struct nvm_id; 47struct nvm_dev; 48 49typedef int (nvm_l2p_update_fn)(u64, u32, __le64 *, void *); 50typedef int (nvm_id_fn)(struct nvm_dev *, struct nvm_id *); 51typedef int (nvm_get_l2p_tbl_fn)(struct nvm_dev *, u64, u32, 52 nvm_l2p_update_fn *, void *); 53typedef int (nvm_op_bb_tbl_fn)(struct nvm_dev *, struct ppa_addr, u8 *); 54typedef int (nvm_op_set_bb_fn)(struct nvm_dev *, struct ppa_addr *, int, int); 55typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *); 56typedef int (nvm_erase_blk_fn)(struct nvm_dev *, struct nvm_rq *); 57typedef void *(nvm_create_dma_pool_fn)(struct nvm_dev *, char *); 58typedef void (nvm_destroy_dma_pool_fn)(void *); 59typedef void *(nvm_dev_dma_alloc_fn)(struct nvm_dev *, void *, gfp_t, 60 dma_addr_t *); 61typedef void (nvm_dev_dma_free_fn)(void *, void*, dma_addr_t); 62 63struct nvm_dev_ops { 64 nvm_id_fn *identity; 65 nvm_get_l2p_tbl_fn *get_l2p_tbl; 66 nvm_op_bb_tbl_fn *get_bb_tbl; 67 nvm_op_set_bb_fn *set_bb_tbl; 68 69 nvm_submit_io_fn *submit_io; 70 nvm_erase_blk_fn *erase_block; 71 72 nvm_create_dma_pool_fn *create_dma_pool; 73 nvm_destroy_dma_pool_fn *destroy_dma_pool; 74 nvm_dev_dma_alloc_fn *dev_dma_alloc; 75 nvm_dev_dma_free_fn *dev_dma_free; 76 77 unsigned int max_phys_sect; 78}; 79 80 81 82#ifdef CONFIG_NVM 83 84#include <linux/blkdev.h> 85#include <linux/file.h> 86#include <linux/dmapool.h> 87#include <uapi/linux/lightnvm.h> 88 89enum { 90 /* HW Responsibilities */ 91 NVM_RSP_L2P = 1 << 0, 92 NVM_RSP_ECC = 1 << 1, 93 94 /* Physical Adressing Mode */ 95 NVM_ADDRMODE_LINEAR = 0, 96 NVM_ADDRMODE_CHANNEL = 1, 97 98 /* Plane programming mode for LUN */ 99 NVM_PLANE_SINGLE = 1, 100 NVM_PLANE_DOUBLE = 2, 101 NVM_PLANE_QUAD = 4, 102 103 /* Status codes */ 104 NVM_RSP_SUCCESS = 0x0, 105 NVM_RSP_NOT_CHANGEABLE = 0x1, 106 NVM_RSP_ERR_FAILWRITE = 0x40ff, 107 NVM_RSP_ERR_EMPTYPAGE = 0x42ff, 108 109 /* Device opcodes */ 110 NVM_OP_HBREAD = 0x02, 111 NVM_OP_HBWRITE = 0x81, 112 NVM_OP_PWRITE = 0x91, 113 NVM_OP_PREAD = 0x92, 114 NVM_OP_ERASE = 0x90, 115 116 /* PPA Command Flags */ 117 NVM_IO_SNGL_ACCESS = 0x0, 118 NVM_IO_DUAL_ACCESS = 0x1, 119 NVM_IO_QUAD_ACCESS = 0x2, 120 121 /* NAND Access Modes */ 122 NVM_IO_SUSPEND = 0x80, 123 NVM_IO_SLC_MODE = 0x100, 124 NVM_IO_SCRAMBLE_DISABLE = 0x200, 125 126 /* Block Types */ 127 NVM_BLK_T_FREE = 0x0, 128 NVM_BLK_T_BAD = 0x1, 129 NVM_BLK_T_GRWN_BAD = 0x2, 130 NVM_BLK_T_DEV = 0x4, 131 NVM_BLK_T_HOST = 0x8, 132 133 /* Memory capabilities */ 134 NVM_ID_CAP_SLC = 0x1, 135 NVM_ID_CAP_CMD_SUSPEND = 0x2, 136 NVM_ID_CAP_SCRAMBLE = 0x4, 137 NVM_ID_CAP_ENCRYPT = 0x8, 138 139 /* Memory types */ 140 NVM_ID_FMTYPE_SLC = 0, 141 NVM_ID_FMTYPE_MLC = 1, 142 143 /* Device capabilities */ 144 NVM_ID_DCAP_BBLKMGMT = 0x1, 145 NVM_UD_DCAP_ECC = 0x2, 146}; 147 148struct nvm_id_lp_mlc { 149 u16 num_pairs; 150 u8 pairs[886]; 151}; 152 153struct nvm_id_lp_tbl { 154 __u8 id[8]; 155 struct nvm_id_lp_mlc mlc; 156}; 157 158struct nvm_id_group { 159 u8 mtype; 160 u8 fmtype; 161 u8 num_ch; 162 u8 num_lun; 163 u8 num_pln; 164 u16 num_blk; 165 u16 num_pg; 166 u16 fpg_sz; 167 u16 csecs; 168 u16 sos; 169 u32 trdt; 170 u32 trdm; 171 u32 tprt; 172 u32 tprm; 173 u32 tbet; 174 u32 tbem; 175 u32 mpos; 176 u32 mccap; 177 u16 cpar; 178 179 struct nvm_id_lp_tbl lptbl; 180}; 181 182struct nvm_addr_format { 183 u8 ch_offset; 184 u8 ch_len; 185 u8 lun_offset; 186 u8 lun_len; 187 u8 pln_offset; 188 u8 pln_len; 189 u8 blk_offset; 190 u8 blk_len; 191 u8 pg_offset; 192 u8 pg_len; 193 u8 sect_offset; 194 u8 sect_len; 195}; 196 197struct nvm_id { 198 u8 ver_id; 199 u8 vmnt; 200 u8 cgrps; 201 u32 cap; 202 u32 dom; 203 struct nvm_addr_format ppaf; 204 struct nvm_id_group groups[4]; 205} __packed; 206 207struct nvm_target { 208 struct list_head list; 209 struct nvm_dev *dev; 210 struct nvm_tgt_type *type; 211 struct gendisk *disk; 212}; 213 214struct nvm_tgt_instance { 215 struct nvm_tgt_type *tt; 216}; 217 218#define ADDR_EMPTY (~0ULL) 219 220#define NVM_VERSION_MAJOR 1 221#define NVM_VERSION_MINOR 0 222#define NVM_VERSION_PATCH 0 223 224struct nvm_rq; 225typedef void (nvm_end_io_fn)(struct nvm_rq *); 226 227struct nvm_rq { 228 struct nvm_tgt_instance *ins; 229 struct nvm_dev *dev; 230 231 struct bio *bio; 232 233 union { 234 struct ppa_addr ppa_addr; 235 dma_addr_t dma_ppa_list; 236 }; 237 238 struct ppa_addr *ppa_list; 239 240 void *meta_list; 241 dma_addr_t dma_meta_list; 242 243 struct completion *wait; 244 nvm_end_io_fn *end_io; 245 246 uint8_t opcode; 247 uint16_t nr_ppas; 248 uint16_t flags; 249 250 u64 ppa_status; /* ppa media status */ 251 int error; 252}; 253 254static inline struct nvm_rq *nvm_rq_from_pdu(void *pdu) 255{ 256 return pdu - sizeof(struct nvm_rq); 257} 258 259static inline void *nvm_rq_to_pdu(struct nvm_rq *rqdata) 260{ 261 return rqdata + 1; 262} 263 264struct nvm_block; 265 266struct nvm_lun { 267 int id; 268 269 int lun_id; 270 int chnl_id; 271 272 /* It is up to the target to mark blocks as closed. If the target does 273 * not do it, all blocks are marked as open, and nr_open_blocks 274 * represents the number of blocks in use 275 */ 276 unsigned int nr_open_blocks; /* Number of used, writable blocks */ 277 unsigned int nr_closed_blocks; /* Number of used, read-only blocks */ 278 unsigned int nr_free_blocks; /* Number of unused blocks */ 279 unsigned int nr_bad_blocks; /* Number of bad blocks */ 280 281 spinlock_t lock; 282 283 struct nvm_block *blocks; 284}; 285 286enum { 287 NVM_BLK_ST_FREE = 0x1, /* Free block */ 288 NVM_BLK_ST_OPEN = 0x2, /* Open block - read-write */ 289 NVM_BLK_ST_CLOSED = 0x4, /* Closed block - read-only */ 290 NVM_BLK_ST_BAD = 0x8, /* Bad block */ 291}; 292 293struct nvm_block { 294 struct list_head list; 295 struct nvm_lun *lun; 296 unsigned long id; 297 298 void *priv; 299 int state; 300}; 301 302/* system block cpu representation */ 303struct nvm_sb_info { 304 unsigned long seqnr; 305 unsigned long erase_cnt; 306 unsigned int version; 307 char mmtype[NVM_MMTYPE_LEN]; 308 struct ppa_addr fs_ppa; 309}; 310 311struct nvm_dev { 312 struct nvm_dev_ops *ops; 313 314 struct list_head devices; 315 316 /* Media manager */ 317 struct nvmm_type *mt; 318 void *mp; 319 320 /* System blocks */ 321 struct nvm_sb_info sb; 322 323 /* Device information */ 324 int nr_chnls; 325 int nr_planes; 326 int luns_per_chnl; 327 int sec_per_pg; /* only sectors for a single page */ 328 int pgs_per_blk; 329 int blks_per_lun; 330 int fpg_size; 331 int pfpg_size; /* size of buffer if all pages are to be read */ 332 int sec_size; 333 int oob_size; 334 int mccap; 335 struct nvm_addr_format ppaf; 336 337 /* Calculated/Cached values. These do not reflect the actual usable 338 * blocks at run-time. 339 */ 340 int max_rq_size; 341 int plane_mode; /* drive device in single, double or quad mode */ 342 343 int sec_per_pl; /* all sectors across planes */ 344 int sec_per_blk; 345 int sec_per_lun; 346 347 /* lower page table */ 348 int lps_per_blk; 349 int *lptbl; 350 351 unsigned long total_blocks; 352 unsigned long total_secs; 353 int nr_luns; 354 355 unsigned long *lun_map; 356 void *dma_pool; 357 358 struct nvm_id identity; 359 360 /* Backend device */ 361 struct request_queue *q; 362 char name[DISK_NAME_LEN]; 363 364 struct mutex mlock; 365 spinlock_t lock; 366}; 367 368static inline struct ppa_addr generic_to_dev_addr(struct nvm_dev *dev, 369 struct ppa_addr r) 370{ 371 struct ppa_addr l; 372 373 l.ppa = ((u64)r.g.blk) << dev->ppaf.blk_offset; 374 l.ppa |= ((u64)r.g.pg) << dev->ppaf.pg_offset; 375 l.ppa |= ((u64)r.g.sec) << dev->ppaf.sect_offset; 376 l.ppa |= ((u64)r.g.pl) << dev->ppaf.pln_offset; 377 l.ppa |= ((u64)r.g.lun) << dev->ppaf.lun_offset; 378 l.ppa |= ((u64)r.g.ch) << dev->ppaf.ch_offset; 379 380 return l; 381} 382 383static inline struct ppa_addr dev_to_generic_addr(struct nvm_dev *dev, 384 struct ppa_addr r) 385{ 386 struct ppa_addr l; 387 388 /* 389 * (r.ppa << X offset) & X len bitmask. X eq. blk, pg, etc. 390 */ 391 l.g.blk = (r.ppa >> dev->ppaf.blk_offset) & 392 (((1 << dev->ppaf.blk_len) - 1)); 393 l.g.pg |= (r.ppa >> dev->ppaf.pg_offset) & 394 (((1 << dev->ppaf.pg_len) - 1)); 395 l.g.sec |= (r.ppa >> dev->ppaf.sect_offset) & 396 (((1 << dev->ppaf.sect_len) - 1)); 397 l.g.pl |= (r.ppa >> dev->ppaf.pln_offset) & 398 (((1 << dev->ppaf.pln_len) - 1)); 399 l.g.lun |= (r.ppa >> dev->ppaf.lun_offset) & 400 (((1 << dev->ppaf.lun_len) - 1)); 401 l.g.ch |= (r.ppa >> dev->ppaf.ch_offset) & 402 (((1 << dev->ppaf.ch_len) - 1)); 403 404 return l; 405} 406 407static inline int ppa_empty(struct ppa_addr ppa_addr) 408{ 409 return (ppa_addr.ppa == ADDR_EMPTY); 410} 411 412static inline void ppa_set_empty(struct ppa_addr *ppa_addr) 413{ 414 ppa_addr->ppa = ADDR_EMPTY; 415} 416 417static inline struct ppa_addr block_to_ppa(struct nvm_dev *dev, 418 struct nvm_block *blk) 419{ 420 struct ppa_addr ppa; 421 struct nvm_lun *lun = blk->lun; 422 423 ppa.ppa = 0; 424 ppa.g.blk = blk->id % dev->blks_per_lun; 425 ppa.g.lun = lun->lun_id; 426 ppa.g.ch = lun->chnl_id; 427 428 return ppa; 429} 430 431static inline int ppa_to_slc(struct nvm_dev *dev, int slc_pg) 432{ 433 return dev->lptbl[slc_pg]; 434} 435 436typedef blk_qc_t (nvm_tgt_make_rq_fn)(struct request_queue *, struct bio *); 437typedef sector_t (nvm_tgt_capacity_fn)(void *); 438typedef void *(nvm_tgt_init_fn)(struct nvm_dev *, struct gendisk *, int, int); 439typedef void (nvm_tgt_exit_fn)(void *); 440 441struct nvm_tgt_type { 442 const char *name; 443 unsigned int version[3]; 444 445 /* target entry points */ 446 nvm_tgt_make_rq_fn *make_rq; 447 nvm_tgt_capacity_fn *capacity; 448 nvm_end_io_fn *end_io; 449 450 /* module-specific init/teardown */ 451 nvm_tgt_init_fn *init; 452 nvm_tgt_exit_fn *exit; 453 454 /* For internal use */ 455 struct list_head list; 456}; 457 458extern int nvm_register_tgt_type(struct nvm_tgt_type *); 459extern void nvm_unregister_tgt_type(struct nvm_tgt_type *); 460 461extern void *nvm_dev_dma_alloc(struct nvm_dev *, gfp_t, dma_addr_t *); 462extern void nvm_dev_dma_free(struct nvm_dev *, void *, dma_addr_t); 463 464typedef int (nvmm_register_fn)(struct nvm_dev *); 465typedef void (nvmm_unregister_fn)(struct nvm_dev *); 466typedef struct nvm_block *(nvmm_get_blk_fn)(struct nvm_dev *, 467 struct nvm_lun *, unsigned long); 468typedef void (nvmm_put_blk_fn)(struct nvm_dev *, struct nvm_block *); 469typedef int (nvmm_open_blk_fn)(struct nvm_dev *, struct nvm_block *); 470typedef int (nvmm_close_blk_fn)(struct nvm_dev *, struct nvm_block *); 471typedef void (nvmm_flush_blk_fn)(struct nvm_dev *, struct nvm_block *); 472typedef int (nvmm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *); 473typedef int (nvmm_erase_blk_fn)(struct nvm_dev *, struct nvm_block *, 474 unsigned long); 475typedef void (nvmm_mark_blk_fn)(struct nvm_dev *, struct ppa_addr, int); 476typedef struct nvm_lun *(nvmm_get_lun_fn)(struct nvm_dev *, int); 477typedef int (nvmm_reserve_lun)(struct nvm_dev *, int); 478typedef void (nvmm_release_lun)(struct nvm_dev *, int); 479typedef void (nvmm_lun_info_print_fn)(struct nvm_dev *); 480 481typedef int (nvmm_get_area_fn)(struct nvm_dev *, sector_t *, sector_t); 482typedef void (nvmm_put_area_fn)(struct nvm_dev *, sector_t); 483 484struct nvmm_type { 485 const char *name; 486 unsigned int version[3]; 487 488 nvmm_register_fn *register_mgr; 489 nvmm_unregister_fn *unregister_mgr; 490 491 /* Block administration callbacks */ 492 nvmm_get_blk_fn *get_blk_unlocked; 493 nvmm_put_blk_fn *put_blk_unlocked; 494 nvmm_get_blk_fn *get_blk; 495 nvmm_put_blk_fn *put_blk; 496 nvmm_open_blk_fn *open_blk; 497 nvmm_close_blk_fn *close_blk; 498 nvmm_flush_blk_fn *flush_blk; 499 500 nvmm_submit_io_fn *submit_io; 501 nvmm_erase_blk_fn *erase_blk; 502 503 /* Bad block mgmt */ 504 nvmm_mark_blk_fn *mark_blk; 505 506 /* Configuration management */ 507 nvmm_get_lun_fn *get_lun; 508 nvmm_reserve_lun *reserve_lun; 509 nvmm_release_lun *release_lun; 510 511 /* Statistics */ 512 nvmm_lun_info_print_fn *lun_info_print; 513 514 nvmm_get_area_fn *get_area; 515 nvmm_put_area_fn *put_area; 516 517 struct list_head list; 518}; 519 520extern int nvm_register_mgr(struct nvmm_type *); 521extern void nvm_unregister_mgr(struct nvmm_type *); 522 523extern struct nvm_block *nvm_get_blk_unlocked(struct nvm_dev *, 524 struct nvm_lun *, unsigned long); 525extern void nvm_put_blk_unlocked(struct nvm_dev *, struct nvm_block *); 526 527extern struct nvm_block *nvm_get_blk(struct nvm_dev *, struct nvm_lun *, 528 unsigned long); 529extern void nvm_put_blk(struct nvm_dev *, struct nvm_block *); 530 531extern int nvm_register(struct request_queue *, char *, 532 struct nvm_dev_ops *); 533extern void nvm_unregister(char *); 534 535extern int nvm_submit_io(struct nvm_dev *, struct nvm_rq *); 536extern void nvm_generic_to_addr_mode(struct nvm_dev *, struct nvm_rq *); 537extern void nvm_addr_to_generic_mode(struct nvm_dev *, struct nvm_rq *); 538extern int nvm_set_rqd_ppalist(struct nvm_dev *, struct nvm_rq *, 539 struct ppa_addr *, int, int); 540extern void nvm_free_rqd_ppalist(struct nvm_dev *, struct nvm_rq *); 541extern int nvm_erase_ppa(struct nvm_dev *, struct ppa_addr *, int); 542extern int nvm_erase_blk(struct nvm_dev *, struct nvm_block *); 543extern void nvm_end_io(struct nvm_rq *, int); 544extern int nvm_submit_ppa(struct nvm_dev *, struct ppa_addr *, int, int, int, 545 void *, int); 546extern int nvm_submit_ppa_list(struct nvm_dev *, struct ppa_addr *, int, int, 547 int, void *, int); 548extern int nvm_bb_tbl_fold(struct nvm_dev *, u8 *, int); 549extern int nvm_get_bb_tbl(struct nvm_dev *, struct ppa_addr, u8 *); 550 551/* sysblk.c */ 552#define NVM_SYSBLK_MAGIC 0x4E564D53 /* "NVMS" */ 553 554/* system block on disk representation */ 555struct nvm_system_block { 556 __be32 magic; /* magic signature */ 557 __be32 seqnr; /* sequence number */ 558 __be32 erase_cnt; /* erase count */ 559 __be16 version; /* version number */ 560 u8 mmtype[NVM_MMTYPE_LEN]; /* media manager name */ 561 __be64 fs_ppa; /* PPA for media manager 562 * superblock */ 563}; 564 565extern int nvm_get_sysblock(struct nvm_dev *, struct nvm_sb_info *); 566extern int nvm_update_sysblock(struct nvm_dev *, struct nvm_sb_info *); 567extern int nvm_init_sysblock(struct nvm_dev *, struct nvm_sb_info *); 568 569extern int nvm_dev_factory(struct nvm_dev *, int flags); 570 571#define nvm_for_each_lun_ppa(dev, ppa, chid, lunid) \ 572 for ((chid) = 0, (ppa).ppa = 0; (chid) < (dev)->nr_chnls; \ 573 (chid)++, (ppa).g.ch = (chid)) \ 574 for ((lunid) = 0; (lunid) < (dev)->luns_per_chnl; \ 575 (lunid)++, (ppa).g.lun = (lunid)) 576 577#else /* CONFIG_NVM */ 578struct nvm_dev_ops; 579 580static inline int nvm_register(struct request_queue *q, char *disk_name, 581 struct nvm_dev_ops *ops) 582{ 583 return -EINVAL; 584} 585static inline void nvm_unregister(char *disk_name) {} 586#endif /* CONFIG_NVM */ 587#endif /* LIGHTNVM.H */