at v4.9 14 kB view raw
1#ifndef NVM_H 2#define NVM_H 3 4#include <linux/blkdev.h> 5#include <linux/types.h> 6#include <uapi/linux/lightnvm.h> 7 8enum { 9 NVM_IO_OK = 0, 10 NVM_IO_REQUEUE = 1, 11 NVM_IO_DONE = 2, 12 NVM_IO_ERR = 3, 13 14 NVM_IOTYPE_NONE = 0, 15 NVM_IOTYPE_GC = 1, 16}; 17 18#define NVM_BLK_BITS (16) 19#define NVM_PG_BITS (16) 20#define NVM_SEC_BITS (8) 21#define NVM_PL_BITS (8) 22#define NVM_LUN_BITS (8) 23#define NVM_CH_BITS (7) 24 25struct ppa_addr { 26 /* Generic structure for all addresses */ 27 union { 28 struct { 29 u64 blk : NVM_BLK_BITS; 30 u64 pg : NVM_PG_BITS; 31 u64 sec : NVM_SEC_BITS; 32 u64 pl : NVM_PL_BITS; 33 u64 lun : NVM_LUN_BITS; 34 u64 ch : NVM_CH_BITS; 35 u64 reserved : 1; 36 } g; 37 38 struct { 39 u64 line : 63; 40 u64 is_cached : 1; 41 } c; 42 43 u64 ppa; 44 }; 45}; 46 47struct nvm_rq; 48struct nvm_id; 49struct nvm_dev; 50 51typedef int (nvm_l2p_update_fn)(u64, u32, __le64 *, void *); 52typedef int (nvm_id_fn)(struct nvm_dev *, struct nvm_id *); 53typedef int (nvm_get_l2p_tbl_fn)(struct nvm_dev *, u64, u32, 54 nvm_l2p_update_fn *, void *); 55typedef int (nvm_op_bb_tbl_fn)(struct nvm_dev *, struct ppa_addr, u8 *); 56typedef int (nvm_op_set_bb_fn)(struct nvm_dev *, struct ppa_addr *, int, int); 57typedef int (nvm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *); 58typedef int (nvm_erase_blk_fn)(struct nvm_dev *, struct nvm_rq *); 59typedef void *(nvm_create_dma_pool_fn)(struct nvm_dev *, char *); 60typedef void (nvm_destroy_dma_pool_fn)(void *); 61typedef void *(nvm_dev_dma_alloc_fn)(struct nvm_dev *, void *, gfp_t, 62 dma_addr_t *); 63typedef void (nvm_dev_dma_free_fn)(void *, void*, dma_addr_t); 64 65struct nvm_dev_ops { 66 nvm_id_fn *identity; 67 nvm_get_l2p_tbl_fn *get_l2p_tbl; 68 nvm_op_bb_tbl_fn *get_bb_tbl; 69 nvm_op_set_bb_fn *set_bb_tbl; 70 71 nvm_submit_io_fn *submit_io; 72 nvm_erase_blk_fn *erase_block; 73 74 nvm_create_dma_pool_fn *create_dma_pool; 75 nvm_destroy_dma_pool_fn *destroy_dma_pool; 76 nvm_dev_dma_alloc_fn *dev_dma_alloc; 77 nvm_dev_dma_free_fn *dev_dma_free; 78 79 unsigned int max_phys_sect; 80}; 81 82 83 84#ifdef CONFIG_NVM 85 86#include <linux/blkdev.h> 87#include <linux/file.h> 88#include <linux/dmapool.h> 89#include <uapi/linux/lightnvm.h> 90 91enum { 92 /* HW Responsibilities */ 93 NVM_RSP_L2P = 1 << 0, 94 NVM_RSP_ECC = 1 << 1, 95 96 /* Physical Adressing Mode */ 97 NVM_ADDRMODE_LINEAR = 0, 98 NVM_ADDRMODE_CHANNEL = 1, 99 100 /* Plane programming mode for LUN */ 101 NVM_PLANE_SINGLE = 1, 102 NVM_PLANE_DOUBLE = 2, 103 NVM_PLANE_QUAD = 4, 104 105 /* Status codes */ 106 NVM_RSP_SUCCESS = 0x0, 107 NVM_RSP_NOT_CHANGEABLE = 0x1, 108 NVM_RSP_ERR_FAILWRITE = 0x40ff, 109 NVM_RSP_ERR_EMPTYPAGE = 0x42ff, 110 111 /* Device opcodes */ 112 NVM_OP_HBREAD = 0x02, 113 NVM_OP_HBWRITE = 0x81, 114 NVM_OP_PWRITE = 0x91, 115 NVM_OP_PREAD = 0x92, 116 NVM_OP_ERASE = 0x90, 117 118 /* PPA Command Flags */ 119 NVM_IO_SNGL_ACCESS = 0x0, 120 NVM_IO_DUAL_ACCESS = 0x1, 121 NVM_IO_QUAD_ACCESS = 0x2, 122 123 /* NAND Access Modes */ 124 NVM_IO_SUSPEND = 0x80, 125 NVM_IO_SLC_MODE = 0x100, 126 NVM_IO_SCRAMBLE_DISABLE = 0x200, 127 128 /* Block Types */ 129 NVM_BLK_T_FREE = 0x0, 130 NVM_BLK_T_BAD = 0x1, 131 NVM_BLK_T_GRWN_BAD = 0x2, 132 NVM_BLK_T_DEV = 0x4, 133 NVM_BLK_T_HOST = 0x8, 134 135 /* Memory capabilities */ 136 NVM_ID_CAP_SLC = 0x1, 137 NVM_ID_CAP_CMD_SUSPEND = 0x2, 138 NVM_ID_CAP_SCRAMBLE = 0x4, 139 NVM_ID_CAP_ENCRYPT = 0x8, 140 141 /* Memory types */ 142 NVM_ID_FMTYPE_SLC = 0, 143 NVM_ID_FMTYPE_MLC = 1, 144 145 /* Device capabilities */ 146 NVM_ID_DCAP_BBLKMGMT = 0x1, 147 NVM_UD_DCAP_ECC = 0x2, 148}; 149 150struct nvm_id_lp_mlc { 151 u16 num_pairs; 152 u8 pairs[886]; 153}; 154 155struct nvm_id_lp_tbl { 156 __u8 id[8]; 157 struct nvm_id_lp_mlc mlc; 158}; 159 160struct nvm_id_group { 161 u8 mtype; 162 u8 fmtype; 163 u8 num_ch; 164 u8 num_lun; 165 u8 num_pln; 166 u16 num_blk; 167 u16 num_pg; 168 u16 fpg_sz; 169 u16 csecs; 170 u16 sos; 171 u32 trdt; 172 u32 trdm; 173 u32 tprt; 174 u32 tprm; 175 u32 tbet; 176 u32 tbem; 177 u32 mpos; 178 u32 mccap; 179 u16 cpar; 180 181 struct nvm_id_lp_tbl lptbl; 182}; 183 184struct nvm_addr_format { 185 u8 ch_offset; 186 u8 ch_len; 187 u8 lun_offset; 188 u8 lun_len; 189 u8 pln_offset; 190 u8 pln_len; 191 u8 blk_offset; 192 u8 blk_len; 193 u8 pg_offset; 194 u8 pg_len; 195 u8 sect_offset; 196 u8 sect_len; 197}; 198 199struct nvm_id { 200 u8 ver_id; 201 u8 vmnt; 202 u8 cgrps; 203 u32 cap; 204 u32 dom; 205 struct nvm_addr_format ppaf; 206 struct nvm_id_group groups[4]; 207} __packed; 208 209struct nvm_target { 210 struct list_head list; 211 struct nvm_dev *dev; 212 struct nvm_tgt_type *type; 213 struct gendisk *disk; 214}; 215 216struct nvm_tgt_instance { 217 struct nvm_tgt_type *tt; 218}; 219 220#define ADDR_EMPTY (~0ULL) 221 222#define NVM_VERSION_MAJOR 1 223#define NVM_VERSION_MINOR 0 224#define NVM_VERSION_PATCH 0 225 226struct nvm_rq; 227typedef void (nvm_end_io_fn)(struct nvm_rq *); 228 229struct nvm_rq { 230 struct nvm_tgt_instance *ins; 231 struct nvm_dev *dev; 232 233 struct bio *bio; 234 235 union { 236 struct ppa_addr ppa_addr; 237 dma_addr_t dma_ppa_list; 238 }; 239 240 struct ppa_addr *ppa_list; 241 242 void *meta_list; 243 dma_addr_t dma_meta_list; 244 245 struct completion *wait; 246 nvm_end_io_fn *end_io; 247 248 uint8_t opcode; 249 uint16_t nr_ppas; 250 uint16_t flags; 251 252 u64 ppa_status; /* ppa media status */ 253 int error; 254}; 255 256static inline struct nvm_rq *nvm_rq_from_pdu(void *pdu) 257{ 258 return pdu - sizeof(struct nvm_rq); 259} 260 261static inline void *nvm_rq_to_pdu(struct nvm_rq *rqdata) 262{ 263 return rqdata + 1; 264} 265 266struct nvm_block; 267 268struct nvm_lun { 269 int id; 270 271 int lun_id; 272 int chnl_id; 273 274 spinlock_t lock; 275 276 unsigned int nr_free_blocks; /* Number of unused blocks */ 277 struct nvm_block *blocks; 278}; 279 280enum { 281 NVM_BLK_ST_FREE = 0x1, /* Free block */ 282 NVM_BLK_ST_TGT = 0x2, /* Block in use by target */ 283 NVM_BLK_ST_BAD = 0x8, /* Bad block */ 284}; 285 286struct nvm_block { 287 struct list_head list; 288 struct nvm_lun *lun; 289 unsigned long id; 290 291 void *priv; 292 int state; 293}; 294 295/* system block cpu representation */ 296struct nvm_sb_info { 297 unsigned long seqnr; 298 unsigned long erase_cnt; 299 unsigned int version; 300 char mmtype[NVM_MMTYPE_LEN]; 301 struct ppa_addr fs_ppa; 302}; 303 304struct nvm_dev { 305 struct nvm_dev_ops *ops; 306 307 struct list_head devices; 308 309 /* Media manager */ 310 struct nvmm_type *mt; 311 void *mp; 312 313 /* System blocks */ 314 struct nvm_sb_info sb; 315 316 /* Device information */ 317 int nr_chnls; 318 int nr_planes; 319 int luns_per_chnl; 320 int sec_per_pg; /* only sectors for a single page */ 321 int pgs_per_blk; 322 int blks_per_lun; 323 int fpg_size; 324 int pfpg_size; /* size of buffer if all pages are to be read */ 325 int sec_size; 326 int oob_size; 327 int mccap; 328 struct nvm_addr_format ppaf; 329 330 /* Calculated/Cached values. These do not reflect the actual usable 331 * blocks at run-time. 332 */ 333 int max_rq_size; 334 int plane_mode; /* drive device in single, double or quad mode */ 335 336 int sec_per_pl; /* all sectors across planes */ 337 int sec_per_blk; 338 int sec_per_lun; 339 340 /* lower page table */ 341 int lps_per_blk; 342 int *lptbl; 343 344 unsigned long total_blocks; 345 unsigned long total_secs; 346 int nr_luns; 347 348 unsigned long *lun_map; 349 void *dma_pool; 350 351 struct nvm_id identity; 352 353 /* Backend device */ 354 struct request_queue *q; 355 struct device dev; 356 struct device *parent_dev; 357 char name[DISK_NAME_LEN]; 358 void *private_data; 359 360 struct mutex mlock; 361 spinlock_t lock; 362}; 363 364static inline struct ppa_addr generic_to_dev_addr(struct nvm_dev *dev, 365 struct ppa_addr r) 366{ 367 struct ppa_addr l; 368 369 l.ppa = ((u64)r.g.blk) << dev->ppaf.blk_offset; 370 l.ppa |= ((u64)r.g.pg) << dev->ppaf.pg_offset; 371 l.ppa |= ((u64)r.g.sec) << dev->ppaf.sect_offset; 372 l.ppa |= ((u64)r.g.pl) << dev->ppaf.pln_offset; 373 l.ppa |= ((u64)r.g.lun) << dev->ppaf.lun_offset; 374 l.ppa |= ((u64)r.g.ch) << dev->ppaf.ch_offset; 375 376 return l; 377} 378 379static inline struct ppa_addr dev_to_generic_addr(struct nvm_dev *dev, 380 struct ppa_addr r) 381{ 382 struct ppa_addr l; 383 384 l.ppa = 0; 385 /* 386 * (r.ppa << X offset) & X len bitmask. X eq. blk, pg, etc. 387 */ 388 l.g.blk = (r.ppa >> dev->ppaf.blk_offset) & 389 (((1 << dev->ppaf.blk_len) - 1)); 390 l.g.pg |= (r.ppa >> dev->ppaf.pg_offset) & 391 (((1 << dev->ppaf.pg_len) - 1)); 392 l.g.sec |= (r.ppa >> dev->ppaf.sect_offset) & 393 (((1 << dev->ppaf.sect_len) - 1)); 394 l.g.pl |= (r.ppa >> dev->ppaf.pln_offset) & 395 (((1 << dev->ppaf.pln_len) - 1)); 396 l.g.lun |= (r.ppa >> dev->ppaf.lun_offset) & 397 (((1 << dev->ppaf.lun_len) - 1)); 398 l.g.ch |= (r.ppa >> dev->ppaf.ch_offset) & 399 (((1 << dev->ppaf.ch_len) - 1)); 400 401 return l; 402} 403 404static inline int ppa_empty(struct ppa_addr ppa_addr) 405{ 406 return (ppa_addr.ppa == ADDR_EMPTY); 407} 408 409static inline void ppa_set_empty(struct ppa_addr *ppa_addr) 410{ 411 ppa_addr->ppa = ADDR_EMPTY; 412} 413 414static inline struct ppa_addr block_to_ppa(struct nvm_dev *dev, 415 struct nvm_block *blk) 416{ 417 struct ppa_addr ppa; 418 struct nvm_lun *lun = blk->lun; 419 420 ppa.ppa = 0; 421 ppa.g.blk = blk->id % dev->blks_per_lun; 422 ppa.g.lun = lun->lun_id; 423 ppa.g.ch = lun->chnl_id; 424 425 return ppa; 426} 427 428static inline int ppa_to_slc(struct nvm_dev *dev, int slc_pg) 429{ 430 return dev->lptbl[slc_pg]; 431} 432 433typedef blk_qc_t (nvm_tgt_make_rq_fn)(struct request_queue *, struct bio *); 434typedef sector_t (nvm_tgt_capacity_fn)(void *); 435typedef void *(nvm_tgt_init_fn)(struct nvm_dev *, struct gendisk *, int, int); 436typedef void (nvm_tgt_exit_fn)(void *); 437 438struct nvm_tgt_type { 439 const char *name; 440 unsigned int version[3]; 441 442 /* target entry points */ 443 nvm_tgt_make_rq_fn *make_rq; 444 nvm_tgt_capacity_fn *capacity; 445 nvm_end_io_fn *end_io; 446 447 /* module-specific init/teardown */ 448 nvm_tgt_init_fn *init; 449 nvm_tgt_exit_fn *exit; 450 451 /* For internal use */ 452 struct list_head list; 453}; 454 455extern struct nvm_tgt_type *nvm_find_target_type(const char *, int); 456 457extern int nvm_register_tgt_type(struct nvm_tgt_type *); 458extern void nvm_unregister_tgt_type(struct nvm_tgt_type *); 459 460extern void *nvm_dev_dma_alloc(struct nvm_dev *, gfp_t, dma_addr_t *); 461extern void nvm_dev_dma_free(struct nvm_dev *, void *, dma_addr_t); 462 463typedef int (nvmm_register_fn)(struct nvm_dev *); 464typedef void (nvmm_unregister_fn)(struct nvm_dev *); 465 466typedef int (nvmm_create_tgt_fn)(struct nvm_dev *, struct nvm_ioctl_create *); 467typedef int (nvmm_remove_tgt_fn)(struct nvm_dev *, struct nvm_ioctl_remove *); 468typedef struct nvm_block *(nvmm_get_blk_fn)(struct nvm_dev *, 469 struct nvm_lun *, unsigned long); 470typedef void (nvmm_put_blk_fn)(struct nvm_dev *, struct nvm_block *); 471typedef int (nvmm_open_blk_fn)(struct nvm_dev *, struct nvm_block *); 472typedef int (nvmm_close_blk_fn)(struct nvm_dev *, struct nvm_block *); 473typedef void (nvmm_flush_blk_fn)(struct nvm_dev *, struct nvm_block *); 474typedef int (nvmm_submit_io_fn)(struct nvm_dev *, struct nvm_rq *); 475typedef int (nvmm_erase_blk_fn)(struct nvm_dev *, struct nvm_block *, 476 unsigned long); 477typedef void (nvmm_mark_blk_fn)(struct nvm_dev *, struct ppa_addr, int); 478typedef struct nvm_lun *(nvmm_get_lun_fn)(struct nvm_dev *, int); 479typedef int (nvmm_reserve_lun)(struct nvm_dev *, int); 480typedef void (nvmm_release_lun)(struct nvm_dev *, int); 481typedef void (nvmm_lun_info_print_fn)(struct nvm_dev *); 482 483typedef int (nvmm_get_area_fn)(struct nvm_dev *, sector_t *, sector_t); 484typedef void (nvmm_put_area_fn)(struct nvm_dev *, sector_t); 485 486struct nvmm_type { 487 const char *name; 488 unsigned int version[3]; 489 490 nvmm_register_fn *register_mgr; 491 nvmm_unregister_fn *unregister_mgr; 492 493 nvmm_create_tgt_fn *create_tgt; 494 nvmm_remove_tgt_fn *remove_tgt; 495 496 /* Block administration callbacks */ 497 nvmm_get_blk_fn *get_blk; 498 nvmm_put_blk_fn *put_blk; 499 nvmm_open_blk_fn *open_blk; 500 nvmm_close_blk_fn *close_blk; 501 nvmm_flush_blk_fn *flush_blk; 502 503 nvmm_submit_io_fn *submit_io; 504 nvmm_erase_blk_fn *erase_blk; 505 506 /* Bad block mgmt */ 507 nvmm_mark_blk_fn *mark_blk; 508 509 /* Configuration management */ 510 nvmm_get_lun_fn *get_lun; 511 nvmm_reserve_lun *reserve_lun; 512 nvmm_release_lun *release_lun; 513 514 /* Statistics */ 515 nvmm_lun_info_print_fn *lun_info_print; 516 517 nvmm_get_area_fn *get_area; 518 nvmm_put_area_fn *put_area; 519 520 struct list_head list; 521}; 522 523extern int nvm_register_mgr(struct nvmm_type *); 524extern void nvm_unregister_mgr(struct nvmm_type *); 525 526extern struct nvm_block *nvm_get_blk(struct nvm_dev *, struct nvm_lun *, 527 unsigned long); 528extern void nvm_put_blk(struct nvm_dev *, struct nvm_block *); 529 530extern struct nvm_dev *nvm_alloc_dev(int); 531extern int nvm_register(struct nvm_dev *); 532extern void nvm_unregister(struct nvm_dev *); 533 534void nvm_mark_blk(struct nvm_dev *dev, struct ppa_addr ppa, int type); 535 536extern int nvm_submit_io(struct nvm_dev *, struct nvm_rq *); 537extern void nvm_generic_to_addr_mode(struct nvm_dev *, struct nvm_rq *); 538extern void nvm_addr_to_generic_mode(struct nvm_dev *, struct nvm_rq *); 539extern int nvm_set_rqd_ppalist(struct nvm_dev *, struct nvm_rq *, 540 const struct ppa_addr *, int, int); 541extern void nvm_free_rqd_ppalist(struct nvm_dev *, struct nvm_rq *); 542extern int nvm_erase_ppa(struct nvm_dev *, struct ppa_addr *, int); 543extern int nvm_erase_blk(struct nvm_dev *, struct nvm_block *); 544extern void nvm_end_io(struct nvm_rq *, int); 545extern int nvm_submit_ppa(struct nvm_dev *, struct ppa_addr *, int, int, int, 546 void *, int); 547extern int nvm_submit_ppa_list(struct nvm_dev *, struct ppa_addr *, int, int, 548 int, void *, int); 549extern int nvm_bb_tbl_fold(struct nvm_dev *, u8 *, int); 550extern int nvm_get_bb_tbl(struct nvm_dev *, struct ppa_addr, u8 *); 551 552/* sysblk.c */ 553#define NVM_SYSBLK_MAGIC 0x4E564D53 /* "NVMS" */ 554 555/* system block on disk representation */ 556struct nvm_system_block { 557 __be32 magic; /* magic signature */ 558 __be32 seqnr; /* sequence number */ 559 __be32 erase_cnt; /* erase count */ 560 __be16 version; /* version number */ 561 u8 mmtype[NVM_MMTYPE_LEN]; /* media manager name */ 562 __be64 fs_ppa; /* PPA for media manager 563 * superblock */ 564}; 565 566extern int nvm_get_sysblock(struct nvm_dev *, struct nvm_sb_info *); 567extern int nvm_update_sysblock(struct nvm_dev *, struct nvm_sb_info *); 568extern int nvm_init_sysblock(struct nvm_dev *, struct nvm_sb_info *); 569 570extern int nvm_dev_factory(struct nvm_dev *, int flags); 571 572#define nvm_for_each_lun_ppa(dev, ppa, chid, lunid) \ 573 for ((chid) = 0, (ppa).ppa = 0; (chid) < (dev)->nr_chnls; \ 574 (chid)++, (ppa).g.ch = (chid)) \ 575 for ((lunid) = 0; (lunid) < (dev)->luns_per_chnl; \ 576 (lunid)++, (ppa).g.lun = (lunid)) 577 578#else /* CONFIG_NVM */ 579struct nvm_dev_ops; 580 581static inline struct nvm_dev *nvm_alloc_dev(int node) 582{ 583 return ERR_PTR(-EINVAL); 584} 585static inline int nvm_register(struct nvm_dev *dev) 586{ 587 return -EINVAL; 588} 589static inline void nvm_unregister(struct nvm_dev *dev) {} 590#endif /* CONFIG_NVM */ 591#endif /* LIGHTNVM.H */