at v5.7 48 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _IDE_H 3#define _IDE_H 4/* 5 * linux/include/linux/ide.h 6 * 7 * Copyright (C) 1994-2002 Linus Torvalds & authors 8 */ 9 10#include <linux/init.h> 11#include <linux/ioport.h> 12#include <linux/ata.h> 13#include <linux/blk-mq.h> 14#include <linux/proc_fs.h> 15#include <linux/interrupt.h> 16#include <linux/bitops.h> 17#include <linux/bio.h> 18#include <linux/pci.h> 19#include <linux/completion.h> 20#include <linux/pm.h> 21#include <linux/mutex.h> 22/* for request_sense */ 23#include <linux/cdrom.h> 24#include <scsi/scsi_cmnd.h> 25#include <asm/byteorder.h> 26#include <asm/io.h> 27 28/* 29 * Probably not wise to fiddle with these 30 */ 31#define SUPPORT_VLB_SYNC 1 32#define IDE_DEFAULT_MAX_FAILURES 1 33#define ERROR_MAX 8 /* Max read/write errors per sector */ 34#define ERROR_RESET 3 /* Reset controller every 4th retry */ 35#define ERROR_RECAL 1 /* Recalibrate every 2nd retry */ 36 37struct device; 38 39/* values for ide_request.type */ 40enum ata_priv_type { 41 ATA_PRIV_MISC, 42 ATA_PRIV_TASKFILE, 43 ATA_PRIV_PC, 44 ATA_PRIV_SENSE, /* sense request */ 45 ATA_PRIV_PM_SUSPEND, /* suspend request */ 46 ATA_PRIV_PM_RESUME, /* resume request */ 47}; 48 49struct ide_request { 50 struct scsi_request sreq; 51 u8 sense[SCSI_SENSE_BUFFERSIZE]; 52 u8 type; 53 void *special; 54}; 55 56static inline struct ide_request *ide_req(struct request *rq) 57{ 58 return blk_mq_rq_to_pdu(rq); 59} 60 61static inline bool ata_misc_request(struct request *rq) 62{ 63 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_MISC; 64} 65 66static inline bool ata_taskfile_request(struct request *rq) 67{ 68 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_TASKFILE; 69} 70 71static inline bool ata_pc_request(struct request *rq) 72{ 73 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_PC; 74} 75 76static inline bool ata_sense_request(struct request *rq) 77{ 78 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_SENSE; 79} 80 81static inline bool ata_pm_request(struct request *rq) 82{ 83 return blk_rq_is_private(rq) && 84 (ide_req(rq)->type == ATA_PRIV_PM_SUSPEND || 85 ide_req(rq)->type == ATA_PRIV_PM_RESUME); 86} 87 88/* Error codes returned in result to the higher part of the driver. */ 89enum { 90 IDE_DRV_ERROR_GENERAL = 101, 91 IDE_DRV_ERROR_FILEMARK = 102, 92 IDE_DRV_ERROR_EOD = 103, 93}; 94 95/* 96 * Definitions for accessing IDE controller registers 97 */ 98#define IDE_NR_PORTS (10) 99 100struct ide_io_ports { 101 unsigned long data_addr; 102 103 union { 104 unsigned long error_addr; /* read: error */ 105 unsigned long feature_addr; /* write: feature */ 106 }; 107 108 unsigned long nsect_addr; 109 unsigned long lbal_addr; 110 unsigned long lbam_addr; 111 unsigned long lbah_addr; 112 113 unsigned long device_addr; 114 115 union { 116 unsigned long status_addr; /*  read: status  */ 117 unsigned long command_addr; /* write: command */ 118 }; 119 120 unsigned long ctl_addr; 121 122 unsigned long irq_addr; 123}; 124 125#define OK_STAT(stat,good,bad) (((stat)&((good)|(bad)))==(good)) 126 127#define BAD_R_STAT (ATA_BUSY | ATA_ERR) 128#define BAD_W_STAT (BAD_R_STAT | ATA_DF) 129#define BAD_STAT (BAD_R_STAT | ATA_DRQ) 130#define DRIVE_READY (ATA_DRDY | ATA_DSC) 131 132#define BAD_CRC (ATA_ABORTED | ATA_ICRC) 133 134#define SATA_NR_PORTS (3) /* 16 possible ?? */ 135 136#define SATA_STATUS_OFFSET (0) 137#define SATA_ERROR_OFFSET (1) 138#define SATA_CONTROL_OFFSET (2) 139 140/* 141 * Our Physical Region Descriptor (PRD) table should be large enough 142 * to handle the biggest I/O request we are likely to see. Since requests 143 * can have no more than 256 sectors, and since the typical blocksize is 144 * two or more sectors, we could get by with a limit of 128 entries here for 145 * the usual worst case. Most requests seem to include some contiguous blocks, 146 * further reducing the number of table entries required. 147 * 148 * The driver reverts to PIO mode for individual requests that exceed 149 * this limit (possible with 512 byte blocksizes, eg. MSDOS f/s), so handling 150 * 100% of all crazy scenarios here is not necessary. 151 * 152 * As it turns out though, we must allocate a full 4KB page for this, 153 * so the two PRD tables (ide0 & ide1) will each get half of that, 154 * allowing each to have about 256 entries (8 bytes each) from this. 155 */ 156#define PRD_BYTES 8 157#define PRD_ENTRIES 256 158 159/* 160 * Some more useful definitions 161 */ 162#define PARTN_BITS 6 /* number of minor dev bits for partitions */ 163#define MAX_DRIVES 2 /* per interface; 2 assumed by lots of code */ 164 165/* 166 * Timeouts for various operations: 167 */ 168enum { 169 /* spec allows up to 20ms, but CF cards and SSD drives need more */ 170 WAIT_DRQ = 1 * HZ, /* 1s */ 171 /* some laptops are very slow */ 172 WAIT_READY = 5 * HZ, /* 5s */ 173 /* should be less than 3ms (?), if all ATAPI CD is closed at boot */ 174 WAIT_PIDENTIFY = 10 * HZ, /* 10s */ 175 /* worst case when spinning up */ 176 WAIT_WORSTCASE = 30 * HZ, /* 30s */ 177 /* maximum wait for an IRQ to happen */ 178 WAIT_CMD = 10 * HZ, /* 10s */ 179 /* Some drives require a longer IRQ timeout. */ 180 WAIT_FLOPPY_CMD = 50 * HZ, /* 50s */ 181 /* 182 * Some drives (for example, Seagate STT3401A Travan) require a very 183 * long timeout, because they don't return an interrupt or clear their 184 * BSY bit until after the command completes (even retension commands). 185 */ 186 WAIT_TAPE_CMD = 900 * HZ, /* 900s */ 187 /* minimum sleep time */ 188 WAIT_MIN_SLEEP = HZ / 50, /* 20ms */ 189}; 190 191/* 192 * Op codes for special requests to be handled by ide_special_rq(). 193 * Values should be in the range of 0x20 to 0x3f. 194 */ 195#define REQ_DRIVE_RESET 0x20 196#define REQ_DEVSET_EXEC 0x21 197#define REQ_PARK_HEADS 0x22 198#define REQ_UNPARK_HEADS 0x23 199 200/* 201 * hwif_chipset_t is used to keep track of the specific hardware 202 * chipset used by each IDE interface, if known. 203 */ 204enum { ide_unknown, ide_generic, ide_pci, 205 ide_cmd640, ide_dtc2278, ide_ali14xx, 206 ide_qd65xx, ide_umc8672, ide_ht6560b, 207 ide_4drives, ide_pmac, ide_acorn, 208 ide_au1xxx, ide_palm3710 209}; 210 211typedef u8 hwif_chipset_t; 212 213/* 214 * Structure to hold all information about the location of this port 215 */ 216struct ide_hw { 217 union { 218 struct ide_io_ports io_ports; 219 unsigned long io_ports_array[IDE_NR_PORTS]; 220 }; 221 222 int irq; /* our irq number */ 223 struct device *dev, *parent; 224 unsigned long config; 225}; 226 227static inline void ide_std_init_ports(struct ide_hw *hw, 228 unsigned long io_addr, 229 unsigned long ctl_addr) 230{ 231 unsigned int i; 232 233 for (i = 0; i <= 7; i++) 234 hw->io_ports_array[i] = io_addr++; 235 236 hw->io_ports.ctl_addr = ctl_addr; 237} 238 239#define MAX_HWIFS 10 240 241/* 242 * Now for the data we need to maintain per-drive: ide_drive_t 243 */ 244 245#define ide_scsi 0x21 246#define ide_disk 0x20 247#define ide_optical 0x7 248#define ide_cdrom 0x5 249#define ide_tape 0x1 250#define ide_floppy 0x0 251 252/* 253 * Special Driver Flags 254 */ 255enum { 256 IDE_SFLAG_SET_GEOMETRY = BIT(0), 257 IDE_SFLAG_RECALIBRATE = BIT(1), 258 IDE_SFLAG_SET_MULTMODE = BIT(2), 259}; 260 261/* 262 * Status returned from various ide_ functions 263 */ 264typedef enum { 265 ide_stopped, /* no drive operation was started */ 266 ide_started, /* a drive operation was started, handler was set */ 267} ide_startstop_t; 268 269enum { 270 IDE_VALID_ERROR = BIT(1), 271 IDE_VALID_FEATURE = IDE_VALID_ERROR, 272 IDE_VALID_NSECT = BIT(2), 273 IDE_VALID_LBAL = BIT(3), 274 IDE_VALID_LBAM = BIT(4), 275 IDE_VALID_LBAH = BIT(5), 276 IDE_VALID_DEVICE = BIT(6), 277 IDE_VALID_LBA = IDE_VALID_LBAL | 278 IDE_VALID_LBAM | 279 IDE_VALID_LBAH, 280 IDE_VALID_OUT_TF = IDE_VALID_FEATURE | 281 IDE_VALID_NSECT | 282 IDE_VALID_LBA, 283 IDE_VALID_IN_TF = IDE_VALID_NSECT | 284 IDE_VALID_LBA, 285 IDE_VALID_OUT_HOB = IDE_VALID_OUT_TF, 286 IDE_VALID_IN_HOB = IDE_VALID_ERROR | 287 IDE_VALID_NSECT | 288 IDE_VALID_LBA, 289}; 290 291enum { 292 IDE_TFLAG_LBA48 = BIT(0), 293 IDE_TFLAG_WRITE = BIT(1), 294 IDE_TFLAG_CUSTOM_HANDLER = BIT(2), 295 IDE_TFLAG_DMA_PIO_FALLBACK = BIT(3), 296 /* force 16-bit I/O operations */ 297 IDE_TFLAG_IO_16BIT = BIT(4), 298 /* struct ide_cmd was allocated using kmalloc() */ 299 IDE_TFLAG_DYN = BIT(5), 300 IDE_TFLAG_FS = BIT(6), 301 IDE_TFLAG_MULTI_PIO = BIT(7), 302 IDE_TFLAG_SET_XFER = BIT(8), 303}; 304 305enum { 306 IDE_FTFLAG_FLAGGED = BIT(0), 307 IDE_FTFLAG_SET_IN_FLAGS = BIT(1), 308 IDE_FTFLAG_OUT_DATA = BIT(2), 309 IDE_FTFLAG_IN_DATA = BIT(3), 310}; 311 312struct ide_taskfile { 313 u8 data; /* 0: data byte (for TASKFILE ioctl) */ 314 union { /* 1: */ 315 u8 error; /* read: error */ 316 u8 feature; /* write: feature */ 317 }; 318 u8 nsect; /* 2: number of sectors */ 319 u8 lbal; /* 3: LBA low */ 320 u8 lbam; /* 4: LBA mid */ 321 u8 lbah; /* 5: LBA high */ 322 u8 device; /* 6: device select */ 323 union { /* 7: */ 324 u8 status; /* read: status */ 325 u8 command; /* write: command */ 326 }; 327}; 328 329struct ide_cmd { 330 struct ide_taskfile tf; 331 struct ide_taskfile hob; 332 struct { 333 struct { 334 u8 tf; 335 u8 hob; 336 } out, in; 337 } valid; 338 339 u16 tf_flags; 340 u8 ftf_flags; /* for TASKFILE ioctl */ 341 int protocol; 342 343 int sg_nents; /* number of sg entries */ 344 int orig_sg_nents; 345 int sg_dma_direction; /* DMA transfer direction */ 346 347 unsigned int nbytes; 348 unsigned int nleft; 349 unsigned int last_xfer_len; 350 351 struct scatterlist *cursg; 352 unsigned int cursg_ofs; 353 354 struct request *rq; /* copy of request */ 355}; 356 357/* ATAPI packet command flags */ 358enum { 359 /* set when an error is considered normal - no retry (ide-tape) */ 360 PC_FLAG_ABORT = BIT(0), 361 PC_FLAG_SUPPRESS_ERROR = BIT(1), 362 PC_FLAG_WAIT_FOR_DSC = BIT(2), 363 PC_FLAG_DMA_OK = BIT(3), 364 PC_FLAG_DMA_IN_PROGRESS = BIT(4), 365 PC_FLAG_DMA_ERROR = BIT(5), 366 PC_FLAG_WRITING = BIT(6), 367}; 368 369#define ATAPI_WAIT_PC (60 * HZ) 370 371struct ide_atapi_pc { 372 /* actual packet bytes */ 373 u8 c[12]; 374 /* incremented on each retry */ 375 int retries; 376 int error; 377 378 /* bytes to transfer */ 379 int req_xfer; 380 381 /* the corresponding request */ 382 struct request *rq; 383 384 unsigned long flags; 385 386 /* 387 * those are more or less driver-specific and some of them are subject 388 * to change/removal later. 389 */ 390 unsigned long timeout; 391}; 392 393struct ide_devset; 394struct ide_driver; 395 396#ifdef CONFIG_BLK_DEV_IDEACPI 397struct ide_acpi_drive_link; 398struct ide_acpi_hwif_link; 399#endif 400 401struct ide_drive_s; 402 403struct ide_disk_ops { 404 int (*check)(struct ide_drive_s *, const char *); 405 int (*get_capacity)(struct ide_drive_s *); 406 void (*unlock_native_capacity)(struct ide_drive_s *); 407 void (*setup)(struct ide_drive_s *); 408 void (*flush)(struct ide_drive_s *); 409 int (*init_media)(struct ide_drive_s *, struct gendisk *); 410 int (*set_doorlock)(struct ide_drive_s *, struct gendisk *, 411 int); 412 ide_startstop_t (*do_request)(struct ide_drive_s *, struct request *, 413 sector_t); 414 int (*ioctl)(struct ide_drive_s *, struct block_device *, 415 fmode_t, unsigned int, unsigned long); 416 int (*compat_ioctl)(struct ide_drive_s *, struct block_device *, 417 fmode_t, unsigned int, unsigned long); 418}; 419 420/* ATAPI device flags */ 421enum { 422 IDE_AFLAG_DRQ_INTERRUPT = BIT(0), 423 424 /* ide-cd */ 425 /* Drive cannot eject the disc. */ 426 IDE_AFLAG_NO_EJECT = BIT(1), 427 /* Drive is a pre ATAPI 1.2 drive. */ 428 IDE_AFLAG_PRE_ATAPI12 = BIT(2), 429 /* TOC addresses are in BCD. */ 430 IDE_AFLAG_TOCADDR_AS_BCD = BIT(3), 431 /* TOC track numbers are in BCD. */ 432 IDE_AFLAG_TOCTRACKS_AS_BCD = BIT(4), 433 /* Saved TOC information is current. */ 434 IDE_AFLAG_TOC_VALID = BIT(6), 435 /* We think that the drive door is locked. */ 436 IDE_AFLAG_DOOR_LOCKED = BIT(7), 437 /* SET_CD_SPEED command is unsupported. */ 438 IDE_AFLAG_NO_SPEED_SELECT = BIT(8), 439 IDE_AFLAG_VERTOS_300_SSD = BIT(9), 440 IDE_AFLAG_VERTOS_600_ESD = BIT(10), 441 IDE_AFLAG_SANYO_3CD = BIT(11), 442 IDE_AFLAG_FULL_CAPS_PAGE = BIT(12), 443 IDE_AFLAG_PLAY_AUDIO_OK = BIT(13), 444 IDE_AFLAG_LE_SPEED_FIELDS = BIT(14), 445 446 /* ide-floppy */ 447 /* Avoid commands not supported in Clik drive */ 448 IDE_AFLAG_CLIK_DRIVE = BIT(15), 449 /* Requires BH algorithm for packets */ 450 IDE_AFLAG_ZIP_DRIVE = BIT(16), 451 /* Supports format progress report */ 452 IDE_AFLAG_SRFP = BIT(17), 453 454 /* ide-tape */ 455 IDE_AFLAG_IGNORE_DSC = BIT(18), 456 /* 0 When the tape position is unknown */ 457 IDE_AFLAG_ADDRESS_VALID = BIT(19), 458 /* Device already opened */ 459 IDE_AFLAG_BUSY = BIT(20), 460 /* Attempt to auto-detect the current user block size */ 461 IDE_AFLAG_DETECT_BS = BIT(21), 462 /* Currently on a filemark */ 463 IDE_AFLAG_FILEMARK = BIT(22), 464 /* 0 = no tape is loaded, so we don't rewind after ejecting */ 465 IDE_AFLAG_MEDIUM_PRESENT = BIT(23), 466 467 IDE_AFLAG_NO_AUTOCLOSE = BIT(24), 468}; 469 470/* device flags */ 471enum { 472 /* restore settings after device reset */ 473 IDE_DFLAG_KEEP_SETTINGS = BIT(0), 474 /* device is using DMA for read/write */ 475 IDE_DFLAG_USING_DMA = BIT(1), 476 /* okay to unmask other IRQs */ 477 IDE_DFLAG_UNMASK = BIT(2), 478 /* don't attempt flushes */ 479 IDE_DFLAG_NOFLUSH = BIT(3), 480 /* DSC overlap */ 481 IDE_DFLAG_DSC_OVERLAP = BIT(4), 482 /* give potential excess bandwidth */ 483 IDE_DFLAG_NICE1 = BIT(5), 484 /* device is physically present */ 485 IDE_DFLAG_PRESENT = BIT(6), 486 /* disable Host Protected Area */ 487 IDE_DFLAG_NOHPA = BIT(7), 488 /* id read from device (synthetic if not set) */ 489 IDE_DFLAG_ID_READ = BIT(8), 490 IDE_DFLAG_NOPROBE = BIT(9), 491 /* need to do check_media_change() */ 492 IDE_DFLAG_REMOVABLE = BIT(10), 493 /* needed for removable devices */ 494 IDE_DFLAG_ATTACH = BIT(11), 495 IDE_DFLAG_FORCED_GEOM = BIT(12), 496 /* disallow setting unmask bit */ 497 IDE_DFLAG_NO_UNMASK = BIT(13), 498 /* disallow enabling 32-bit I/O */ 499 IDE_DFLAG_NO_IO_32BIT = BIT(14), 500 /* for removable only: door lock/unlock works */ 501 IDE_DFLAG_DOORLOCKING = BIT(15), 502 /* disallow DMA */ 503 IDE_DFLAG_NODMA = BIT(16), 504 /* powermanagement told us not to do anything, so sleep nicely */ 505 IDE_DFLAG_BLOCKED = BIT(17), 506 /* sleeping & sleep field valid */ 507 IDE_DFLAG_SLEEPING = BIT(18), 508 IDE_DFLAG_POST_RESET = BIT(19), 509 IDE_DFLAG_UDMA33_WARNED = BIT(20), 510 IDE_DFLAG_LBA48 = BIT(21), 511 /* status of write cache */ 512 IDE_DFLAG_WCACHE = BIT(22), 513 /* used for ignoring ATA_DF */ 514 IDE_DFLAG_NOWERR = BIT(23), 515 /* retrying in PIO */ 516 IDE_DFLAG_DMA_PIO_RETRY = BIT(24), 517 IDE_DFLAG_LBA = BIT(25), 518 /* don't unload heads */ 519 IDE_DFLAG_NO_UNLOAD = BIT(26), 520 /* heads unloaded, please don't reset port */ 521 IDE_DFLAG_PARKED = BIT(27), 522 IDE_DFLAG_MEDIA_CHANGED = BIT(28), 523 /* write protect */ 524 IDE_DFLAG_WP = BIT(29), 525 IDE_DFLAG_FORMAT_IN_PROGRESS = BIT(30), 526 IDE_DFLAG_NIEN_QUIRK = BIT(31), 527}; 528 529struct ide_drive_s { 530 char name[4]; /* drive name, such as "hda" */ 531 char driver_req[10]; /* requests specific driver */ 532 533 struct request_queue *queue; /* request queue */ 534 535 bool (*prep_rq)(struct ide_drive_s *, struct request *); 536 537 struct blk_mq_tag_set tag_set; 538 539 struct request *rq; /* current request */ 540 void *driver_data; /* extra driver data */ 541 u16 *id; /* identification info */ 542#ifdef CONFIG_IDE_PROC_FS 543 struct proc_dir_entry *proc; /* /proc/ide/ directory entry */ 544 const struct ide_proc_devset *settings; /* /proc/ide/ drive settings */ 545#endif 546 struct hwif_s *hwif; /* actually (ide_hwif_t *) */ 547 548 const struct ide_disk_ops *disk_ops; 549 550 unsigned long dev_flags; 551 552 unsigned long sleep; /* sleep until this time */ 553 unsigned long timeout; /* max time to wait for irq */ 554 555 u8 special_flags; /* special action flags */ 556 557 u8 select; /* basic drive/head select reg value */ 558 u8 retry_pio; /* retrying dma capable host in pio */ 559 u8 waiting_for_dma; /* dma currently in progress */ 560 u8 dma; /* atapi dma flag */ 561 562 u8 init_speed; /* transfer rate set at boot */ 563 u8 current_speed; /* current transfer rate set */ 564 u8 desired_speed; /* desired transfer rate set */ 565 u8 pio_mode; /* for ->set_pio_mode _only_ */ 566 u8 dma_mode; /* for ->set_dma_mode _only_ */ 567 u8 dn; /* now wide spread use */ 568 u8 acoustic; /* acoustic management */ 569 u8 media; /* disk, cdrom, tape, floppy, ... */ 570 u8 ready_stat; /* min status value for drive ready */ 571 u8 mult_count; /* current multiple sector setting */ 572 u8 mult_req; /* requested multiple sector setting */ 573 u8 io_32bit; /* 0=16-bit, 1=32-bit, 2/3=32bit+sync */ 574 u8 bad_wstat; /* used for ignoring ATA_DF */ 575 u8 head; /* "real" number of heads */ 576 u8 sect; /* "real" sectors per track */ 577 u8 bios_head; /* BIOS/fdisk/LILO number of heads */ 578 u8 bios_sect; /* BIOS/fdisk/LILO sectors per track */ 579 580 /* delay this long before sending packet command */ 581 u8 pc_delay; 582 583 unsigned int bios_cyl; /* BIOS/fdisk/LILO number of cyls */ 584 unsigned int cyl; /* "real" number of cyls */ 585 void *drive_data; /* used by set_pio_mode/dev_select() */ 586 unsigned int failures; /* current failure count */ 587 unsigned int max_failures; /* maximum allowed failure count */ 588 u64 probed_capacity;/* initial/native media capacity */ 589 u64 capacity64; /* total number of sectors */ 590 591 int lun; /* logical unit */ 592 int crc_count; /* crc counter to reduce drive speed */ 593 594 unsigned long debug_mask; /* debugging levels switch */ 595 596#ifdef CONFIG_BLK_DEV_IDEACPI 597 struct ide_acpi_drive_link *acpidata; 598#endif 599 struct list_head list; 600 struct device gendev; 601 struct completion gendev_rel_comp; /* to deal with device release() */ 602 603 /* current packet command */ 604 struct ide_atapi_pc *pc; 605 606 /* last failed packet command */ 607 struct ide_atapi_pc *failed_pc; 608 609 /* callback for packet commands */ 610 int (*pc_callback)(struct ide_drive_s *, int); 611 612 ide_startstop_t (*irq_handler)(struct ide_drive_s *); 613 614 unsigned long atapi_flags; 615 616 struct ide_atapi_pc request_sense_pc; 617 618 /* current sense rq and buffer */ 619 bool sense_rq_armed; 620 bool sense_rq_active; 621 struct request *sense_rq; 622 struct request_sense sense_data; 623 624 /* async sense insertion */ 625 struct work_struct rq_work; 626 struct list_head rq_list; 627}; 628 629typedef struct ide_drive_s ide_drive_t; 630 631#define to_ide_device(dev) container_of(dev, ide_drive_t, gendev) 632 633#define to_ide_drv(obj, cont_type) \ 634 container_of(obj, struct cont_type, dev) 635 636#define ide_drv_g(disk, cont_type) \ 637 container_of((disk)->private_data, struct cont_type, driver) 638 639struct ide_port_info; 640 641struct ide_tp_ops { 642 void (*exec_command)(struct hwif_s *, u8); 643 u8 (*read_status)(struct hwif_s *); 644 u8 (*read_altstatus)(struct hwif_s *); 645 void (*write_devctl)(struct hwif_s *, u8); 646 647 void (*dev_select)(ide_drive_t *); 648 void (*tf_load)(ide_drive_t *, struct ide_taskfile *, u8); 649 void (*tf_read)(ide_drive_t *, struct ide_taskfile *, u8); 650 651 void (*input_data)(ide_drive_t *, struct ide_cmd *, 652 void *, unsigned int); 653 void (*output_data)(ide_drive_t *, struct ide_cmd *, 654 void *, unsigned int); 655}; 656 657extern const struct ide_tp_ops default_tp_ops; 658 659/** 660 * struct ide_port_ops - IDE port operations 661 * 662 * @init_dev: host specific initialization of a device 663 * @set_pio_mode: routine to program host for PIO mode 664 * @set_dma_mode: routine to program host for DMA mode 665 * @reset_poll: chipset polling based on hba specifics 666 * @pre_reset: chipset specific changes to default for device-hba resets 667 * @resetproc: routine to reset controller after a disk reset 668 * @maskproc: special host masking for drive selection 669 * @quirkproc: check host's drive quirk list 670 * @clear_irq: clear IRQ 671 * 672 * @mdma_filter: filter MDMA modes 673 * @udma_filter: filter UDMA modes 674 * 675 * @cable_detect: detect cable type 676 */ 677struct ide_port_ops { 678 void (*init_dev)(ide_drive_t *); 679 void (*set_pio_mode)(struct hwif_s *, ide_drive_t *); 680 void (*set_dma_mode)(struct hwif_s *, ide_drive_t *); 681 blk_status_t (*reset_poll)(ide_drive_t *); 682 void (*pre_reset)(ide_drive_t *); 683 void (*resetproc)(ide_drive_t *); 684 void (*maskproc)(ide_drive_t *, int); 685 void (*quirkproc)(ide_drive_t *); 686 void (*clear_irq)(ide_drive_t *); 687 int (*test_irq)(struct hwif_s *); 688 689 u8 (*mdma_filter)(ide_drive_t *); 690 u8 (*udma_filter)(ide_drive_t *); 691 692 u8 (*cable_detect)(struct hwif_s *); 693}; 694 695struct ide_dma_ops { 696 void (*dma_host_set)(struct ide_drive_s *, int); 697 int (*dma_setup)(struct ide_drive_s *, struct ide_cmd *); 698 void (*dma_start)(struct ide_drive_s *); 699 int (*dma_end)(struct ide_drive_s *); 700 int (*dma_test_irq)(struct ide_drive_s *); 701 void (*dma_lost_irq)(struct ide_drive_s *); 702 /* below ones are optional */ 703 int (*dma_check)(struct ide_drive_s *, struct ide_cmd *); 704 int (*dma_timer_expiry)(struct ide_drive_s *); 705 void (*dma_clear)(struct ide_drive_s *); 706 /* 707 * The following method is optional and only required to be 708 * implemented for the SFF-8038i compatible controllers. 709 */ 710 u8 (*dma_sff_read_status)(struct hwif_s *); 711}; 712 713enum { 714 IDE_PFLAG_PROBING = BIT(0), 715}; 716 717struct ide_host; 718 719typedef struct hwif_s { 720 struct hwif_s *mate; /* other hwif from same PCI chip */ 721 struct proc_dir_entry *proc; /* /proc/ide/ directory entry */ 722 723 struct ide_host *host; 724 725 char name[6]; /* name of interface, eg. "ide0" */ 726 727 struct ide_io_ports io_ports; 728 729 unsigned long sata_scr[SATA_NR_PORTS]; 730 731 ide_drive_t *devices[MAX_DRIVES + 1]; 732 733 unsigned long port_flags; 734 735 u8 major; /* our major number */ 736 u8 index; /* 0 for ide0; 1 for ide1; ... */ 737 u8 channel; /* for dual-port chips: 0=primary, 1=secondary */ 738 739 u32 host_flags; 740 741 u8 pio_mask; 742 743 u8 ultra_mask; 744 u8 mwdma_mask; 745 u8 swdma_mask; 746 747 u8 cbl; /* cable type */ 748 749 hwif_chipset_t chipset; /* sub-module for tuning.. */ 750 751 struct device *dev; 752 753 void (*rw_disk)(ide_drive_t *, struct request *); 754 755 const struct ide_tp_ops *tp_ops; 756 const struct ide_port_ops *port_ops; 757 const struct ide_dma_ops *dma_ops; 758 759 /* dma physical region descriptor table (cpu view) */ 760 unsigned int *dmatable_cpu; 761 /* dma physical region descriptor table (dma view) */ 762 dma_addr_t dmatable_dma; 763 764 /* maximum number of PRD table entries */ 765 int prd_max_nents; 766 /* PRD entry size in bytes */ 767 int prd_ent_size; 768 769 /* Scatter-gather list used to build the above */ 770 struct scatterlist *sg_table; 771 int sg_max_nents; /* Maximum number of entries in it */ 772 773 struct ide_cmd cmd; /* current command */ 774 775 int rqsize; /* max sectors per request */ 776 int irq; /* our irq number */ 777 778 unsigned long dma_base; /* base addr for dma ports */ 779 780 unsigned long config_data; /* for use by chipset-specific code */ 781 unsigned long select_data; /* for use by chipset-specific code */ 782 783 unsigned long extra_base; /* extra addr for dma ports */ 784 unsigned extra_ports; /* number of extra dma ports */ 785 786 unsigned present : 1; /* this interface exists */ 787 unsigned busy : 1; /* serializes devices on a port */ 788 789 struct device gendev; 790 struct device *portdev; 791 792 struct completion gendev_rel_comp; /* To deal with device release() */ 793 794 void *hwif_data; /* extra hwif data */ 795 796#ifdef CONFIG_BLK_DEV_IDEACPI 797 struct ide_acpi_hwif_link *acpidata; 798#endif 799 800 /* IRQ handler, if active */ 801 ide_startstop_t (*handler)(ide_drive_t *); 802 803 /* BOOL: polling active & poll_timeout field valid */ 804 unsigned int polling : 1; 805 806 /* current drive */ 807 ide_drive_t *cur_dev; 808 809 /* current request */ 810 struct request *rq; 811 812 /* failsafe timer */ 813 struct timer_list timer; 814 /* timeout value during long polls */ 815 unsigned long poll_timeout; 816 /* queried upon timeouts */ 817 int (*expiry)(ide_drive_t *); 818 819 int req_gen; 820 int req_gen_timer; 821 822 spinlock_t lock; 823} ____cacheline_internodealigned_in_smp ide_hwif_t; 824 825#define MAX_HOST_PORTS 4 826 827struct ide_host { 828 ide_hwif_t *ports[MAX_HOST_PORTS + 1]; 829 unsigned int n_ports; 830 struct device *dev[2]; 831 832 int (*init_chipset)(struct pci_dev *); 833 834 void (*get_lock)(irq_handler_t, void *); 835 void (*release_lock)(void); 836 837 irq_handler_t irq_handler; 838 839 unsigned long host_flags; 840 841 int irq_flags; 842 843 void *host_priv; 844 ide_hwif_t *cur_port; /* for hosts requiring serialization */ 845 846 /* used for hosts requiring serialization */ 847 volatile unsigned long host_busy; 848}; 849 850#define IDE_HOST_BUSY 0 851 852/* 853 * internal ide interrupt handler type 854 */ 855typedef ide_startstop_t (ide_handler_t)(ide_drive_t *); 856typedef int (ide_expiry_t)(ide_drive_t *); 857 858/* used by ide-cd, ide-floppy, etc. */ 859typedef void (xfer_func_t)(ide_drive_t *, struct ide_cmd *, void *, unsigned); 860 861extern struct mutex ide_setting_mtx; 862 863/* 864 * configurable drive settings 865 */ 866 867#define DS_SYNC BIT(0) 868 869struct ide_devset { 870 int (*get)(ide_drive_t *); 871 int (*set)(ide_drive_t *, int); 872 unsigned int flags; 873}; 874 875#define __DEVSET(_flags, _get, _set) { \ 876 .flags = _flags, \ 877 .get = _get, \ 878 .set = _set, \ 879} 880 881#define ide_devset_get(name, field) \ 882static int get_##name(ide_drive_t *drive) \ 883{ \ 884 return drive->field; \ 885} 886 887#define ide_devset_set(name, field) \ 888static int set_##name(ide_drive_t *drive, int arg) \ 889{ \ 890 drive->field = arg; \ 891 return 0; \ 892} 893 894#define ide_devset_get_flag(name, flag) \ 895static int get_##name(ide_drive_t *drive) \ 896{ \ 897 return !!(drive->dev_flags & flag); \ 898} 899 900#define ide_devset_set_flag(name, flag) \ 901static int set_##name(ide_drive_t *drive, int arg) \ 902{ \ 903 if (arg) \ 904 drive->dev_flags |= flag; \ 905 else \ 906 drive->dev_flags &= ~flag; \ 907 return 0; \ 908} 909 910#define __IDE_DEVSET(_name, _flags, _get, _set) \ 911const struct ide_devset ide_devset_##_name = \ 912 __DEVSET(_flags, _get, _set) 913 914#define IDE_DEVSET(_name, _flags, _get, _set) \ 915static __IDE_DEVSET(_name, _flags, _get, _set) 916 917#define ide_devset_rw(_name, _func) \ 918IDE_DEVSET(_name, 0, get_##_func, set_##_func) 919 920#define ide_devset_w(_name, _func) \ 921IDE_DEVSET(_name, 0, NULL, set_##_func) 922 923#define ide_ext_devset_rw(_name, _func) \ 924__IDE_DEVSET(_name, 0, get_##_func, set_##_func) 925 926#define ide_ext_devset_rw_sync(_name, _func) \ 927__IDE_DEVSET(_name, DS_SYNC, get_##_func, set_##_func) 928 929#define ide_decl_devset(_name) \ 930extern const struct ide_devset ide_devset_##_name 931 932ide_decl_devset(io_32bit); 933ide_decl_devset(keepsettings); 934ide_decl_devset(pio_mode); 935ide_decl_devset(unmaskirq); 936ide_decl_devset(using_dma); 937 938#ifdef CONFIG_IDE_PROC_FS 939/* 940 * /proc/ide interface 941 */ 942 943#define ide_devset_rw_field(_name, _field) \ 944ide_devset_get(_name, _field); \ 945ide_devset_set(_name, _field); \ 946IDE_DEVSET(_name, DS_SYNC, get_##_name, set_##_name) 947 948#define ide_devset_ro_field(_name, _field) \ 949ide_devset_get(_name, _field); \ 950IDE_DEVSET(_name, 0, get_##_name, NULL) 951 952#define ide_devset_rw_flag(_name, _field) \ 953ide_devset_get_flag(_name, _field); \ 954ide_devset_set_flag(_name, _field); \ 955IDE_DEVSET(_name, DS_SYNC, get_##_name, set_##_name) 956 957struct ide_proc_devset { 958 const char *name; 959 const struct ide_devset *setting; 960 int min, max; 961 int (*mulf)(ide_drive_t *); 962 int (*divf)(ide_drive_t *); 963}; 964 965#define __IDE_PROC_DEVSET(_name, _min, _max, _mulf, _divf) { \ 966 .name = __stringify(_name), \ 967 .setting = &ide_devset_##_name, \ 968 .min = _min, \ 969 .max = _max, \ 970 .mulf = _mulf, \ 971 .divf = _divf, \ 972} 973 974#define IDE_PROC_DEVSET(_name, _min, _max) \ 975__IDE_PROC_DEVSET(_name, _min, _max, NULL, NULL) 976 977typedef struct { 978 const char *name; 979 umode_t mode; 980 int (*show)(struct seq_file *, void *); 981} ide_proc_entry_t; 982 983void proc_ide_create(void); 984void proc_ide_destroy(void); 985void ide_proc_register_port(ide_hwif_t *); 986void ide_proc_port_register_devices(ide_hwif_t *); 987void ide_proc_unregister_device(ide_drive_t *); 988void ide_proc_unregister_port(ide_hwif_t *); 989void ide_proc_register_driver(ide_drive_t *, struct ide_driver *); 990void ide_proc_unregister_driver(ide_drive_t *, struct ide_driver *); 991 992int ide_capacity_proc_show(struct seq_file *m, void *v); 993int ide_geometry_proc_show(struct seq_file *m, void *v); 994#else 995static inline void proc_ide_create(void) { ; } 996static inline void proc_ide_destroy(void) { ; } 997static inline void ide_proc_register_port(ide_hwif_t *hwif) { ; } 998static inline void ide_proc_port_register_devices(ide_hwif_t *hwif) { ; } 999static inline void ide_proc_unregister_device(ide_drive_t *drive) { ; } 1000static inline void ide_proc_unregister_port(ide_hwif_t *hwif) { ; } 1001static inline void ide_proc_register_driver(ide_drive_t *drive, 1002 struct ide_driver *driver) { ; } 1003static inline void ide_proc_unregister_driver(ide_drive_t *drive, 1004 struct ide_driver *driver) { ; } 1005#endif 1006 1007enum { 1008 /* enter/exit functions */ 1009 IDE_DBG_FUNC = BIT(0), 1010 /* sense key/asc handling */ 1011 IDE_DBG_SENSE = BIT(1), 1012 /* packet commands handling */ 1013 IDE_DBG_PC = BIT(2), 1014 /* request handling */ 1015 IDE_DBG_RQ = BIT(3), 1016 /* driver probing/setup */ 1017 IDE_DBG_PROBE = BIT(4), 1018}; 1019 1020/* DRV_NAME has to be defined in the driver before using the macro below */ 1021#define __ide_debug_log(lvl, fmt, args...) \ 1022{ \ 1023 if (unlikely(drive->debug_mask & lvl)) \ 1024 printk(KERN_INFO DRV_NAME ": %s: " fmt "\n", \ 1025 __func__, ## args); \ 1026} 1027 1028/* 1029 * Power Management state machine (rq->pm->pm_step). 1030 * 1031 * For each step, the core calls ide_start_power_step() first. 1032 * This can return: 1033 * - ide_stopped : In this case, the core calls us back again unless 1034 * step have been set to ide_power_state_completed. 1035 * - ide_started : In this case, the channel is left busy until an 1036 * async event (interrupt) occurs. 1037 * Typically, ide_start_power_step() will issue a taskfile request with 1038 * do_rw_taskfile(). 1039 * 1040 * Upon reception of the interrupt, the core will call ide_complete_power_step() 1041 * with the error code if any. This routine should update the step value 1042 * and return. It should not start a new request. The core will call 1043 * ide_start_power_step() for the new step value, unless step have been 1044 * set to IDE_PM_COMPLETED. 1045 */ 1046enum { 1047 IDE_PM_START_SUSPEND, 1048 IDE_PM_FLUSH_CACHE = IDE_PM_START_SUSPEND, 1049 IDE_PM_STANDBY, 1050 1051 IDE_PM_START_RESUME, 1052 IDE_PM_RESTORE_PIO = IDE_PM_START_RESUME, 1053 IDE_PM_IDLE, 1054 IDE_PM_RESTORE_DMA, 1055 1056 IDE_PM_COMPLETED, 1057}; 1058 1059int generic_ide_suspend(struct device *, pm_message_t); 1060int generic_ide_resume(struct device *); 1061 1062void ide_complete_power_step(ide_drive_t *, struct request *); 1063ide_startstop_t ide_start_power_step(ide_drive_t *, struct request *); 1064void ide_complete_pm_rq(ide_drive_t *, struct request *); 1065void ide_check_pm_state(ide_drive_t *, struct request *); 1066 1067/* 1068 * Subdrivers support. 1069 * 1070 * The gendriver.owner field should be set to the module owner of this driver. 1071 * The gendriver.name field should be set to the name of this driver 1072 */ 1073struct ide_driver { 1074 const char *version; 1075 ide_startstop_t (*do_request)(ide_drive_t *, struct request *, sector_t); 1076 struct device_driver gen_driver; 1077 int (*probe)(ide_drive_t *); 1078 void (*remove)(ide_drive_t *); 1079 void (*resume)(ide_drive_t *); 1080 void (*shutdown)(ide_drive_t *); 1081#ifdef CONFIG_IDE_PROC_FS 1082 ide_proc_entry_t * (*proc_entries)(ide_drive_t *); 1083 const struct ide_proc_devset * (*proc_devsets)(ide_drive_t *); 1084#endif 1085}; 1086 1087#define to_ide_driver(drv) container_of(drv, struct ide_driver, gen_driver) 1088 1089int ide_device_get(ide_drive_t *); 1090void ide_device_put(ide_drive_t *); 1091 1092struct ide_ioctl_devset { 1093 unsigned int get_ioctl; 1094 unsigned int set_ioctl; 1095 const struct ide_devset *setting; 1096}; 1097 1098int ide_setting_ioctl(ide_drive_t *, struct block_device *, unsigned int, 1099 unsigned long, const struct ide_ioctl_devset *); 1100 1101int generic_ide_ioctl(ide_drive_t *, struct block_device *, unsigned, unsigned long); 1102 1103extern int ide_vlb_clk; 1104extern int ide_pci_clk; 1105 1106int ide_end_rq(ide_drive_t *, struct request *, blk_status_t, unsigned int); 1107void ide_kill_rq(ide_drive_t *, struct request *); 1108void ide_insert_request_head(ide_drive_t *, struct request *); 1109 1110void __ide_set_handler(ide_drive_t *, ide_handler_t *, unsigned int); 1111void ide_set_handler(ide_drive_t *, ide_handler_t *, unsigned int); 1112 1113void ide_execute_command(ide_drive_t *, struct ide_cmd *, ide_handler_t *, 1114 unsigned int); 1115 1116void ide_pad_transfer(ide_drive_t *, int, int); 1117 1118ide_startstop_t ide_error(ide_drive_t *, const char *, u8); 1119 1120void ide_fix_driveid(u16 *); 1121 1122extern void ide_fixstring(u8 *, const int, const int); 1123 1124int ide_busy_sleep(ide_drive_t *, unsigned long, int); 1125 1126int __ide_wait_stat(ide_drive_t *, u8, u8, unsigned long, u8 *); 1127int ide_wait_stat(ide_startstop_t *, ide_drive_t *, u8, u8, unsigned long); 1128 1129ide_startstop_t ide_do_park_unpark(ide_drive_t *, struct request *); 1130ide_startstop_t ide_do_devset(ide_drive_t *, struct request *); 1131 1132extern ide_startstop_t ide_do_reset (ide_drive_t *); 1133 1134extern int ide_devset_execute(ide_drive_t *drive, 1135 const struct ide_devset *setting, int arg); 1136 1137void ide_complete_cmd(ide_drive_t *, struct ide_cmd *, u8, u8); 1138int ide_complete_rq(ide_drive_t *, blk_status_t, unsigned int); 1139 1140void ide_tf_readback(ide_drive_t *drive, struct ide_cmd *cmd); 1141void ide_tf_dump(const char *, struct ide_cmd *); 1142 1143void ide_exec_command(ide_hwif_t *, u8); 1144u8 ide_read_status(ide_hwif_t *); 1145u8 ide_read_altstatus(ide_hwif_t *); 1146void ide_write_devctl(ide_hwif_t *, u8); 1147 1148void ide_dev_select(ide_drive_t *); 1149void ide_tf_load(ide_drive_t *, struct ide_taskfile *, u8); 1150void ide_tf_read(ide_drive_t *, struct ide_taskfile *, u8); 1151 1152void ide_input_data(ide_drive_t *, struct ide_cmd *, void *, unsigned int); 1153void ide_output_data(ide_drive_t *, struct ide_cmd *, void *, unsigned int); 1154 1155void SELECT_MASK(ide_drive_t *, int); 1156 1157u8 ide_read_error(ide_drive_t *); 1158void ide_read_bcount_and_ireason(ide_drive_t *, u16 *, u8 *); 1159 1160int ide_check_ireason(ide_drive_t *, struct request *, int, int, int); 1161 1162int ide_check_atapi_device(ide_drive_t *, const char *); 1163 1164void ide_init_pc(struct ide_atapi_pc *); 1165 1166/* Disk head parking */ 1167extern wait_queue_head_t ide_park_wq; 1168ssize_t ide_park_show(struct device *dev, struct device_attribute *attr, 1169 char *buf); 1170ssize_t ide_park_store(struct device *dev, struct device_attribute *attr, 1171 const char *buf, size_t len); 1172 1173/* 1174 * Special requests for ide-tape block device strategy routine. 1175 * 1176 * In order to service a character device command, we add special requests to 1177 * the tail of our block device request queue and wait for their completion. 1178 */ 1179enum { 1180 REQ_IDETAPE_PC1 = BIT(0), /* packet command (first stage) */ 1181 REQ_IDETAPE_PC2 = BIT(1), /* packet command (second stage) */ 1182 REQ_IDETAPE_READ = BIT(2), 1183 REQ_IDETAPE_WRITE = BIT(3), 1184}; 1185 1186int ide_queue_pc_tail(ide_drive_t *, struct gendisk *, struct ide_atapi_pc *, 1187 void *, unsigned int); 1188 1189int ide_do_test_unit_ready(ide_drive_t *, struct gendisk *); 1190int ide_do_start_stop(ide_drive_t *, struct gendisk *, int); 1191int ide_set_media_lock(ide_drive_t *, struct gendisk *, int); 1192void ide_create_request_sense_cmd(ide_drive_t *, struct ide_atapi_pc *); 1193void ide_retry_pc(ide_drive_t *drive); 1194 1195void ide_prep_sense(ide_drive_t *drive, struct request *rq); 1196int ide_queue_sense_rq(ide_drive_t *drive, void *special); 1197 1198int ide_cd_expiry(ide_drive_t *); 1199 1200int ide_cd_get_xferlen(struct request *); 1201 1202ide_startstop_t ide_issue_pc(ide_drive_t *, struct ide_cmd *); 1203 1204ide_startstop_t do_rw_taskfile(ide_drive_t *, struct ide_cmd *); 1205 1206void ide_pio_bytes(ide_drive_t *, struct ide_cmd *, unsigned int, unsigned int); 1207 1208void ide_finish_cmd(ide_drive_t *, struct ide_cmd *, u8); 1209 1210int ide_raw_taskfile(ide_drive_t *, struct ide_cmd *, u8 *, u16); 1211int ide_no_data_taskfile(ide_drive_t *, struct ide_cmd *); 1212 1213int ide_taskfile_ioctl(ide_drive_t *, unsigned long); 1214 1215int ide_dev_read_id(ide_drive_t *, u8, u16 *, int); 1216 1217extern int ide_driveid_update(ide_drive_t *); 1218extern int ide_config_drive_speed(ide_drive_t *, u8); 1219extern u8 eighty_ninty_three (ide_drive_t *); 1220extern int taskfile_lib_get_identify(ide_drive_t *drive, u8 *); 1221 1222extern int ide_wait_not_busy(ide_hwif_t *hwif, unsigned long timeout); 1223 1224extern void ide_stall_queue(ide_drive_t *drive, unsigned long timeout); 1225 1226extern void ide_timer_expiry(struct timer_list *t); 1227extern irqreturn_t ide_intr(int irq, void *dev_id); 1228extern blk_status_t ide_queue_rq(struct blk_mq_hw_ctx *, const struct blk_mq_queue_data *); 1229extern blk_status_t ide_issue_rq(ide_drive_t *, struct request *, bool); 1230extern void ide_requeue_and_plug(ide_drive_t *drive, struct request *rq); 1231 1232void ide_init_disk(struct gendisk *, ide_drive_t *); 1233 1234#ifdef CONFIG_IDEPCI_PCIBUS_ORDER 1235extern int __ide_pci_register_driver(struct pci_driver *driver, struct module *owner, const char *mod_name); 1236#define ide_pci_register_driver(d) __ide_pci_register_driver(d, THIS_MODULE, KBUILD_MODNAME) 1237#else 1238#define ide_pci_register_driver(d) pci_register_driver(d) 1239#endif 1240 1241static inline int ide_pci_is_in_compatibility_mode(struct pci_dev *dev) 1242{ 1243 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE && (dev->class & 5) != 5) 1244 return 1; 1245 return 0; 1246} 1247 1248void ide_pci_setup_ports(struct pci_dev *, const struct ide_port_info *, 1249 struct ide_hw *, struct ide_hw **); 1250void ide_setup_pci_noise(struct pci_dev *, const struct ide_port_info *); 1251 1252#ifdef CONFIG_BLK_DEV_IDEDMA_PCI 1253int ide_pci_set_master(struct pci_dev *, const char *); 1254unsigned long ide_pci_dma_base(ide_hwif_t *, const struct ide_port_info *); 1255int ide_pci_check_simplex(ide_hwif_t *, const struct ide_port_info *); 1256int ide_hwif_setup_dma(ide_hwif_t *, const struct ide_port_info *); 1257#else 1258static inline int ide_hwif_setup_dma(ide_hwif_t *hwif, 1259 const struct ide_port_info *d) 1260{ 1261 return -EINVAL; 1262} 1263#endif 1264 1265struct ide_pci_enablebit { 1266 u8 reg; /* byte pci reg holding the enable-bit */ 1267 u8 mask; /* mask to isolate the enable-bit */ 1268 u8 val; /* value of masked reg when "enabled" */ 1269}; 1270 1271enum { 1272 /* Uses ISA control ports not PCI ones. */ 1273 IDE_HFLAG_ISA_PORTS = BIT(0), 1274 /* single port device */ 1275 IDE_HFLAG_SINGLE = BIT(1), 1276 /* don't use legacy PIO blacklist */ 1277 IDE_HFLAG_PIO_NO_BLACKLIST = BIT(2), 1278 /* set for the second port of QD65xx */ 1279 IDE_HFLAG_QD_2ND_PORT = BIT(3), 1280 /* use PIO8/9 for prefetch off/on */ 1281 IDE_HFLAG_ABUSE_PREFETCH = BIT(4), 1282 /* use PIO6/7 for fast-devsel off/on */ 1283 IDE_HFLAG_ABUSE_FAST_DEVSEL = BIT(5), 1284 /* use 100-102 and 200-202 PIO values to set DMA modes */ 1285 IDE_HFLAG_ABUSE_DMA_MODES = BIT(6), 1286 /* 1287 * keep DMA setting when programming PIO mode, may be used only 1288 * for hosts which have separate PIO and DMA timings (ie. PMAC) 1289 */ 1290 IDE_HFLAG_SET_PIO_MODE_KEEP_DMA = BIT(7), 1291 /* program host for the transfer mode after programming device */ 1292 IDE_HFLAG_POST_SET_MODE = BIT(8), 1293 /* don't program host/device for the transfer mode ("smart" hosts) */ 1294 IDE_HFLAG_NO_SET_MODE = BIT(9), 1295 /* trust BIOS for programming chipset/device for DMA */ 1296 IDE_HFLAG_TRUST_BIOS_FOR_DMA = BIT(10), 1297 /* host is CS5510/CS5520 */ 1298 IDE_HFLAG_CS5520 = BIT(11), 1299 /* ATAPI DMA is unsupported */ 1300 IDE_HFLAG_NO_ATAPI_DMA = BIT(12), 1301 /* set if host is a "non-bootable" controller */ 1302 IDE_HFLAG_NON_BOOTABLE = BIT(13), 1303 /* host doesn't support DMA */ 1304 IDE_HFLAG_NO_DMA = BIT(14), 1305 /* check if host is PCI IDE device before allowing DMA */ 1306 IDE_HFLAG_NO_AUTODMA = BIT(15), 1307 /* host uses MMIO */ 1308 IDE_HFLAG_MMIO = BIT(16), 1309 /* no LBA48 */ 1310 IDE_HFLAG_NO_LBA48 = BIT(17), 1311 /* no LBA48 DMA */ 1312 IDE_HFLAG_NO_LBA48_DMA = BIT(18), 1313 /* data FIFO is cleared by an error */ 1314 IDE_HFLAG_ERROR_STOPS_FIFO = BIT(19), 1315 /* serialize ports */ 1316 IDE_HFLAG_SERIALIZE = BIT(20), 1317 /* host is DTC2278 */ 1318 IDE_HFLAG_DTC2278 = BIT(21), 1319 /* 4 devices on a single set of I/O ports */ 1320 IDE_HFLAG_4DRIVES = BIT(22), 1321 /* host is TRM290 */ 1322 IDE_HFLAG_TRM290 = BIT(23), 1323 /* use 32-bit I/O ops */ 1324 IDE_HFLAG_IO_32BIT = BIT(24), 1325 /* unmask IRQs */ 1326 IDE_HFLAG_UNMASK_IRQS = BIT(25), 1327 IDE_HFLAG_BROKEN_ALTSTATUS = BIT(26), 1328 /* serialize ports if DMA is possible (for sl82c105) */ 1329 IDE_HFLAG_SERIALIZE_DMA = BIT(27), 1330 /* force host out of "simplex" mode */ 1331 IDE_HFLAG_CLEAR_SIMPLEX = BIT(28), 1332 /* DSC overlap is unsupported */ 1333 IDE_HFLAG_NO_DSC = BIT(29), 1334 /* never use 32-bit I/O ops */ 1335 IDE_HFLAG_NO_IO_32BIT = BIT(30), 1336 /* never unmask IRQs */ 1337 IDE_HFLAG_NO_UNMASK_IRQS = BIT(31), 1338}; 1339 1340#ifdef CONFIG_BLK_DEV_OFFBOARD 1341# define IDE_HFLAG_OFF_BOARD 0 1342#else 1343# define IDE_HFLAG_OFF_BOARD IDE_HFLAG_NON_BOOTABLE 1344#endif 1345 1346struct ide_port_info { 1347 char *name; 1348 1349 int (*init_chipset)(struct pci_dev *); 1350 1351 void (*get_lock)(irq_handler_t, void *); 1352 void (*release_lock)(void); 1353 1354 void (*init_iops)(ide_hwif_t *); 1355 void (*init_hwif)(ide_hwif_t *); 1356 int (*init_dma)(ide_hwif_t *, 1357 const struct ide_port_info *); 1358 1359 const struct ide_tp_ops *tp_ops; 1360 const struct ide_port_ops *port_ops; 1361 const struct ide_dma_ops *dma_ops; 1362 1363 struct ide_pci_enablebit enablebits[2]; 1364 1365 hwif_chipset_t chipset; 1366 1367 u16 max_sectors; /* if < than the default one */ 1368 1369 u32 host_flags; 1370 1371 int irq_flags; 1372 1373 u8 pio_mask; 1374 u8 swdma_mask; 1375 u8 mwdma_mask; 1376 u8 udma_mask; 1377}; 1378 1379/* 1380 * State information carried for REQ_TYPE_ATA_PM_SUSPEND and REQ_TYPE_ATA_PM_RESUME 1381 * requests. 1382 */ 1383struct ide_pm_state { 1384 /* PM state machine step value, currently driver specific */ 1385 int pm_step; 1386 /* requested PM state value (S1, S2, S3, S4, ...) */ 1387 u32 pm_state; 1388 void* data; /* for driver use */ 1389}; 1390 1391 1392int ide_pci_init_one(struct pci_dev *, const struct ide_port_info *, void *); 1393int ide_pci_init_two(struct pci_dev *, struct pci_dev *, 1394 const struct ide_port_info *, void *); 1395void ide_pci_remove(struct pci_dev *); 1396 1397#ifdef CONFIG_PM 1398int ide_pci_suspend(struct pci_dev *, pm_message_t); 1399int ide_pci_resume(struct pci_dev *); 1400#else 1401#define ide_pci_suspend NULL 1402#define ide_pci_resume NULL 1403#endif 1404 1405void ide_map_sg(ide_drive_t *, struct ide_cmd *); 1406void ide_init_sg_cmd(struct ide_cmd *, unsigned int); 1407 1408#define BAD_DMA_DRIVE 0 1409#define GOOD_DMA_DRIVE 1 1410 1411struct drive_list_entry { 1412 const char *id_model; 1413 const char *id_firmware; 1414}; 1415 1416int ide_in_drive_list(u16 *, const struct drive_list_entry *); 1417 1418#ifdef CONFIG_BLK_DEV_IDEDMA 1419int ide_dma_good_drive(ide_drive_t *); 1420int __ide_dma_bad_drive(ide_drive_t *); 1421 1422u8 ide_find_dma_mode(ide_drive_t *, u8); 1423 1424static inline u8 ide_max_dma_mode(ide_drive_t *drive) 1425{ 1426 return ide_find_dma_mode(drive, XFER_UDMA_6); 1427} 1428 1429void ide_dma_off_quietly(ide_drive_t *); 1430void ide_dma_off(ide_drive_t *); 1431void ide_dma_on(ide_drive_t *); 1432int ide_set_dma(ide_drive_t *); 1433void ide_check_dma_crc(ide_drive_t *); 1434ide_startstop_t ide_dma_intr(ide_drive_t *); 1435 1436int ide_allocate_dma_engine(ide_hwif_t *); 1437void ide_release_dma_engine(ide_hwif_t *); 1438 1439int ide_dma_prepare(ide_drive_t *, struct ide_cmd *); 1440void ide_dma_unmap_sg(ide_drive_t *, struct ide_cmd *); 1441 1442#ifdef CONFIG_BLK_DEV_IDEDMA_SFF 1443int config_drive_for_dma(ide_drive_t *); 1444int ide_build_dmatable(ide_drive_t *, struct ide_cmd *); 1445void ide_dma_host_set(ide_drive_t *, int); 1446int ide_dma_setup(ide_drive_t *, struct ide_cmd *); 1447extern void ide_dma_start(ide_drive_t *); 1448int ide_dma_end(ide_drive_t *); 1449int ide_dma_test_irq(ide_drive_t *); 1450int ide_dma_sff_timer_expiry(ide_drive_t *); 1451u8 ide_dma_sff_read_status(ide_hwif_t *); 1452extern const struct ide_dma_ops sff_dma_ops; 1453#else 1454static inline int config_drive_for_dma(ide_drive_t *drive) { return 0; } 1455#endif /* CONFIG_BLK_DEV_IDEDMA_SFF */ 1456 1457void ide_dma_lost_irq(ide_drive_t *); 1458ide_startstop_t ide_dma_timeout_retry(ide_drive_t *, int); 1459 1460#else 1461static inline u8 ide_find_dma_mode(ide_drive_t *drive, u8 speed) { return 0; } 1462static inline u8 ide_max_dma_mode(ide_drive_t *drive) { return 0; } 1463static inline void ide_dma_off_quietly(ide_drive_t *drive) { ; } 1464static inline void ide_dma_off(ide_drive_t *drive) { ; } 1465static inline void ide_dma_on(ide_drive_t *drive) { ; } 1466static inline void ide_dma_verbose(ide_drive_t *drive) { ; } 1467static inline int ide_set_dma(ide_drive_t *drive) { return 1; } 1468static inline void ide_check_dma_crc(ide_drive_t *drive) { ; } 1469static inline ide_startstop_t ide_dma_intr(ide_drive_t *drive) { return ide_stopped; } 1470static inline ide_startstop_t ide_dma_timeout_retry(ide_drive_t *drive, int error) { return ide_stopped; } 1471static inline void ide_release_dma_engine(ide_hwif_t *hwif) { ; } 1472static inline int ide_dma_prepare(ide_drive_t *drive, 1473 struct ide_cmd *cmd) { return 1; } 1474static inline void ide_dma_unmap_sg(ide_drive_t *drive, 1475 struct ide_cmd *cmd) { ; } 1476#endif /* CONFIG_BLK_DEV_IDEDMA */ 1477 1478#ifdef CONFIG_BLK_DEV_IDEACPI 1479int ide_acpi_init(void); 1480bool ide_port_acpi(ide_hwif_t *hwif); 1481extern int ide_acpi_exec_tfs(ide_drive_t *drive); 1482extern void ide_acpi_get_timing(ide_hwif_t *hwif); 1483extern void ide_acpi_push_timing(ide_hwif_t *hwif); 1484void ide_acpi_init_port(ide_hwif_t *); 1485void ide_acpi_port_init_devices(ide_hwif_t *); 1486extern void ide_acpi_set_state(ide_hwif_t *hwif, int on); 1487#else 1488static inline int ide_acpi_init(void) { return 0; } 1489static inline bool ide_port_acpi(ide_hwif_t *hwif) { return 0; } 1490static inline int ide_acpi_exec_tfs(ide_drive_t *drive) { return 0; } 1491static inline void ide_acpi_get_timing(ide_hwif_t *hwif) { ; } 1492static inline void ide_acpi_push_timing(ide_hwif_t *hwif) { ; } 1493static inline void ide_acpi_init_port(ide_hwif_t *hwif) { ; } 1494static inline void ide_acpi_port_init_devices(ide_hwif_t *hwif) { ; } 1495static inline void ide_acpi_set_state(ide_hwif_t *hwif, int on) {} 1496#endif 1497 1498void ide_register_region(struct gendisk *); 1499void ide_unregister_region(struct gendisk *); 1500 1501void ide_check_nien_quirk_list(ide_drive_t *); 1502void ide_undecoded_slave(ide_drive_t *); 1503 1504void ide_port_apply_params(ide_hwif_t *); 1505int ide_sysfs_register_port(ide_hwif_t *); 1506 1507struct ide_host *ide_host_alloc(const struct ide_port_info *, struct ide_hw **, 1508 unsigned int); 1509void ide_host_free(struct ide_host *); 1510int ide_host_register(struct ide_host *, const struct ide_port_info *, 1511 struct ide_hw **); 1512int ide_host_add(const struct ide_port_info *, struct ide_hw **, unsigned int, 1513 struct ide_host **); 1514void ide_host_remove(struct ide_host *); 1515int ide_legacy_device_add(const struct ide_port_info *, unsigned long); 1516void ide_port_unregister_devices(ide_hwif_t *); 1517void ide_port_scan(ide_hwif_t *); 1518 1519static inline void *ide_get_hwifdata (ide_hwif_t * hwif) 1520{ 1521 return hwif->hwif_data; 1522} 1523 1524static inline void ide_set_hwifdata (ide_hwif_t * hwif, void *data) 1525{ 1526 hwif->hwif_data = data; 1527} 1528 1529u64 ide_get_lba_addr(struct ide_cmd *, int); 1530u8 ide_dump_status(ide_drive_t *, const char *, u8); 1531 1532struct ide_timing { 1533 u8 mode; 1534 u8 setup; /* t1 */ 1535 u16 act8b; /* t2 for 8-bit io */ 1536 u16 rec8b; /* t2i for 8-bit io */ 1537 u16 cyc8b; /* t0 for 8-bit io */ 1538 u16 active; /* t2 or tD */ 1539 u16 recover; /* t2i or tK */ 1540 u16 cycle; /* t0 */ 1541 u16 udma; /* t2CYCTYP/2 */ 1542}; 1543 1544enum { 1545 IDE_TIMING_SETUP = BIT(0), 1546 IDE_TIMING_ACT8B = BIT(1), 1547 IDE_TIMING_REC8B = BIT(2), 1548 IDE_TIMING_CYC8B = BIT(3), 1549 IDE_TIMING_8BIT = IDE_TIMING_ACT8B | IDE_TIMING_REC8B | 1550 IDE_TIMING_CYC8B, 1551 IDE_TIMING_ACTIVE = BIT(4), 1552 IDE_TIMING_RECOVER = BIT(5), 1553 IDE_TIMING_CYCLE = BIT(6), 1554 IDE_TIMING_UDMA = BIT(7), 1555 IDE_TIMING_ALL = IDE_TIMING_SETUP | IDE_TIMING_8BIT | 1556 IDE_TIMING_ACTIVE | IDE_TIMING_RECOVER | 1557 IDE_TIMING_CYCLE | IDE_TIMING_UDMA, 1558}; 1559 1560struct ide_timing *ide_timing_find_mode(u8); 1561u16 ide_pio_cycle_time(ide_drive_t *, u8); 1562void ide_timing_merge(struct ide_timing *, struct ide_timing *, 1563 struct ide_timing *, unsigned int); 1564int ide_timing_compute(ide_drive_t *, u8, struct ide_timing *, int, int); 1565 1566#ifdef CONFIG_IDE_XFER_MODE 1567int ide_scan_pio_blacklist(char *); 1568const char *ide_xfer_verbose(u8); 1569int ide_pio_need_iordy(ide_drive_t *, const u8); 1570int ide_set_pio_mode(ide_drive_t *, u8); 1571int ide_set_dma_mode(ide_drive_t *, u8); 1572void ide_set_pio(ide_drive_t *, u8); 1573int ide_set_xfer_rate(ide_drive_t *, u8); 1574#else 1575static inline void ide_set_pio(ide_drive_t *drive, u8 pio) { ; } 1576static inline int ide_set_xfer_rate(ide_drive_t *drive, u8 rate) { return -1; } 1577#endif 1578 1579static inline void ide_set_max_pio(ide_drive_t *drive) 1580{ 1581 ide_set_pio(drive, 255); 1582} 1583 1584char *ide_media_string(ide_drive_t *); 1585 1586extern const struct attribute_group *ide_dev_groups[]; 1587extern struct bus_type ide_bus_type; 1588extern struct class *ide_port_class; 1589 1590static inline void ide_dump_identify(u8 *id) 1591{ 1592 print_hex_dump(KERN_INFO, "", DUMP_PREFIX_NONE, 16, 2, id, 512, 0); 1593} 1594 1595static inline int hwif_to_node(ide_hwif_t *hwif) 1596{ 1597 return hwif->dev ? dev_to_node(hwif->dev) : -1; 1598} 1599 1600static inline ide_drive_t *ide_get_pair_dev(ide_drive_t *drive) 1601{ 1602 ide_drive_t *peer = drive->hwif->devices[(drive->dn ^ 1) & 1]; 1603 1604 return (peer->dev_flags & IDE_DFLAG_PRESENT) ? peer : NULL; 1605} 1606 1607static inline void *ide_get_drivedata(ide_drive_t *drive) 1608{ 1609 return drive->drive_data; 1610} 1611 1612static inline void ide_set_drivedata(ide_drive_t *drive, void *data) 1613{ 1614 drive->drive_data = data; 1615} 1616 1617#define ide_port_for_each_dev(i, dev, port) \ 1618 for ((i) = 0; ((dev) = (port)->devices[i]) || (i) < MAX_DRIVES; (i)++) 1619 1620#define ide_port_for_each_present_dev(i, dev, port) \ 1621 for ((i) = 0; ((dev) = (port)->devices[i]) || (i) < MAX_DRIVES; (i)++) \ 1622 if ((dev)->dev_flags & IDE_DFLAG_PRESENT) 1623 1624#define ide_host_for_each_port(i, port, host) \ 1625 for ((i) = 0; ((port) = (host)->ports[i]) || (i) < MAX_HOST_PORTS; (i)++) 1626 1627 1628#endif /* _IDE_H */