Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v4.8-rc4 278 lines 7.3 kB view raw
1#include <linux/kernel.h> 2#include <linux/gfp.h> 3#include <linux/ide.h> 4 5int generic_ide_suspend(struct device *dev, pm_message_t mesg) 6{ 7 ide_drive_t *drive = to_ide_device(dev); 8 ide_drive_t *pair = ide_get_pair_dev(drive); 9 ide_hwif_t *hwif = drive->hwif; 10 struct request *rq; 11 struct ide_pm_state rqpm; 12 int ret; 13 14 if (ide_port_acpi(hwif)) { 15 /* call ACPI _GTM only once */ 16 if ((drive->dn & 1) == 0 || pair == NULL) 17 ide_acpi_get_timing(hwif); 18 } 19 20 memset(&rqpm, 0, sizeof(rqpm)); 21 rq = blk_get_request(drive->queue, READ, __GFP_RECLAIM); 22 rq->cmd_type = REQ_TYPE_ATA_PM_SUSPEND; 23 rq->special = &rqpm; 24 rqpm.pm_step = IDE_PM_START_SUSPEND; 25 if (mesg.event == PM_EVENT_PRETHAW) 26 mesg.event = PM_EVENT_FREEZE; 27 rqpm.pm_state = mesg.event; 28 29 ret = blk_execute_rq(drive->queue, NULL, rq, 0); 30 blk_put_request(rq); 31 32 if (ret == 0 && ide_port_acpi(hwif)) { 33 /* call ACPI _PS3 only after both devices are suspended */ 34 if ((drive->dn & 1) || pair == NULL) 35 ide_acpi_set_state(hwif, 0); 36 } 37 38 return ret; 39} 40 41static void ide_end_sync_rq(struct request *rq, int error) 42{ 43 complete(rq->end_io_data); 44} 45 46static int ide_pm_execute_rq(struct request *rq) 47{ 48 struct request_queue *q = rq->q; 49 DECLARE_COMPLETION_ONSTACK(wait); 50 51 rq->end_io_data = &wait; 52 rq->end_io = ide_end_sync_rq; 53 54 spin_lock_irq(q->queue_lock); 55 if (unlikely(blk_queue_dying(q))) { 56 rq->cmd_flags |= REQ_QUIET; 57 rq->errors = -ENXIO; 58 __blk_end_request_all(rq, rq->errors); 59 spin_unlock_irq(q->queue_lock); 60 return -ENXIO; 61 } 62 __elv_add_request(q, rq, ELEVATOR_INSERT_FRONT); 63 __blk_run_queue_uncond(q); 64 spin_unlock_irq(q->queue_lock); 65 66 wait_for_completion_io(&wait); 67 68 return rq->errors ? -EIO : 0; 69} 70 71int generic_ide_resume(struct device *dev) 72{ 73 ide_drive_t *drive = to_ide_device(dev); 74 ide_drive_t *pair = ide_get_pair_dev(drive); 75 ide_hwif_t *hwif = drive->hwif; 76 struct request *rq; 77 struct ide_pm_state rqpm; 78 int err; 79 80 if (ide_port_acpi(hwif)) { 81 /* call ACPI _PS0 / _STM only once */ 82 if ((drive->dn & 1) == 0 || pair == NULL) { 83 ide_acpi_set_state(hwif, 1); 84 ide_acpi_push_timing(hwif); 85 } 86 87 ide_acpi_exec_tfs(drive); 88 } 89 90 memset(&rqpm, 0, sizeof(rqpm)); 91 rq = blk_get_request(drive->queue, READ, __GFP_RECLAIM); 92 rq->cmd_type = REQ_TYPE_ATA_PM_RESUME; 93 rq->cmd_flags |= REQ_PREEMPT; 94 rq->special = &rqpm; 95 rqpm.pm_step = IDE_PM_START_RESUME; 96 rqpm.pm_state = PM_EVENT_ON; 97 98 err = ide_pm_execute_rq(rq); 99 blk_put_request(rq); 100 101 if (err == 0 && dev->driver) { 102 struct ide_driver *drv = to_ide_driver(dev->driver); 103 104 if (drv->resume) 105 drv->resume(drive); 106 } 107 108 return err; 109} 110 111void ide_complete_power_step(ide_drive_t *drive, struct request *rq) 112{ 113 struct ide_pm_state *pm = rq->special; 114 115#ifdef DEBUG_PM 116 printk(KERN_INFO "%s: complete_power_step(step: %d)\n", 117 drive->name, pm->pm_step); 118#endif 119 if (drive->media != ide_disk) 120 return; 121 122 switch (pm->pm_step) { 123 case IDE_PM_FLUSH_CACHE: /* Suspend step 1 (flush cache) */ 124 if (pm->pm_state == PM_EVENT_FREEZE) 125 pm->pm_step = IDE_PM_COMPLETED; 126 else 127 pm->pm_step = IDE_PM_STANDBY; 128 break; 129 case IDE_PM_STANDBY: /* Suspend step 2 (standby) */ 130 pm->pm_step = IDE_PM_COMPLETED; 131 break; 132 case IDE_PM_RESTORE_PIO: /* Resume step 1 (restore PIO) */ 133 pm->pm_step = IDE_PM_IDLE; 134 break; 135 case IDE_PM_IDLE: /* Resume step 2 (idle)*/ 136 pm->pm_step = IDE_PM_RESTORE_DMA; 137 break; 138 } 139} 140 141ide_startstop_t ide_start_power_step(ide_drive_t *drive, struct request *rq) 142{ 143 struct ide_pm_state *pm = rq->special; 144 struct ide_cmd cmd = { }; 145 146 switch (pm->pm_step) { 147 case IDE_PM_FLUSH_CACHE: /* Suspend step 1 (flush cache) */ 148 if (drive->media != ide_disk) 149 break; 150 /* Not supported? Switch to next step now. */ 151 if (ata_id_flush_enabled(drive->id) == 0 || 152 (drive->dev_flags & IDE_DFLAG_WCACHE) == 0) { 153 ide_complete_power_step(drive, rq); 154 return ide_stopped; 155 } 156 if (ata_id_flush_ext_enabled(drive->id)) 157 cmd.tf.command = ATA_CMD_FLUSH_EXT; 158 else 159 cmd.tf.command = ATA_CMD_FLUSH; 160 goto out_do_tf; 161 case IDE_PM_STANDBY: /* Suspend step 2 (standby) */ 162 cmd.tf.command = ATA_CMD_STANDBYNOW1; 163 goto out_do_tf; 164 case IDE_PM_RESTORE_PIO: /* Resume step 1 (restore PIO) */ 165 ide_set_max_pio(drive); 166 /* 167 * skip IDE_PM_IDLE for ATAPI devices 168 */ 169 if (drive->media != ide_disk) 170 pm->pm_step = IDE_PM_RESTORE_DMA; 171 else 172 ide_complete_power_step(drive, rq); 173 return ide_stopped; 174 case IDE_PM_IDLE: /* Resume step 2 (idle) */ 175 cmd.tf.command = ATA_CMD_IDLEIMMEDIATE; 176 goto out_do_tf; 177 case IDE_PM_RESTORE_DMA: /* Resume step 3 (restore DMA) */ 178 /* 179 * Right now, all we do is call ide_set_dma(drive), 180 * we could be smarter and check for current xfer_speed 181 * in struct drive etc... 182 */ 183 if (drive->hwif->dma_ops == NULL) 184 break; 185 /* 186 * TODO: respect IDE_DFLAG_USING_DMA 187 */ 188 ide_set_dma(drive); 189 break; 190 } 191 192 pm->pm_step = IDE_PM_COMPLETED; 193 194 return ide_stopped; 195 196out_do_tf: 197 cmd.valid.out.tf = IDE_VALID_OUT_TF | IDE_VALID_DEVICE; 198 cmd.valid.in.tf = IDE_VALID_IN_TF | IDE_VALID_DEVICE; 199 cmd.protocol = ATA_PROT_NODATA; 200 201 return do_rw_taskfile(drive, &cmd); 202} 203 204/** 205 * ide_complete_pm_rq - end the current Power Management request 206 * @drive: target drive 207 * @rq: request 208 * 209 * This function cleans up the current PM request and stops the queue 210 * if necessary. 211 */ 212void ide_complete_pm_rq(ide_drive_t *drive, struct request *rq) 213{ 214 struct request_queue *q = drive->queue; 215 struct ide_pm_state *pm = rq->special; 216 unsigned long flags; 217 218 ide_complete_power_step(drive, rq); 219 if (pm->pm_step != IDE_PM_COMPLETED) 220 return; 221 222#ifdef DEBUG_PM 223 printk("%s: completing PM request, %s\n", drive->name, 224 (rq->cmd_type == REQ_TYPE_ATA_PM_SUSPEND) ? "suspend" : "resume"); 225#endif 226 spin_lock_irqsave(q->queue_lock, flags); 227 if (rq->cmd_type == REQ_TYPE_ATA_PM_SUSPEND) 228 blk_stop_queue(q); 229 else 230 drive->dev_flags &= ~IDE_DFLAG_BLOCKED; 231 spin_unlock_irqrestore(q->queue_lock, flags); 232 233 drive->hwif->rq = NULL; 234 235 if (blk_end_request(rq, 0, 0)) 236 BUG(); 237} 238 239void ide_check_pm_state(ide_drive_t *drive, struct request *rq) 240{ 241 struct ide_pm_state *pm = rq->special; 242 243 if (rq->cmd_type == REQ_TYPE_ATA_PM_SUSPEND && 244 pm->pm_step == IDE_PM_START_SUSPEND) 245 /* Mark drive blocked when starting the suspend sequence. */ 246 drive->dev_flags |= IDE_DFLAG_BLOCKED; 247 else if (rq->cmd_type == REQ_TYPE_ATA_PM_RESUME && 248 pm->pm_step == IDE_PM_START_RESUME) { 249 /* 250 * The first thing we do on wakeup is to wait for BSY bit to 251 * go away (with a looong timeout) as a drive on this hwif may 252 * just be POSTing itself. 253 * We do that before even selecting as the "other" device on 254 * the bus may be broken enough to walk on our toes at this 255 * point. 256 */ 257 ide_hwif_t *hwif = drive->hwif; 258 const struct ide_tp_ops *tp_ops = hwif->tp_ops; 259 struct request_queue *q = drive->queue; 260 unsigned long flags; 261 int rc; 262#ifdef DEBUG_PM 263 printk("%s: Wakeup request inited, waiting for !BSY...\n", drive->name); 264#endif 265 rc = ide_wait_not_busy(hwif, 35000); 266 if (rc) 267 printk(KERN_WARNING "%s: bus not ready on wakeup\n", drive->name); 268 tp_ops->dev_select(drive); 269 tp_ops->write_devctl(hwif, ATA_DEVCTL_OBS); 270 rc = ide_wait_not_busy(hwif, 100000); 271 if (rc) 272 printk(KERN_WARNING "%s: drive not ready on wakeup\n", drive->name); 273 274 spin_lock_irqsave(q->queue_lock, flags); 275 blk_start_queue(q); 276 spin_unlock_irqrestore(q->queue_lock, flags); 277 } 278}