Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

[SCSI] bfa: remove unused functions and misc cleanups

Remove unused functions and fix checkpatch errors. Misc cleanups in
comment and formatting.

Signed-off-by: Maggie Zhang <xmzhang@brocade.com>
Signed-off-by: Jing Huang <huangj@brocade.com>
Signed-off-by: James Bottomley <James.Bottomley@suse.de>

authored by

Maggie Zhang and committed by
James Bottomley
da99dcc9 f16a1750

+110 -457
+1 -1
drivers/scsi/bfa/bfa_core.c
··· 342 342 intr = readl(bfa->iocfc.bfa_regs.intr_status); 343 343 344 344 if (intr & (__HFN_INT_MBOX_LPU0 | __HFN_INT_MBOX_LPU1)) 345 - bfa_ioc_mbox_isr(&bfa->ioc); 345 + bfa_ioc_mbox_isr(&bfa->ioc); 346 346 347 347 intr &= (__HFN_INT_ERR_EMC | __HFN_INT_ERR_LPU0 | 348 348 __HFN_INT_ERR_LPU1 | __HFN_INT_ERR_PSS | __HFN_INT_LL_HALT);
+1 -6
drivers/scsi/bfa/bfa_fcbuild.c
··· 456 456 return sizeof(struct fc_logo_s); 457 457 } 458 458 459 - static u16 459 + static u16 460 460 fc_adisc_x_build(struct fchs_s *fchs, struct fc_adisc_s *adisc, u32 d_id, 461 461 u32 s_id, __be16 ox_id, wwn_t port_name, 462 462 wwn_t node_name, u8 els_code) ··· 806 806 807 807 return sizeof(struct fc_rpsc_acc_s); 808 808 } 809 - 810 - /* 811 - * TBD - 812 - * . get rid of unnecessary memsets 813 - */ 814 809 815 810 u16 816 811 fc_logo_rsp_parse(struct fchs_s *fchs, int len)
+46 -294
drivers/scsi/bfa/bfa_fcpim.c
··· 21 21 BFA_TRC_FILE(HAL, FCPIM); 22 22 BFA_MODULE(fcpim); 23 23 24 - 25 - #define bfa_fcpim_add_iostats(__l, __r, __stats) \ 26 - (__l->__stats += __r->__stats) 27 - 28 - 29 24 /* 30 25 * BFA ITNIM Related definitions 31 26 */ ··· 68 73 } while (0) 69 74 70 75 /* 71 - * bfa_itnim_sm BFA itnim state machine 76 + * itnim state machine event 72 77 */ 73 - 74 - 75 78 enum bfa_itnim_event { 76 79 BFA_ITNIM_SM_CREATE = 1, /* itnim is created */ 77 80 BFA_ITNIM_SM_ONLINE = 2, /* itnim is online */ ··· 221 228 static void __bfa_cb_ioim_pathtov(void *cbarg, bfa_boolean_t complete); 222 229 static bfa_boolean_t bfa_ioim_is_abortable(struct bfa_ioim_s *ioim); 223 230 224 - 225 231 /* 226 232 * forward declaration of BFA IO state machine 227 233 */ ··· 248 256 enum bfa_ioim_event event); 249 257 static void bfa_ioim_sm_cmnd_retry(struct bfa_ioim_s *ioim, 250 258 enum bfa_ioim_event event); 251 - 252 259 /* 253 260 * forward declaration for BFA TSKIM functions 254 261 */ ··· 261 270 static bfa_boolean_t bfa_tskim_send(struct bfa_tskim_s *tskim); 262 271 static bfa_boolean_t bfa_tskim_send_abort(struct bfa_tskim_s *tskim); 263 272 static void bfa_tskim_iocdisable_ios(struct bfa_tskim_s *tskim); 264 - 265 273 266 274 /* 267 275 * forward declaration of BFA TSKIM state machine ··· 279 289 enum bfa_tskim_event event); 280 290 static void bfa_tskim_sm_hcb(struct bfa_tskim_s *tskim, 281 291 enum bfa_tskim_event event); 282 - 283 292 /* 284 293 * BFA FCP Initiator Mode module 285 294 */ 286 295 287 296 /* 288 - * Compute and return memory needed by FCP(im) module. 297 + * Compute and return memory needed by FCP(im) module. 289 298 */ 290 299 static void 291 300 bfa_fcpim_meminfo(struct bfa_iocfc_cfg_s *cfg, u32 *km_len, ··· 368 379 } 369 380 370 381 void 371 - bfa_fcpim_add_stats(struct bfa_itnim_iostats_s *lstats, 372 - struct bfa_itnim_iostats_s *rstats) 373 - { 374 - bfa_fcpim_add_iostats(lstats, rstats, total_ios); 375 - bfa_fcpim_add_iostats(lstats, rstats, qresumes); 376 - bfa_fcpim_add_iostats(lstats, rstats, no_iotags); 377 - bfa_fcpim_add_iostats(lstats, rstats, io_aborts); 378 - bfa_fcpim_add_iostats(lstats, rstats, no_tskims); 379 - bfa_fcpim_add_iostats(lstats, rstats, iocomp_ok); 380 - bfa_fcpim_add_iostats(lstats, rstats, iocomp_underrun); 381 - bfa_fcpim_add_iostats(lstats, rstats, iocomp_overrun); 382 - bfa_fcpim_add_iostats(lstats, rstats, iocomp_aborted); 383 - bfa_fcpim_add_iostats(lstats, rstats, iocomp_timedout); 384 - bfa_fcpim_add_iostats(lstats, rstats, iocom_nexus_abort); 385 - bfa_fcpim_add_iostats(lstats, rstats, iocom_proto_err); 386 - bfa_fcpim_add_iostats(lstats, rstats, iocom_dif_err); 387 - bfa_fcpim_add_iostats(lstats, rstats, iocom_sqer_needed); 388 - bfa_fcpim_add_iostats(lstats, rstats, iocom_res_free); 389 - bfa_fcpim_add_iostats(lstats, rstats, iocom_hostabrts); 390 - bfa_fcpim_add_iostats(lstats, rstats, iocom_utags); 391 - bfa_fcpim_add_iostats(lstats, rstats, io_cleanups); 392 - bfa_fcpim_add_iostats(lstats, rstats, io_tmaborts); 393 - bfa_fcpim_add_iostats(lstats, rstats, onlines); 394 - bfa_fcpim_add_iostats(lstats, rstats, offlines); 395 - bfa_fcpim_add_iostats(lstats, rstats, creates); 396 - bfa_fcpim_add_iostats(lstats, rstats, deletes); 397 - bfa_fcpim_add_iostats(lstats, rstats, create_comps); 398 - bfa_fcpim_add_iostats(lstats, rstats, delete_comps); 399 - bfa_fcpim_add_iostats(lstats, rstats, sler_events); 400 - bfa_fcpim_add_iostats(lstats, rstats, fw_create); 401 - bfa_fcpim_add_iostats(lstats, rstats, fw_delete); 402 - bfa_fcpim_add_iostats(lstats, rstats, ioc_disabled); 403 - bfa_fcpim_add_iostats(lstats, rstats, cleanup_comps); 404 - bfa_fcpim_add_iostats(lstats, rstats, tm_cmnds); 405 - bfa_fcpim_add_iostats(lstats, rstats, tm_fw_rsps); 406 - bfa_fcpim_add_iostats(lstats, rstats, tm_success); 407 - bfa_fcpim_add_iostats(lstats, rstats, tm_failures); 408 - bfa_fcpim_add_iostats(lstats, rstats, tm_io_comps); 409 - bfa_fcpim_add_iostats(lstats, rstats, tm_qresumes); 410 - bfa_fcpim_add_iostats(lstats, rstats, tm_iocdowns); 411 - bfa_fcpim_add_iostats(lstats, rstats, tm_cleanups); 412 - bfa_fcpim_add_iostats(lstats, rstats, tm_cleanup_comps); 413 - bfa_fcpim_add_iostats(lstats, rstats, io_comps); 414 - bfa_fcpim_add_iostats(lstats, rstats, input_reqs); 415 - bfa_fcpim_add_iostats(lstats, rstats, output_reqs); 416 - bfa_fcpim_add_iostats(lstats, rstats, rd_throughput); 417 - bfa_fcpim_add_iostats(lstats, rstats, wr_throughput); 418 - } 419 - 420 - void 421 382 bfa_fcpim_path_tov_set(struct bfa_s *bfa, u16 path_tov) 422 383 { 423 384 struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); ··· 385 446 return fcpim->path_tov / 1000; 386 447 } 387 448 388 - bfa_status_t 389 - bfa_fcpim_port_iostats(struct bfa_s *bfa, struct bfa_itnim_iostats_s *stats, 390 - u8 lp_tag) 391 - { 392 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 393 - struct list_head *qe, *qen; 394 - struct bfa_itnim_s *itnim; 395 - 396 - /* accumulate IO stats from itnim */ 397 - memset(stats, 0, sizeof(struct bfa_itnim_iostats_s)); 398 - list_for_each_safe(qe, qen, &fcpim->itnim_q) { 399 - itnim = (struct bfa_itnim_s *) qe; 400 - if (itnim->rport->rport_info.lp_tag != lp_tag) 401 - continue; 402 - bfa_fcpim_add_stats(stats, &(itnim->stats)); 403 - } 404 - return BFA_STATUS_OK; 405 - } 406 - bfa_status_t 407 - bfa_fcpim_get_modstats(struct bfa_s *bfa, struct bfa_itnim_iostats_s *modstats) 408 - { 409 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 410 - struct list_head *qe, *qen; 411 - struct bfa_itnim_s *itnim; 412 - 413 - /* accumulate IO stats from itnim */ 414 - memset(modstats, 0, sizeof(struct bfa_itnim_iostats_s)); 415 - list_for_each_safe(qe, qen, &fcpim->itnim_q) { 416 - itnim = (struct bfa_itnim_s *) qe; 417 - bfa_fcpim_add_stats(modstats, &(itnim->stats)); 418 - } 419 - return BFA_STATUS_OK; 420 - } 421 - 422 - bfa_status_t 423 - bfa_fcpim_get_del_itn_stats(struct bfa_s *bfa, 424 - struct bfa_fcpim_del_itn_stats_s *modstats) 425 - { 426 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 427 - 428 - *modstats = fcpim->del_itn_stats; 429 - 430 - return BFA_STATUS_OK; 431 - } 432 - 433 - 434 - bfa_status_t 435 - bfa_fcpim_profile_on(struct bfa_s *bfa, u32 time) 436 - { 437 - struct bfa_itnim_s *itnim; 438 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 439 - struct list_head *qe, *qen; 440 - 441 - /* accumulate IO stats from itnim */ 442 - list_for_each_safe(qe, qen, &fcpim->itnim_q) { 443 - itnim = (struct bfa_itnim_s *) qe; 444 - bfa_itnim_clear_stats(itnim); 445 - } 446 - fcpim->io_profile = BFA_TRUE; 447 - fcpim->io_profile_start_time = time; 448 - fcpim->profile_comp = bfa_ioim_profile_comp; 449 - fcpim->profile_start = bfa_ioim_profile_start; 450 - 451 - return BFA_STATUS_OK; 452 - } 453 - bfa_status_t 454 - bfa_fcpim_profile_off(struct bfa_s *bfa) 455 - { 456 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 457 - fcpim->io_profile = BFA_FALSE; 458 - fcpim->io_profile_start_time = 0; 459 - fcpim->profile_comp = NULL; 460 - fcpim->profile_start = NULL; 461 - return BFA_STATUS_OK; 462 - } 463 - 464 - bfa_status_t 465 - bfa_fcpim_port_clear_iostats(struct bfa_s *bfa, u8 lp_tag) 466 - { 467 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 468 - struct list_head *qe, *qen; 469 - struct bfa_itnim_s *itnim; 470 - 471 - /* clear IO stats from all active itnims */ 472 - list_for_each_safe(qe, qen, &fcpim->itnim_q) { 473 - itnim = (struct bfa_itnim_s *) qe; 474 - if (itnim->rport->rport_info.lp_tag != lp_tag) 475 - continue; 476 - bfa_itnim_clear_stats(itnim); 477 - } 478 - return BFA_STATUS_OK; 479 - 480 - } 481 - 482 - bfa_status_t 483 - bfa_fcpim_clr_modstats(struct bfa_s *bfa) 484 - { 485 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 486 - struct list_head *qe, *qen; 487 - struct bfa_itnim_s *itnim; 488 - 489 - /* clear IO stats from all active itnims */ 490 - list_for_each_safe(qe, qen, &fcpim->itnim_q) { 491 - itnim = (struct bfa_itnim_s *) qe; 492 - bfa_itnim_clear_stats(itnim); 493 - } 494 - memset(&fcpim->del_itn_stats, 0, 495 - sizeof(struct bfa_fcpim_del_itn_stats_s)); 496 - 497 - return BFA_STATUS_OK; 498 - } 499 - 500 - void 501 - bfa_fcpim_qdepth_set(struct bfa_s *bfa, u16 q_depth) 502 - { 503 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 504 - 505 - bfa_assert(q_depth <= BFA_IOCFC_QDEPTH_MAX); 506 - 507 - fcpim->q_depth = q_depth; 508 - } 509 - 510 449 u16 511 450 bfa_fcpim_qdepth_get(struct bfa_s *bfa) 512 451 { ··· 393 576 return fcpim->q_depth; 394 577 } 395 578 396 - void 397 - bfa_fcpim_update_ioredirect(struct bfa_s *bfa) 398 - { 399 - bfa_boolean_t ioredirect; 400 - 401 - /* 402 - * IO redirection is turned off when QoS is enabled and vice versa 403 - */ 404 - ioredirect = bfa_fcport_is_qos_enabled(bfa) ? BFA_FALSE : BFA_TRUE; 405 - } 406 - 407 - void 408 - bfa_fcpim_set_ioredirect(struct bfa_s *bfa, bfa_boolean_t state) 409 - { 410 - struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); 411 - fcpim->ioredirect = state; 412 - } 413 - 414 - 415 - 416 579 /* 417 580 * BFA ITNIM module state machine functions 418 581 */ 419 582 420 583 /* 421 - * Beginning/unallocated state - no events expected. 584 + * Beginning/unallocated state - no events expected. 422 585 */ 423 586 static void 424 587 bfa_itnim_sm_uninit(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 419 622 } 420 623 421 624 /* 422 - * Beginning state, only online event expected. 625 + * Beginning state, only online event expected. 423 626 */ 424 627 static void 425 628 bfa_itnim_sm_created(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 522 725 } 523 726 524 727 /* 525 - * Waiting for itnim create response from firmware, a delete is pending. 728 + * Waiting for itnim create response from firmware, a delete is pending. 526 729 */ 527 730 static void 528 731 bfa_itnim_sm_delete_pending(struct bfa_itnim_s *itnim, ··· 550 753 } 551 754 552 755 /* 553 - * Online state - normal parking state. 756 + * Online state - normal parking state. 554 757 */ 555 758 static void 556 759 bfa_itnim_sm_online(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 592 795 } 593 796 594 797 /* 595 - * Second level error recovery need. 798 + * Second level error recovery need. 596 799 */ 597 800 static void 598 801 bfa_itnim_sm_sler(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 623 826 } 624 827 625 828 /* 626 - * Going offline. Waiting for active IO cleanup. 829 + * Going offline. Waiting for active IO cleanup. 627 830 */ 628 831 static void 629 832 bfa_itnim_sm_cleanup_offline(struct bfa_itnim_s *itnim, ··· 660 863 } 661 864 662 865 /* 663 - * Deleting itnim. Waiting for active IO cleanup. 866 + * Deleting itnim. Waiting for active IO cleanup. 664 867 */ 665 868 static void 666 869 bfa_itnim_sm_cleanup_delete(struct bfa_itnim_s *itnim, ··· 745 948 } 746 949 747 950 /* 748 - * Offline state. 951 + * Offline state. 749 952 */ 750 953 static void 751 954 bfa_itnim_sm_offline(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 776 979 } 777 980 } 778 981 779 - /* 780 - * IOC h/w failed state. 781 - */ 782 982 static void 783 983 bfa_itnim_sm_iocdisable(struct bfa_itnim_s *itnim, 784 984 enum bfa_itnim_event event) ··· 810 1016 } 811 1017 812 1018 /* 813 - * Itnim is deleted, waiting for firmware response to delete. 1019 + * Itnim is deleted, waiting for firmware response to delete. 814 1020 */ 815 1021 static void 816 1022 bfa_itnim_sm_deleting(struct bfa_itnim_s *itnim, enum bfa_itnim_event event) ··· 855 1061 } 856 1062 857 1063 /* 858 - * Initiate cleanup of all IOs on an IOC failure. 1064 + * Initiate cleanup of all IOs on an IOC failure. 859 1065 */ 860 1066 static void 861 1067 bfa_itnim_iocdisable_cleanup(struct bfa_itnim_s *itnim) ··· 889 1095 } 890 1096 891 1097 /* 892 - * IO cleanup completion 1098 + * IO cleanup completion 893 1099 */ 894 1100 static void 895 1101 bfa_itnim_cleanp_comp(void *itnim_cbarg) ··· 901 1107 } 902 1108 903 1109 /* 904 - * Initiate cleanup of all IOs. 1110 + * Initiate cleanup of all IOs. 905 1111 */ 906 1112 static void 907 1113 bfa_itnim_cleanup(struct bfa_itnim_s *itnim) ··· 972 1178 973 1179 bfa_sm_send_event(itnim, BFA_ITNIM_SM_QRESUME); 974 1180 } 975 - 976 - 977 - 978 1181 979 1182 /* 980 1183 * bfa_itnim_public ··· 1240 1449 fcpim->del_itn_stats.del_tm_iocdowns += itnim->stats.tm_iocdowns; 1241 1450 } 1242 1451 1243 - 1244 - 1245 1452 /* 1246 - * bfa_itnim_public 1453 + * bfa_itnim_public 1247 1454 */ 1248 1455 1249 1456 /* 1250 - * Itnim interrupt processing. 1457 + * Itnim interrupt processing. 1251 1458 */ 1252 1459 void 1253 1460 bfa_itnim_isr(struct bfa_s *bfa, struct bfi_msg_s *m) ··· 1288 1499 } 1289 1500 } 1290 1501 1291 - 1292 - 1293 1502 /* 1294 - * bfa_itnim_api 1503 + * bfa_itnim_api 1295 1504 */ 1296 1505 1297 1506 struct bfa_itnim_s * ··· 1362 1575 */ 1363 1576 1364 1577 /* 1365 - * IO is not started (unallocated). 1578 + * IO is not started (unallocated). 1366 1579 */ 1367 1580 static void 1368 1581 bfa_ioim_sm_uninit(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1427 1640 } 1428 1641 1429 1642 /* 1430 - * IO is waiting for SG pages. 1643 + * IO is waiting for SG pages. 1431 1644 */ 1432 1645 static void 1433 1646 bfa_ioim_sm_sgalloc(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1474 1687 } 1475 1688 1476 1689 /* 1477 - * IO is active. 1690 + * IO is active. 1478 1691 */ 1479 1692 static void 1480 1693 bfa_ioim_sm_active(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1558 1771 } 1559 1772 1560 1773 /* 1561 - * IO is retried with new tag. 1562 - */ 1774 + * IO is retried with new tag. 1775 + */ 1563 1776 static void 1564 1777 bfa_ioim_sm_cmnd_retry(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) 1565 1778 { ··· 1612 1825 } 1613 1826 1614 1827 /* 1615 - * IO is being aborted, waiting for completion from firmware. 1828 + * IO is being aborted, waiting for completion from firmware. 1616 1829 */ 1617 1830 static void 1618 1831 bfa_ioim_sm_abort(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1735 1948 } 1736 1949 1737 1950 /* 1738 - * IO is waiting for room in request CQ 1951 + * IO is waiting for room in request CQ 1739 1952 */ 1740 1953 static void 1741 1954 bfa_ioim_sm_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1779 1992 } 1780 1993 1781 1994 /* 1782 - * Active IO is being aborted, waiting for room in request CQ. 1995 + * Active IO is being aborted, waiting for room in request CQ. 1783 1996 */ 1784 1997 static void 1785 1998 bfa_ioim_sm_abort_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 1830 2043 } 1831 2044 1832 2045 /* 1833 - * Active IO is being cleaned up, waiting for room in request CQ. 2046 + * Active IO is being cleaned up, waiting for room in request CQ. 1834 2047 */ 1835 2048 static void 1836 2049 bfa_ioim_sm_cleanup_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) ··· 2084 2297 { 2085 2298 struct bfa_itnim_s *itnim = ioim->itnim; 2086 2299 struct bfi_ioim_req_s *m; 2087 - static struct fcp_cmnd_s cmnd_z0 = {{{0}}}; 2300 + static struct fcp_cmnd_s cmnd_z0 = { { { 0 } } }; 2088 2301 struct bfi_sge_s *sge, *sgpge; 2089 2302 u32 pgdlen = 0; 2090 2303 u32 fcp_dl; ··· 2340 2553 return BFA_TRUE; 2341 2554 } 2342 2555 2343 - /* 2344 - * or after the link comes back. 2345 - */ 2346 2556 void 2347 2557 bfa_ioim_delayed_comp(struct bfa_ioim_s *ioim, bfa_boolean_t iotov) 2348 2558 { ··· 2534 2750 bfa_sm_send_event(ioim, BFA_IOIM_SM_COMP_GOOD); 2535 2751 } 2536 2752 2537 - void 2538 - bfa_ioim_profile_start(struct bfa_ioim_s *ioim) 2539 - { 2540 - ioim->start_time = jiffies; 2541 - } 2542 - 2543 - void 2544 - bfa_ioim_profile_comp(struct bfa_ioim_s *ioim) 2545 - { 2546 - struct scsi_cmnd *cmnd = (struct scsi_cmnd *) ioim->dio; 2547 - u32 fcp_dl = scsi_bufflen(cmnd); 2548 - u32 index = bfa_ioim_get_index(fcp_dl); 2549 - u64 end_time = jiffies; 2550 - struct bfa_itnim_latency_s *io_lat = 2551 - &(ioim->itnim->ioprofile.io_latency); 2552 - u32 val = (u32)(end_time - ioim->start_time); 2553 - 2554 - bfa_itnim_ioprofile_update(ioim->itnim, index); 2555 - 2556 - io_lat->count[index]++; 2557 - io_lat->min[index] = (io_lat->min[index] < val) ? 2558 - io_lat->min[index] : val; 2559 - io_lat->max[index] = (io_lat->max[index] > val) ? 2560 - io_lat->max[index] : val; 2561 - io_lat->avg[index] += val; 2562 - } 2563 2753 /* 2564 2754 * Called by itnim to clean up IO while going offline. 2565 2755 */ ··· 2667 2909 return BFA_STATUS_OK; 2668 2910 } 2669 2911 2670 - 2671 2912 /* 2672 2913 * BFA TSKIM state machine functions 2673 2914 */ 2674 2915 2675 2916 /* 2676 - * Task management command beginning state. 2917 + * Task management command beginning state. 2677 2918 */ 2678 2919 static void 2679 2920 bfa_tskim_sm_uninit(struct bfa_tskim_s *tskim, enum bfa_tskim_event event) ··· 2709 2952 } 2710 2953 2711 2954 /* 2712 - * brief 2713 - * TM command is active, awaiting completion from firmware to 2714 - * cleanup IO requests in TM scope. 2955 + * TM command is active, awaiting completion from firmware to 2956 + * cleanup IO requests in TM scope. 2715 2957 */ 2716 2958 static void 2717 2959 bfa_tskim_sm_active(struct bfa_tskim_s *tskim, enum bfa_tskim_event event) ··· 2745 2989 } 2746 2990 2747 2991 /* 2748 - * An active TM is being cleaned up since ITN is offline. Awaiting cleanup 2749 - * completion event from firmware. 2992 + * An active TM is being cleaned up since ITN is offline. Awaiting cleanup 2993 + * completion event from firmware. 2750 2994 */ 2751 2995 static void 2752 2996 bfa_tskim_sm_cleanup(struct bfa_tskim_s *tskim, enum bfa_tskim_event event) ··· 2806 3050 } 2807 3051 2808 3052 /* 2809 - * Task management command is waiting for room in request CQ 3053 + * Task management command is waiting for room in request CQ 2810 3054 */ 2811 3055 static void 2812 3056 bfa_tskim_sm_qfull(struct bfa_tskim_s *tskim, enum bfa_tskim_event event) ··· 2841 3085 } 2842 3086 2843 3087 /* 2844 - * Task management command is active, awaiting for room in request CQ 2845 - * to send clean up request. 3088 + * Task management command is active, awaiting for room in request CQ 3089 + * to send clean up request. 2846 3090 */ 2847 3091 static void 2848 3092 bfa_tskim_sm_cleanup_qfull(struct bfa_tskim_s *tskim, ··· 2854 3098 case BFA_TSKIM_SM_DONE: 2855 3099 bfa_reqq_wcancel(&tskim->reqq_wait); 2856 3100 /* 2857 - * 2858 3101 * Fall through !!! 2859 3102 */ 2860 - 2861 3103 case BFA_TSKIM_SM_QRESUME: 2862 3104 bfa_sm_set_state(tskim, bfa_tskim_sm_cleanup); 2863 3105 bfa_tskim_send_abort(tskim); ··· 2874 3120 } 2875 3121 2876 3122 /* 2877 - * BFA callback is pending 3123 + * BFA callback is pending 2878 3124 */ 2879 3125 static void 2880 3126 bfa_tskim_sm_hcb(struct bfa_tskim_s *tskim, enum bfa_tskim_event event) ··· 2898 3144 bfa_sm_fault(tskim->bfa, event); 2899 3145 } 2900 3146 } 2901 - 2902 3147 2903 3148 static void 2904 3149 __bfa_cb_tskim_done(void *cbarg, bfa_boolean_t complete) ··· 2928 3175 BFI_TSKIM_STS_FAILED); 2929 3176 } 2930 3177 2931 - static bfa_boolean_t 3178 + static bfa_boolean_t 2932 3179 bfa_tskim_match_scope(struct bfa_tskim_s *tskim, struct scsi_lun lun) 2933 3180 { 2934 3181 switch (tskim->tm_cmnd) { ··· 2939 3186 case FCP_TM_CLEAR_TASK_SET: 2940 3187 case FCP_TM_LUN_RESET: 2941 3188 case FCP_TM_CLEAR_ACA: 2942 - return (!memcmp(&tskim->lun, &lun, sizeof(lun))); 3189 + return !memcmp(&tskim->lun, &lun, sizeof(lun)); 2943 3190 2944 3191 default: 2945 3192 bfa_assert(0); ··· 2949 3196 } 2950 3197 2951 3198 /* 2952 - * Gather affected IO requests and task management commands. 3199 + * Gather affected IO requests and task management commands. 2953 3200 */ 2954 3201 static void 2955 3202 bfa_tskim_gather_ios(struct bfa_tskim_s *tskim) ··· 2991 3238 } 2992 3239 2993 3240 /* 2994 - * IO cleanup completion 3241 + * IO cleanup completion 2995 3242 */ 2996 3243 static void 2997 3244 bfa_tskim_cleanp_comp(void *tskim_cbarg) ··· 3003 3250 } 3004 3251 3005 3252 /* 3006 - * Gather affected IO requests and task management commands. 3253 + * Gather affected IO requests and task management commands. 3007 3254 */ 3008 3255 static void 3009 3256 bfa_tskim_cleanup_ios(struct bfa_tskim_s *tskim) ··· 3023 3270 } 3024 3271 3025 3272 /* 3026 - * Send task management request to firmware. 3273 + * Send task management request to firmware. 3027 3274 */ 3028 3275 static bfa_boolean_t 3029 3276 bfa_tskim_send(struct bfa_tskim_s *tskim) ··· 3058 3305 } 3059 3306 3060 3307 /* 3061 - * Send abort request to cleanup an active TM to firmware. 3308 + * Send abort request to cleanup an active TM to firmware. 3062 3309 */ 3063 3310 static bfa_boolean_t 3064 3311 bfa_tskim_send_abort(struct bfa_tskim_s *tskim) ··· 3089 3336 } 3090 3337 3091 3338 /* 3092 - * Call to resume task management cmnd waiting for room in request queue. 3339 + * Call to resume task management cmnd waiting for room in request queue. 3093 3340 */ 3094 3341 static void 3095 3342 bfa_tskim_qresume(void *cbarg) ··· 3114 3361 bfa_ioim_iocdisable(ioim); 3115 3362 } 3116 3363 } 3117 - 3118 3364 3119 3365 /* 3120 3366 * Notification on completions from related ioim. ··· 3147 3395 } 3148 3396 3149 3397 /* 3150 - * Memory allocation and initialization. 3398 + * Memory allocation and initialization. 3151 3399 */ 3152 3400 void 3153 3401 bfa_tskim_attach(struct bfa_fcpim_mod_s *fcpim, struct bfa_meminfo_s *minfo) ··· 3229 3477 } 3230 3478 3231 3479 /* 3232 - * Start a task management command. 3480 + * Start a task management command. 3233 3481 * 3234 3482 * @param[in] tskim BFA task management command instance 3235 3483 * @param[in] itnim i-t nexus for the task management command
+57 -101
drivers/scsi/bfa/bfa_fcpim.h
··· 94 94 struct list_head ioim_resfree_q; /* IOs waiting for f/w */ 95 95 struct list_head ioim_comp_q; /* IO global comp Q */ 96 96 struct list_head tskim_free_q; 97 - u32 ios_active; /* current active IOs */ 98 - u32 delay_comp; 97 + u32 ios_active; /* current active IOs */ 98 + u32 delay_comp; 99 99 struct bfa_fcpim_del_itn_stats_s del_itn_stats; 100 100 bfa_boolean_t ioredirect; 101 101 bfa_boolean_t io_profile; 102 - u32 io_profile_start_time; 102 + u32 io_profile_start_time; 103 103 bfa_fcpim_profile_t profile_comp; 104 104 bfa_fcpim_profile_t profile_start; 105 105 }; ··· 114 114 struct bfa_fcpim_mod_s *fcpim; /* parent fcpim module */ 115 115 struct bfa_itnim_s *itnim; /* i-t-n nexus for this IO */ 116 116 struct bfad_ioim_s *dio; /* driver IO handle */ 117 - u16 iotag; /* FWI IO tag */ 118 - u16 abort_tag; /* unqiue abort request tag */ 119 - u16 nsges; /* number of SG elements */ 120 - u16 nsgpgs; /* number of SG pages */ 117 + u16 iotag; /* FWI IO tag */ 118 + u16 abort_tag; /* unqiue abort request tag */ 119 + u16 nsges; /* number of SG elements */ 120 + u16 nsgpgs; /* number of SG pages */ 121 121 struct bfa_sgpg_s *sgpg; /* first SG page */ 122 122 struct list_head sgpg_q; /* allocated SG pages */ 123 123 struct bfa_cb_qe_s hcb_qe; /* bfa callback qelem */ 124 124 bfa_cb_cbfn_t io_cbfn; /* IO completion handler */ 125 - struct bfa_ioim_sp_s *iosp; /* slow-path IO handling */ 126 - u8 reqq; /* Request queue for I/O */ 127 - u64 start_time; /* IO's Profile start val */ 125 + struct bfa_ioim_sp_s *iosp; /* slow-path IO handling */ 126 + u8 reqq; /* Request queue for I/O */ 127 + u64 start_time; /* IO's Profile start val */ 128 128 }; 129 - 130 129 131 130 struct bfa_ioim_sp_s { 132 131 struct bfi_msg_s comp_rspmsg; /* IO comp f/w response */ 133 132 u8 *snsinfo; /* sense info for this IO */ 134 - struct bfa_sgpg_wqe_s sgpg_wqe; /* waitq elem for sgpg */ 135 - struct bfa_reqq_wait_s reqq_wait; /* to wait for room in reqq */ 133 + struct bfa_sgpg_wqe_s sgpg_wqe; /* waitq elem for sgpg */ 134 + struct bfa_reqq_wait_s reqq_wait; /* to wait for room in reqq */ 136 135 bfa_boolean_t abort_explicit; /* aborted by OS */ 137 136 struct bfa_tskim_s *tskim; /* Relevant TM cmd */ 138 137 }; ··· 142 143 struct bfa_tskim_s { 143 144 struct list_head qe; 144 145 bfa_sm_t sm; 145 - struct bfa_s *bfa; /* BFA module */ 146 + struct bfa_s *bfa; /* BFA module */ 146 147 struct bfa_fcpim_mod_s *fcpim; /* parent fcpim module */ 147 148 struct bfa_itnim_s *itnim; /* i-t-n nexus for this IO */ 148 149 struct bfad_tskim_s *dtsk; /* driver task mgmt cmnd */ 149 - bfa_boolean_t notify; /* notify itnim on TM comp */ 150 + bfa_boolean_t notify; /* notify itnim on TM comp */ 150 151 struct scsi_lun lun; /* lun if applicable */ 151 - enum fcp_tm_cmnd tm_cmnd; /* task management command */ 152 - u16 tsk_tag; /* FWI IO tag */ 153 - u8 tsecs; /* timeout in seconds */ 152 + enum fcp_tm_cmnd tm_cmnd; /* task management command */ 153 + u16 tsk_tag; /* FWI IO tag */ 154 + u8 tsecs; /* timeout in seconds */ 154 155 struct bfa_reqq_wait_s reqq_wait; /* to wait for room in reqq */ 155 156 struct list_head io_q; /* queue of affected IOs */ 156 - struct bfa_wc_s wc; /* waiting counter */ 157 + struct bfa_wc_s wc; /* waiting counter */ 157 158 struct bfa_cb_qe_s hcb_qe; /* bfa callback qelem */ 158 159 enum bfi_tskim_status tsk_status; /* TM status */ 159 160 }; 160 - 161 161 162 162 /* 163 163 * BFA i-t-n (initiator mode) 164 164 */ 165 165 struct bfa_itnim_s { 166 - struct list_head qe; /* queue element */ 167 - bfa_sm_t sm; /* i-t-n im BFA state machine */ 168 - struct bfa_s *bfa; /* bfa instance */ 169 - struct bfa_rport_s *rport; /* bfa rport */ 170 - void *ditn; /* driver i-t-n structure */ 166 + struct list_head qe; /* queue element */ 167 + bfa_sm_t sm; /* i-t-n im BFA state machine */ 168 + struct bfa_s *bfa; /* bfa instance */ 169 + struct bfa_rport_s *rport; /* bfa rport */ 170 + void *ditn; /* driver i-t-n structure */ 171 171 struct bfi_mhdr_s mhdr; /* pre-built mhdr */ 172 - u8 msg_no; /* itnim/rport firmware handle */ 173 - u8 reqq; /* CQ for requests */ 172 + u8 msg_no; /* itnim/rport firmware handle */ 173 + u8 reqq; /* CQ for requests */ 174 174 struct bfa_cb_qe_s hcb_qe; /* bfa callback qelem */ 175 175 struct list_head pending_q; /* queue of pending IO requests */ 176 176 struct list_head io_q; /* queue of active IO requests */ ··· 179 181 bfa_boolean_t seq_rec; /* SQER supported */ 180 182 bfa_boolean_t is_online; /* itnim is ONLINE for IO */ 181 183 bfa_boolean_t iotov_active; /* IO TOV timer is active */ 182 - struct bfa_wc_s wc; /* waiting counter */ 183 - struct bfa_timer_s timer; /* pending IO TOV */ 184 + struct bfa_wc_s wc; /* waiting counter */ 185 + struct bfa_timer_s timer; /* pending IO TOV */ 184 186 struct bfa_reqq_wait_s reqq_wait; /* to wait for room in reqq */ 185 187 struct bfa_fcpim_mod_s *fcpim; /* fcpim module */ 186 188 struct bfa_itnim_iostats_s stats; 187 189 struct bfa_itnim_ioprofile_s ioprofile; 188 190 }; 189 - 190 191 191 192 #define bfa_itnim_is_online(_itnim) ((_itnim)->is_online) 192 193 #define BFA_FCPIM_MOD(_hal) (&(_hal)->modules.fcpim_mod) ··· 243 246 void bfa_itnim_iodone(struct bfa_itnim_s *itnim); 244 247 void bfa_itnim_tskdone(struct bfa_itnim_s *itnim); 245 248 bfa_boolean_t bfa_itnim_hold_io(struct bfa_itnim_s *itnim); 246 - void bfa_ioim_profile_comp(struct bfa_ioim_s *ioim); 247 - void bfa_ioim_profile_start(struct bfa_ioim_s *ioim); 248 - 249 249 250 250 /* 251 251 * bfa fcpim module API functions 252 252 */ 253 - void bfa_fcpim_path_tov_set(struct bfa_s *bfa, u16 path_tov); 253 + void bfa_fcpim_path_tov_set(struct bfa_s *bfa, u16 path_tov); 254 254 u16 bfa_fcpim_path_tov_get(struct bfa_s *bfa); 255 - void bfa_fcpim_qdepth_set(struct bfa_s *bfa, u16 q_depth); 256 255 u16 bfa_fcpim_qdepth_get(struct bfa_s *bfa); 257 - bfa_status_t bfa_fcpim_get_modstats(struct bfa_s *bfa, 258 - struct bfa_itnim_iostats_s *modstats); 259 - bfa_status_t bfa_fcpim_port_iostats(struct bfa_s *bfa, 260 - struct bfa_itnim_iostats_s *stats, u8 lp_tag); 261 - bfa_status_t bfa_fcpim_get_del_itn_stats(struct bfa_s *bfa, 262 - struct bfa_fcpim_del_itn_stats_s *modstats); 263 - bfa_status_t bfa_fcpim_port_clear_iostats(struct bfa_s *bfa, u8 lp_tag); 264 - void bfa_fcpim_add_stats(struct bfa_itnim_iostats_s *fcpim_stats, 265 - struct bfa_itnim_iostats_s *itnim_stats); 266 - bfa_status_t bfa_fcpim_clr_modstats(struct bfa_s *bfa); 267 - void bfa_fcpim_set_ioredirect(struct bfa_s *bfa, 268 - bfa_boolean_t state); 269 - void bfa_fcpim_update_ioredirect(struct bfa_s *bfa); 270 - bfa_status_t bfa_fcpim_profile_on(struct bfa_s *bfa, u32 time); 271 - bfa_status_t bfa_fcpim_profile_off(struct bfa_s *bfa); 256 + 272 257 #define bfa_fcpim_ioredirect_enabled(__bfa) \ 273 258 (((struct bfa_fcpim_mod_s *)(BFA_FCPIM_MOD(__bfa)))->ioredirect) 274 259 ··· 268 289 * bfa itnim API functions 269 290 */ 270 291 struct bfa_itnim_s *bfa_itnim_create(struct bfa_s *bfa, 271 - struct bfa_rport_s *rport, void *itnim); 272 - void bfa_itnim_delete(struct bfa_itnim_s *itnim); 273 - void bfa_itnim_online(struct bfa_itnim_s *itnim, 274 - bfa_boolean_t seq_rec); 275 - void bfa_itnim_offline(struct bfa_itnim_s *itnim); 276 - void bfa_itnim_clear_stats(struct bfa_itnim_s *itnim); 277 - bfa_status_t bfa_itnim_get_ioprofile(struct bfa_itnim_s *itnim, 278 - struct bfa_itnim_ioprofile_s *ioprofile); 292 + struct bfa_rport_s *rport, void *itnim); 293 + void bfa_itnim_delete(struct bfa_itnim_s *itnim); 294 + void bfa_itnim_online(struct bfa_itnim_s *itnim, bfa_boolean_t seq_rec); 295 + void bfa_itnim_offline(struct bfa_itnim_s *itnim); 296 + void bfa_itnim_clear_stats(struct bfa_itnim_s *itnim); 297 + bfa_status_t bfa_itnim_get_ioprofile(struct bfa_itnim_s *itnim, 298 + struct bfa_itnim_ioprofile_s *ioprofile); 299 + 279 300 #define bfa_itnim_get_reqq(__ioim) (((struct bfa_ioim_s *)__ioim)->itnim->reqq) 280 301 281 302 /* 282 - * BFA completion callback for bfa_itnim_online(). 283 - * 284 - * @param[in] itnim FCS or driver itnim instance 285 - * 286 - * return None 303 + * BFA completion callback for bfa_itnim_online(). 287 304 */ 288 305 void bfa_cb_itnim_online(void *itnim); 289 306 290 307 /* 291 - * BFA completion callback for bfa_itnim_offline(). 292 - * 293 - * @param[in] itnim FCS or driver itnim instance 294 - * 295 - * return None 308 + * BFA completion callback for bfa_itnim_offline(). 296 309 */ 297 310 void bfa_cb_itnim_offline(void *itnim); 298 311 void bfa_cb_itnim_tov_begin(void *itnim); 299 312 void bfa_cb_itnim_tov(void *itnim); 300 313 301 314 /* 302 - * BFA notification to FCS/driver for second level error recovery. 303 - * 315 + * BFA notification to FCS/driver for second level error recovery. 304 316 * Atleast one I/O request has timedout and target is unresponsive to 305 317 * repeated abort requests. Second level error recovery should be initiated 306 318 * by starting implicit logout and recovery procedures. 307 - * 308 - * @param[in] itnim FCS or driver itnim instance 309 - * 310 - * return None 311 319 */ 312 320 void bfa_cb_itnim_sler(void *itnim); 313 321 ··· 311 345 bfa_status_t bfa_ioim_abort(struct bfa_ioim_s *ioim); 312 346 void bfa_ioim_delayed_comp(struct bfa_ioim_s *ioim, 313 347 bfa_boolean_t iotov); 314 - 315 - 316 348 /* 317 - * I/O completion notification. 349 + * I/O completion notification. 318 350 * 319 351 * @param[in] dio driver IO structure 320 352 * @param[in] io_status IO completion status ··· 323 359 * 324 360 * @return None 325 361 */ 326 - void bfa_cb_ioim_done(void *bfad, struct bfad_ioim_s *dio, 327 - enum bfi_ioim_status io_status, 328 - u8 scsi_status, int sns_len, 329 - u8 *sns_info, s32 residue); 362 + void bfa_cb_ioim_done(void *bfad, struct bfad_ioim_s *dio, 363 + enum bfi_ioim_status io_status, 364 + u8 scsi_status, int sns_len, 365 + u8 *sns_info, s32 residue); 330 366 331 367 /* 332 - * I/O good completion notification. 333 - * 334 - * @param[in] dio driver IO structure 335 - * 336 - * @return None 368 + * I/O good completion notification. 337 369 */ 338 - void bfa_cb_ioim_good_comp(void *bfad, struct bfad_ioim_s *dio); 370 + void bfa_cb_ioim_good_comp(void *bfad, struct bfad_ioim_s *dio); 339 371 340 372 /* 341 - * I/O abort completion notification 342 - * 343 - * @param[in] dio driver IO that was aborted 344 - * 345 - * @return None 373 + * I/O abort completion notification 346 374 */ 347 - void bfa_cb_ioim_abort(void *bfad, struct bfad_ioim_s *dio); 375 + void bfa_cb_ioim_abort(void *bfad, struct bfad_ioim_s *dio); 348 376 349 377 /* 350 378 * bfa tskim API functions 351 379 */ 352 - struct bfa_tskim_s *bfa_tskim_alloc(struct bfa_s *bfa, 353 - struct bfad_tskim_s *dtsk); 354 - void bfa_tskim_free(struct bfa_tskim_s *tskim); 355 - void bfa_tskim_start(struct bfa_tskim_s *tskim, 356 - struct bfa_itnim_s *itnim, struct scsi_lun lun, 357 - enum fcp_tm_cmnd tm, u8 t_secs); 358 - void bfa_cb_tskim_done(void *bfad, struct bfad_tskim_s *dtsk, 359 - enum bfi_tskim_status tsk_status); 380 + struct bfa_tskim_s *bfa_tskim_alloc(struct bfa_s *bfa, 381 + struct bfad_tskim_s *dtsk); 382 + void bfa_tskim_free(struct bfa_tskim_s *tskim); 383 + void bfa_tskim_start(struct bfa_tskim_s *tskim, 384 + struct bfa_itnim_s *itnim, struct scsi_lun lun, 385 + enum fcp_tm_cmnd tm, u8 t_secs); 386 + void bfa_cb_tskim_done(void *bfad, struct bfad_tskim_s *dtsk, 387 + enum bfi_tskim_status tsk_status); 360 388 361 389 #endif /* __BFA_FCPIM_H__ */
+2 -1
drivers/scsi/bfa/bfa_fcs_lport.c
··· 883 883 884 884 if (port->fabric) { 885 885 port_attr->port_type = port->fabric->oper_type; 886 - port_attr->loopback = bfa_sm_cmp_state(port->fabric, bfa_fcs_fabric_sm_loopback); 886 + port_attr->loopback = bfa_sm_cmp_state(port->fabric, 887 + bfa_fcs_fabric_sm_loopback); 887 888 port_attr->authfail = 888 889 bfa_sm_cmp_state(port->fabric, 889 890 bfa_fcs_fabric_sm_auth_failed);
+2 -1
drivers/scsi/bfa/bfa_fcs_rport.c
··· 1974 1974 rport->itnim = bfa_fcs_itnim_create(rport); 1975 1975 if (!rport->itnim) { 1976 1976 bfa_trc(fcs, rpid); 1977 - bfa_sm_send_event(rport->bfa_rport, BFA_RPORT_SM_DELETE); 1977 + bfa_sm_send_event(rport->bfa_rport, 1978 + BFA_RPORT_SM_DELETE); 1978 1979 kfree(rport_drv); 1979 1980 return NULL; 1980 1981 }
-16
drivers/scsi/bfa/bfa_ioc.c
··· 586 586 } 587 587 } 588 588 589 - 590 - 591 589 /* 592 590 * IOCPF State Machine 593 591 */ 594 - 595 592 596 593 /* 597 594 * Reset entry actions -- initialize state machine ··· 751 754 } 752 755 } 753 756 754 - 755 757 static void 756 758 bfa_iocpf_sm_hwinit_entry(struct bfa_iocpf_s *iocpf) 757 759 { ··· 806 810 bfa_sm_fault(ioc, event); 807 811 } 808 812 } 809 - 810 813 811 814 static void 812 815 bfa_iocpf_sm_enabling_entry(struct bfa_iocpf_s *iocpf) ··· 868 873 } 869 874 } 870 875 871 - 872 - 873 876 static void 874 877 bfa_iocpf_sm_ready_entry(struct bfa_iocpf_s *iocpf) 875 878 { ··· 907 914 bfa_sm_fault(ioc, event); 908 915 } 909 916 } 910 - 911 917 912 918 static void 913 919 bfa_iocpf_sm_disabling_entry(struct bfa_iocpf_s *iocpf) ··· 982 990 } 983 991 } 984 992 985 - 986 993 static void 987 994 bfa_iocpf_sm_initfail_entry(struct bfa_iocpf_s *iocpf) 988 995 { ··· 1018 1027 bfa_sm_fault(ioc, event); 1019 1028 } 1020 1029 } 1021 - 1022 1030 1023 1031 static void 1024 1032 bfa_iocpf_sm_fail_entry(struct bfa_iocpf_s *iocpf) ··· 1068 1078 } 1069 1079 } 1070 1080 1071 - 1072 - 1073 1081 /* 1074 1082 * BFA IOC private functions 1075 1083 */ ··· 1110 1122 bfa_assert(cnt < BFA_SEM_SPINCNT); 1111 1123 return BFA_FALSE; 1112 1124 } 1113 - 1114 1125 1115 1126 static void 1116 1127 bfa_ioc_hw_sem_get(struct bfa_ioc_s *ioc) ··· 1291 1304 if (r32) 1292 1305 writel(1, ioc->ioc_regs.lpu_mbox_cmd); 1293 1306 } 1294 - 1295 1307 1296 1308 static void 1297 1309 bfa_ioc_hwinit(struct bfa_ioc_s *ioc, bfa_boolean_t force) ··· 1724 1738 return BFA_STATUS_OK; 1725 1739 } 1726 1740 1727 - 1728 1741 static void 1729 1742 bfa_ioc_pf_fwmismatch(struct bfa_ioc_s *ioc) 1730 1743 { ··· 1736 1751 "Running firmware version is incompatible " 1737 1752 "with the driver version\n"); 1738 1753 } 1739 - 1740 1754 1741 1755 bfa_status_t 1742 1756 bfa_ioc_pll_init(struct bfa_ioc_s *ioc)
+1 -37
drivers/scsi/bfa/bfa_svc.c
··· 52 52 ((bfa_fcport_is_disabled(bfa) == BFA_TRUE) || \ 53 53 (bfa_ioc_is_disabled(&bfa->ioc) == BFA_TRUE)) 54 54 55 - 56 55 /* 57 56 * BFA port state machine events 58 57 */ ··· 916 917 } 917 918 918 919 /* 919 - * hal_fcxp_api BFA FCXP API 920 - */ 921 - 922 - /* 923 920 * Allocate an FCXP instance to send a response or to send a request 924 921 * that has a response. Request/response buffers are allocated by caller. 925 922 * ··· 1009 1014 } 1010 1015 1011 1016 /* 1012 - * Free the BFA FCXP 1017 + * Free the BFA FCXP 1013 1018 * 1014 1019 * @param[in] fcxp BFA fcxp pointer 1015 1020 * ··· 1155 1160 1156 1161 fcxp->send_cbfn = bfa_fcxp_null_comp; 1157 1162 } 1158 - 1159 - 1160 - 1161 - /* 1162 - * hal_fcxp_public BFA FCXP public functions 1163 - */ 1164 1163 1165 1164 void 1166 1165 bfa_fcxp_isr(struct bfa_s *bfa, struct bfi_msg_s *msg) ··· 2625 2636 } 2626 2637 } 2627 2638 2628 - 2629 - 2630 - /* 2631 - * hal_port_private 2632 - */ 2633 - 2634 2639 static void 2635 2640 __bfa_cb_fcport_event(void *cbarg, bfa_boolean_t complete) 2636 2641 { ··· 3212 3229 } 3213 3230 } 3214 3231 3215 - 3216 - 3217 - /* 3218 - * hal_port_public 3219 - */ 3220 - 3221 3232 /* 3222 3233 * Called to initialize port attributes 3223 3234 */ ··· 3312 3335 break; 3313 3336 } 3314 3337 } 3315 - 3316 - 3317 - 3318 - /* 3319 - * hal_port_api 3320 - */ 3321 3338 3322 3339 /* 3323 3340 * Registered callback for port events. ··· 3644 3673 return fcport->cfg.ratelimit ? BFA_TRUE : BFA_FALSE; 3645 3674 3646 3675 } 3647 - 3648 3676 3649 3677 /* 3650 3678 * Get default minimum ratelim speed ··· 4542 4572 bfa_sgpg_iocdisable(struct bfa_s *bfa) 4543 4573 { 4544 4574 } 4545 - 4546 - 4547 - 4548 - /* 4549 - * hal_sgpg_public BFA SGPG public functions 4550 - */ 4551 4575 4552 4576 bfa_status_t 4553 4577 bfa_sgpg_malloc(struct bfa_s *bfa, struct list_head *sgpg_q, int nsgpgs)