Lines Matching refs:orb
260 void (*callback)(struct sbp2_orb * orb, struct sbp2_status * status);
400 struct sbp2_orb *orb = container_of(kref, struct sbp2_orb, kref); in free_orb() local
402 kfree(orb); in free_orb()
411 struct sbp2_orb *orb; in sbp2_status_write() local
436 list_for_each_entry(orb, &lu->orb_list, link) { in sbp2_status_write()
438 STATUS_GET_ORB_LOW(status) == orb->request_bus) { in sbp2_status_write()
439 orb->rcode = RCODE_COMPLETE; in sbp2_status_write()
440 list_del(&orb->link); in sbp2_status_write()
446 if (&orb->link != &lu->orb_list) { in sbp2_status_write()
447 orb->callback(orb, &status); in sbp2_status_write()
448 kref_put(&orb->kref, free_orb); /* orb callback reference */ in sbp2_status_write()
459 struct sbp2_orb *orb = data; in complete_transaction() local
471 spin_lock_irqsave(&orb->lu->tgt->lock, flags); in complete_transaction()
473 if (orb->rcode == -1) in complete_transaction()
474 orb->rcode = rcode; in complete_transaction()
475 if (orb->rcode != RCODE_COMPLETE) { in complete_transaction()
476 list_del(&orb->link); in complete_transaction()
477 spin_unlock_irqrestore(&orb->lu->tgt->lock, flags); in complete_transaction()
479 orb->callback(orb, NULL); in complete_transaction()
480 kref_put(&orb->kref, free_orb); /* orb callback reference */ in complete_transaction()
482 spin_unlock_irqrestore(&orb->lu->tgt->lock, flags); in complete_transaction()
485 kref_put(&orb->kref, free_orb); /* transaction callback reference */ in complete_transaction()
488 static void sbp2_send_orb(struct sbp2_orb *orb, struct sbp2_logical_unit *lu, in sbp2_send_orb() argument
496 orb_pointer.low = cpu_to_be32(orb->request_bus); in sbp2_send_orb()
498 orb->lu = lu; in sbp2_send_orb()
500 list_add_tail(&orb->link, &lu->orb_list); in sbp2_send_orb()
503 kref_get(&orb->kref); /* transaction callback reference */ in sbp2_send_orb()
504 kref_get(&orb->kref); /* orb callback reference */ in sbp2_send_orb()
506 fw_send_request(device->card, &orb->t, TCODE_WRITE_BLOCK_REQUEST, in sbp2_send_orb()
508 &orb_pointer, 8, complete_transaction, orb); in sbp2_send_orb()
514 struct sbp2_orb *orb, *next; in sbp2_cancel_orbs() local
523 list_for_each_entry_safe(orb, next, &list, link) { in sbp2_cancel_orbs()
525 if (fw_cancel_transaction(device->card, &orb->t) == 0) in sbp2_cancel_orbs()
528 orb->rcode = RCODE_CANCELLED; in sbp2_cancel_orbs()
529 orb->callback(orb, NULL); in sbp2_cancel_orbs()
530 kref_put(&orb->kref, free_orb); /* orb callback reference */ in sbp2_cancel_orbs()
539 struct sbp2_management_orb *orb = in complete_management_orb() local
543 memcpy(&orb->status, status, sizeof(*status)); in complete_management_orb()
544 complete(&orb->done); in complete_management_orb()
552 struct sbp2_management_orb *orb; in sbp2_send_management_orb() local
559 orb = kzalloc(sizeof(*orb), GFP_NOIO); in sbp2_send_management_orb()
560 if (orb == NULL) in sbp2_send_management_orb()
563 kref_init(&orb->base.kref); in sbp2_send_management_orb()
564 orb->response_bus = in sbp2_send_management_orb()
565 dma_map_single(device->card->device, &orb->response, in sbp2_send_management_orb()
566 sizeof(orb->response), DMA_FROM_DEVICE); in sbp2_send_management_orb()
567 if (dma_mapping_error(device->card->device, orb->response_bus)) in sbp2_send_management_orb()
570 orb->request.response.high = 0; in sbp2_send_management_orb()
571 orb->request.response.low = cpu_to_be32(orb->response_bus); in sbp2_send_management_orb()
573 orb->request.misc = cpu_to_be32( in sbp2_send_management_orb()
577 orb->request.length = cpu_to_be32( in sbp2_send_management_orb()
578 MANAGEMENT_ORB_RESPONSE_LENGTH(sizeof(orb->response))); in sbp2_send_management_orb()
580 orb->request.status_fifo.high = in sbp2_send_management_orb()
582 orb->request.status_fifo.low = in sbp2_send_management_orb()
587 orb->request.misc |= cpu_to_be32( in sbp2_send_management_orb()
595 init_completion(&orb->done); in sbp2_send_management_orb()
596 orb->base.callback = complete_management_orb; in sbp2_send_management_orb()
598 orb->base.request_bus = in sbp2_send_management_orb()
599 dma_map_single(device->card->device, &orb->request, in sbp2_send_management_orb()
600 sizeof(orb->request), DMA_TO_DEVICE); in sbp2_send_management_orb()
601 if (dma_mapping_error(device->card->device, orb->base.request_bus)) in sbp2_send_management_orb()
604 sbp2_send_orb(&orb->base, lu, node_id, generation, in sbp2_send_management_orb()
607 wait_for_completion_timeout(&orb->done, msecs_to_jiffies(timeout)); in sbp2_send_management_orb()
612 orb->base.rcode); in sbp2_send_management_orb()
616 if (orb->base.rcode != RCODE_COMPLETE) { in sbp2_send_management_orb()
618 orb->base.rcode); in sbp2_send_management_orb()
622 if (STATUS_GET_RESPONSE(orb->status) != 0 || in sbp2_send_management_orb()
623 STATUS_GET_SBP_STATUS(orb->status) != 0) { in sbp2_send_management_orb()
625 STATUS_GET_RESPONSE(orb->status), in sbp2_send_management_orb()
626 STATUS_GET_SBP_STATUS(orb->status)); in sbp2_send_management_orb()
632 dma_unmap_single(device->card->device, orb->base.request_bus, in sbp2_send_management_orb()
633 sizeof(orb->request), DMA_TO_DEVICE); in sbp2_send_management_orb()
635 dma_unmap_single(device->card->device, orb->response_bus, in sbp2_send_management_orb()
636 sizeof(orb->response), DMA_FROM_DEVICE); in sbp2_send_management_orb()
639 memcpy(response, orb->response, sizeof(orb->response)); in sbp2_send_management_orb()
640 kref_put(&orb->base.kref, free_orb); in sbp2_send_management_orb()
1279 struct sbp2_command_orb *orb) in sbp2_unmap_scatterlist() argument
1281 scsi_dma_unmap(orb->cmd); in sbp2_unmap_scatterlist()
1283 if (orb->request.misc & cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT)) in sbp2_unmap_scatterlist()
1284 dma_unmap_single(card_device, orb->page_table_bus, in sbp2_unmap_scatterlist()
1285 sizeof(orb->page_table), DMA_TO_DEVICE); in sbp2_unmap_scatterlist()
1337 struct sbp2_command_orb *orb = in complete_command_orb() local
1362 orb->cmd->sense_buffer); in complete_command_orb()
1373 dma_unmap_single(device->card->device, orb->base.request_bus, in complete_command_orb()
1374 sizeof(orb->request), DMA_TO_DEVICE); in complete_command_orb()
1375 sbp2_unmap_scatterlist(device->card->device, orb); in complete_command_orb()
1377 orb->cmd->result = result; in complete_command_orb()
1378 scsi_done(orb->cmd); in complete_command_orb()
1381 static int sbp2_map_scatterlist(struct sbp2_command_orb *orb, in sbp2_map_scatterlist() argument
1384 struct scatterlist *sg = scsi_sglist(orb->cmd); in sbp2_map_scatterlist()
1387 n = scsi_dma_map(orb->cmd); in sbp2_map_scatterlist()
1399 orb->request.data_descriptor.high = in sbp2_map_scatterlist()
1401 orb->request.data_descriptor.low = in sbp2_map_scatterlist()
1403 orb->request.misc |= in sbp2_map_scatterlist()
1409 orb->page_table[i].high = cpu_to_be32(sg_dma_len(sg) << 16); in sbp2_map_scatterlist()
1410 orb->page_table[i].low = cpu_to_be32(sg_dma_address(sg)); in sbp2_map_scatterlist()
1413 orb->page_table_bus = in sbp2_map_scatterlist()
1414 dma_map_single(device->card->device, orb->page_table, in sbp2_map_scatterlist()
1415 sizeof(orb->page_table), DMA_TO_DEVICE); in sbp2_map_scatterlist()
1416 if (dma_mapping_error(device->card->device, orb->page_table_bus)) in sbp2_map_scatterlist()
1426 orb->request.data_descriptor.high = cpu_to_be32(lu->tgt->address_high); in sbp2_map_scatterlist()
1427 orb->request.data_descriptor.low = cpu_to_be32(orb->page_table_bus); in sbp2_map_scatterlist()
1428 orb->request.misc |= cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT | in sbp2_map_scatterlist()
1434 scsi_dma_unmap(orb->cmd); in sbp2_map_scatterlist()
1446 struct sbp2_command_orb *orb; in sbp2_scsi_queuecommand() local
1449 orb = kzalloc(sizeof(*orb), GFP_ATOMIC); in sbp2_scsi_queuecommand()
1450 if (orb == NULL) in sbp2_scsi_queuecommand()
1454 orb->base.rcode = -1; in sbp2_scsi_queuecommand()
1455 kref_init(&orb->base.kref); in sbp2_scsi_queuecommand()
1456 orb->cmd = cmd; in sbp2_scsi_queuecommand()
1457 orb->request.next.high = cpu_to_be32(SBP2_ORB_NULL); in sbp2_scsi_queuecommand()
1458 orb->request.misc = cpu_to_be32( in sbp2_scsi_queuecommand()
1464 orb->request.misc |= cpu_to_be32(COMMAND_ORB_DIRECTION); in sbp2_scsi_queuecommand()
1469 if (scsi_sg_count(cmd) && sbp2_map_scatterlist(orb, device, lu) < 0) in sbp2_scsi_queuecommand()
1472 memcpy(orb->request.command_block, cmd->cmnd, cmd->cmd_len); in sbp2_scsi_queuecommand()
1474 orb->base.callback = complete_command_orb; in sbp2_scsi_queuecommand()
1475 orb->base.request_bus = in sbp2_scsi_queuecommand()
1476 dma_map_single(device->card->device, &orb->request, in sbp2_scsi_queuecommand()
1477 sizeof(orb->request), DMA_TO_DEVICE); in sbp2_scsi_queuecommand()
1478 if (dma_mapping_error(device->card->device, orb->base.request_bus)) { in sbp2_scsi_queuecommand()
1479 sbp2_unmap_scatterlist(device->card->device, orb); in sbp2_scsi_queuecommand()
1483 sbp2_send_orb(&orb->base, lu, lu->tgt->node_id, generation, in sbp2_scsi_queuecommand()
1487 kref_put(&orb->base.kref, free_orb); in sbp2_scsi_queuecommand()