ieee1394: sbp2: convert from PCI DMA to generic DMA
API conversion without change in functionality Signed-off-by: Stefan Richter <stefanr@s5r6.in-berlin.de>
This commit is contained in:
@@ -120,7 +120,7 @@ comment "SBP-2 support (for storage devices) requires SCSI"
|
|||||||
|
|
||||||
config IEEE1394_SBP2
|
config IEEE1394_SBP2
|
||||||
tristate "SBP-2 support (Harddisks etc.)"
|
tristate "SBP-2 support (Harddisks etc.)"
|
||||||
depends on IEEE1394 && SCSI && (PCI || BROKEN)
|
depends on IEEE1394 && SCSI
|
||||||
help
|
help
|
||||||
This option enables you to use SBP-2 devices connected to an IEEE
|
This option enables you to use SBP-2 devices connected to an IEEE
|
||||||
1394 bus. SBP-2 devices include storage devices like harddisks and
|
1394 bus. SBP-2 devices include storage devices like harddisks and
|
||||||
|
@@ -48,7 +48,6 @@
|
|||||||
* - make the parameter serialize_io configurable per device
|
* - make the parameter serialize_io configurable per device
|
||||||
* - move all requests to fetch agent registers into non-atomic context,
|
* - move all requests to fetch agent registers into non-atomic context,
|
||||||
* replace all usages of sbp2util_node_write_no_wait by true transactions
|
* replace all usages of sbp2util_node_write_no_wait by true transactions
|
||||||
* - convert to generic DMA mapping API to eliminate dependency on PCI
|
|
||||||
* Grep for inline FIXME comments below.
|
* Grep for inline FIXME comments below.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
@@ -63,7 +62,6 @@
|
|||||||
#include <linux/list.h>
|
#include <linux/list.h>
|
||||||
#include <linux/module.h>
|
#include <linux/module.h>
|
||||||
#include <linux/moduleparam.h>
|
#include <linux/moduleparam.h>
|
||||||
#include <linux/pci.h>
|
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/spinlock.h>
|
#include <linux/spinlock.h>
|
||||||
#include <linux/stat.h>
|
#include <linux/stat.h>
|
||||||
@@ -491,14 +489,14 @@ static int sbp2util_create_command_orb_pool(struct sbp2_lu *lu)
|
|||||||
spin_unlock_irqrestore(&lu->cmd_orb_lock, flags);
|
spin_unlock_irqrestore(&lu->cmd_orb_lock, flags);
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
cmd->command_orb_dma = pci_map_single(hi->host->pdev,
|
cmd->command_orb_dma = dma_map_single(&hi->host->device,
|
||||||
&cmd->command_orb,
|
&cmd->command_orb,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
cmd->sge_dma = pci_map_single(hi->host->pdev,
|
cmd->sge_dma = dma_map_single(&hi->host->device,
|
||||||
&cmd->scatter_gather_element,
|
&cmd->scatter_gather_element,
|
||||||
sizeof(cmd->scatter_gather_element),
|
sizeof(cmd->scatter_gather_element),
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
DMA_BIDIRECTIONAL);
|
||||||
INIT_LIST_HEAD(&cmd->list);
|
INIT_LIST_HEAD(&cmd->list);
|
||||||
list_add_tail(&cmd->list, &lu->cmd_orb_completed);
|
list_add_tail(&cmd->list, &lu->cmd_orb_completed);
|
||||||
}
|
}
|
||||||
@@ -517,12 +515,12 @@ static void sbp2util_remove_command_orb_pool(struct sbp2_lu *lu)
|
|||||||
if (!list_empty(&lu->cmd_orb_completed))
|
if (!list_empty(&lu->cmd_orb_completed))
|
||||||
list_for_each_safe(lh, next, &lu->cmd_orb_completed) {
|
list_for_each_safe(lh, next, &lu->cmd_orb_completed) {
|
||||||
cmd = list_entry(lh, struct sbp2_command_info, list);
|
cmd = list_entry(lh, struct sbp2_command_info, list);
|
||||||
pci_unmap_single(host->pdev, cmd->command_orb_dma,
|
dma_unmap_single(&host->device, cmd->command_orb_dma,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
pci_unmap_single(host->pdev, cmd->sge_dma,
|
dma_unmap_single(&host->device, cmd->sge_dma,
|
||||||
sizeof(cmd->scatter_gather_element),
|
sizeof(cmd->scatter_gather_element),
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
DMA_BIDIRECTIONAL);
|
||||||
kfree(cmd);
|
kfree(cmd);
|
||||||
}
|
}
|
||||||
spin_unlock_irqrestore(&lu->cmd_orb_lock, flags);
|
spin_unlock_irqrestore(&lu->cmd_orb_lock, flags);
|
||||||
@@ -602,17 +600,17 @@ static void sbp2util_mark_command_completed(struct sbp2_lu *lu,
|
|||||||
|
|
||||||
if (cmd->cmd_dma) {
|
if (cmd->cmd_dma) {
|
||||||
if (cmd->dma_type == CMD_DMA_SINGLE)
|
if (cmd->dma_type == CMD_DMA_SINGLE)
|
||||||
pci_unmap_single(host->pdev, cmd->cmd_dma,
|
dma_unmap_single(&host->device, cmd->cmd_dma,
|
||||||
cmd->dma_size, cmd->dma_dir);
|
cmd->dma_size, cmd->dma_dir);
|
||||||
else if (cmd->dma_type == CMD_DMA_PAGE)
|
else if (cmd->dma_type == CMD_DMA_PAGE)
|
||||||
pci_unmap_page(host->pdev, cmd->cmd_dma,
|
dma_unmap_page(&host->device, cmd->cmd_dma,
|
||||||
cmd->dma_size, cmd->dma_dir);
|
cmd->dma_size, cmd->dma_dir);
|
||||||
/* XXX: Check for CMD_DMA_NONE bug */
|
/* XXX: Check for CMD_DMA_NONE bug */
|
||||||
cmd->dma_type = CMD_DMA_NONE;
|
cmd->dma_type = CMD_DMA_NONE;
|
||||||
cmd->cmd_dma = 0;
|
cmd->cmd_dma = 0;
|
||||||
}
|
}
|
||||||
if (cmd->sge_buffer) {
|
if (cmd->sge_buffer) {
|
||||||
pci_unmap_sg(host->pdev, cmd->sge_buffer,
|
dma_unmap_sg(&host->device, cmd->sge_buffer,
|
||||||
cmd->dma_size, cmd->dma_dir);
|
cmd->dma_size, cmd->dma_dir);
|
||||||
cmd->sge_buffer = NULL;
|
cmd->sge_buffer = NULL;
|
||||||
}
|
}
|
||||||
@@ -837,39 +835,39 @@ static int sbp2_start_device(struct sbp2_lu *lu)
|
|||||||
struct sbp2_fwhost_info *hi = lu->hi;
|
struct sbp2_fwhost_info *hi = lu->hi;
|
||||||
int error;
|
int error;
|
||||||
|
|
||||||
lu->login_response = pci_alloc_consistent(hi->host->pdev,
|
lu->login_response = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_login_response),
|
sizeof(struct sbp2_login_response),
|
||||||
&lu->login_response_dma);
|
&lu->login_response_dma, GFP_KERNEL);
|
||||||
if (!lu->login_response)
|
if (!lu->login_response)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
lu->query_logins_orb = pci_alloc_consistent(hi->host->pdev,
|
lu->query_logins_orb = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_query_logins_orb),
|
sizeof(struct sbp2_query_logins_orb),
|
||||||
&lu->query_logins_orb_dma);
|
&lu->query_logins_orb_dma, GFP_KERNEL);
|
||||||
if (!lu->query_logins_orb)
|
if (!lu->query_logins_orb)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
lu->query_logins_response = pci_alloc_consistent(hi->host->pdev,
|
lu->query_logins_response = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_query_logins_response),
|
sizeof(struct sbp2_query_logins_response),
|
||||||
&lu->query_logins_response_dma);
|
&lu->query_logins_response_dma, GFP_KERNEL);
|
||||||
if (!lu->query_logins_response)
|
if (!lu->query_logins_response)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
lu->reconnect_orb = pci_alloc_consistent(hi->host->pdev,
|
lu->reconnect_orb = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_reconnect_orb),
|
sizeof(struct sbp2_reconnect_orb),
|
||||||
&lu->reconnect_orb_dma);
|
&lu->reconnect_orb_dma, GFP_KERNEL);
|
||||||
if (!lu->reconnect_orb)
|
if (!lu->reconnect_orb)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
lu->logout_orb = pci_alloc_consistent(hi->host->pdev,
|
lu->logout_orb = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_logout_orb),
|
sizeof(struct sbp2_logout_orb),
|
||||||
&lu->logout_orb_dma);
|
&lu->logout_orb_dma, GFP_KERNEL);
|
||||||
if (!lu->logout_orb)
|
if (!lu->logout_orb)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
lu->login_orb = pci_alloc_consistent(hi->host->pdev,
|
lu->login_orb = dma_alloc_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_login_orb),
|
sizeof(struct sbp2_login_orb),
|
||||||
&lu->login_orb_dma);
|
&lu->login_orb_dma, GFP_KERNEL);
|
||||||
if (!lu->login_orb)
|
if (!lu->login_orb)
|
||||||
goto alloc_fail;
|
goto alloc_fail;
|
||||||
|
|
||||||
@@ -930,32 +928,32 @@ static void sbp2_remove_device(struct sbp2_lu *lu)
|
|||||||
list_del(&lu->lu_list);
|
list_del(&lu->lu_list);
|
||||||
|
|
||||||
if (lu->login_response)
|
if (lu->login_response)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_login_response),
|
sizeof(struct sbp2_login_response),
|
||||||
lu->login_response,
|
lu->login_response,
|
||||||
lu->login_response_dma);
|
lu->login_response_dma);
|
||||||
if (lu->login_orb)
|
if (lu->login_orb)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_login_orb),
|
sizeof(struct sbp2_login_orb),
|
||||||
lu->login_orb,
|
lu->login_orb,
|
||||||
lu->login_orb_dma);
|
lu->login_orb_dma);
|
||||||
if (lu->reconnect_orb)
|
if (lu->reconnect_orb)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_reconnect_orb),
|
sizeof(struct sbp2_reconnect_orb),
|
||||||
lu->reconnect_orb,
|
lu->reconnect_orb,
|
||||||
lu->reconnect_orb_dma);
|
lu->reconnect_orb_dma);
|
||||||
if (lu->logout_orb)
|
if (lu->logout_orb)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_logout_orb),
|
sizeof(struct sbp2_logout_orb),
|
||||||
lu->logout_orb,
|
lu->logout_orb,
|
||||||
lu->logout_orb_dma);
|
lu->logout_orb_dma);
|
||||||
if (lu->query_logins_orb)
|
if (lu->query_logins_orb)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_query_logins_orb),
|
sizeof(struct sbp2_query_logins_orb),
|
||||||
lu->query_logins_orb,
|
lu->query_logins_orb,
|
||||||
lu->query_logins_orb_dma);
|
lu->query_logins_orb_dma);
|
||||||
if (lu->query_logins_response)
|
if (lu->query_logins_response)
|
||||||
pci_free_consistent(hi->host->pdev,
|
dma_free_coherent(&hi->host->device,
|
||||||
sizeof(struct sbp2_query_logins_response),
|
sizeof(struct sbp2_query_logins_response),
|
||||||
lu->query_logins_response,
|
lu->query_logins_response,
|
||||||
lu->query_logins_response_dma);
|
lu->query_logins_response_dma);
|
||||||
@@ -1446,7 +1444,7 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
|
|||||||
|
|
||||||
cmd->dma_size = sgpnt[0].length;
|
cmd->dma_size = sgpnt[0].length;
|
||||||
cmd->dma_type = CMD_DMA_PAGE;
|
cmd->dma_type = CMD_DMA_PAGE;
|
||||||
cmd->cmd_dma = pci_map_page(hi->host->pdev,
|
cmd->cmd_dma = dma_map_page(&hi->host->device,
|
||||||
sgpnt[0].page, sgpnt[0].offset,
|
sgpnt[0].page, sgpnt[0].offset,
|
||||||
cmd->dma_size, cmd->dma_dir);
|
cmd->dma_size, cmd->dma_dir);
|
||||||
|
|
||||||
@@ -1458,7 +1456,7 @@ static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
|
|||||||
&cmd->scatter_gather_element[0];
|
&cmd->scatter_gather_element[0];
|
||||||
u32 sg_count, sg_len;
|
u32 sg_count, sg_len;
|
||||||
dma_addr_t sg_addr;
|
dma_addr_t sg_addr;
|
||||||
int i, count = pci_map_sg(hi->host->pdev, sgpnt, scsi_use_sg,
|
int i, count = dma_map_sg(&hi->host->device, sgpnt, scsi_use_sg,
|
||||||
dma_dir);
|
dma_dir);
|
||||||
|
|
||||||
cmd->dma_size = scsi_use_sg;
|
cmd->dma_size = scsi_use_sg;
|
||||||
@@ -1509,7 +1507,7 @@ static void sbp2_prep_command_orb_no_sg(struct sbp2_command_orb *orb,
|
|||||||
cmd->dma_dir = dma_dir;
|
cmd->dma_dir = dma_dir;
|
||||||
cmd->dma_size = scsi_request_bufflen;
|
cmd->dma_size = scsi_request_bufflen;
|
||||||
cmd->dma_type = CMD_DMA_SINGLE;
|
cmd->dma_type = CMD_DMA_SINGLE;
|
||||||
cmd->cmd_dma = pci_map_single(hi->host->pdev, scsi_request_buffer,
|
cmd->cmd_dma = dma_map_single(&hi->host->device, scsi_request_buffer,
|
||||||
cmd->dma_size, cmd->dma_dir);
|
cmd->dma_size, cmd->dma_dir);
|
||||||
orb->data_descriptor_hi = ORB_SET_NODE_ID(hi->host->node_id);
|
orb->data_descriptor_hi = ORB_SET_NODE_ID(hi->host->node_id);
|
||||||
orb->misc |= ORB_SET_DIRECTION(orb_direction);
|
orb->misc |= ORB_SET_DIRECTION(orb_direction);
|
||||||
@@ -1627,12 +1625,12 @@ static void sbp2_link_orb_command(struct sbp2_lu *lu,
|
|||||||
size_t length;
|
size_t length;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
|
|
||||||
pci_dma_sync_single_for_device(hi->host->pdev, cmd->command_orb_dma,
|
dma_sync_single_for_device(&hi->host->device, cmd->command_orb_dma,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
pci_dma_sync_single_for_device(hi->host->pdev, cmd->sge_dma,
|
dma_sync_single_for_device(&hi->host->device, cmd->sge_dma,
|
||||||
sizeof(cmd->scatter_gather_element),
|
sizeof(cmd->scatter_gather_element),
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
DMA_BIDIRECTIONAL);
|
||||||
|
|
||||||
/* check to see if there are any previous orbs to use */
|
/* check to see if there are any previous orbs to use */
|
||||||
spin_lock_irqsave(&lu->cmd_orb_lock, flags);
|
spin_lock_irqsave(&lu->cmd_orb_lock, flags);
|
||||||
@@ -1656,16 +1654,16 @@ static void sbp2_link_orb_command(struct sbp2_lu *lu,
|
|||||||
* The target's fetch agent may or may not have read this
|
* The target's fetch agent may or may not have read this
|
||||||
* previous ORB yet.
|
* previous ORB yet.
|
||||||
*/
|
*/
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev, last_orb_dma,
|
dma_sync_single_for_cpu(&hi->host->device, last_orb_dma,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
last_orb->next_ORB_lo = cpu_to_be32(cmd->command_orb_dma);
|
last_orb->next_ORB_lo = cpu_to_be32(cmd->command_orb_dma);
|
||||||
wmb();
|
wmb();
|
||||||
/* Tells hardware that this pointer is valid */
|
/* Tells hardware that this pointer is valid */
|
||||||
last_orb->next_ORB_hi = 0;
|
last_orb->next_ORB_hi = 0;
|
||||||
pci_dma_sync_single_for_device(hi->host->pdev, last_orb_dma,
|
dma_sync_single_for_device(&hi->host->device, last_orb_dma,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
addr += SBP2_DOORBELL_OFFSET;
|
addr += SBP2_DOORBELL_OFFSET;
|
||||||
data[0] = 0;
|
data[0] = 0;
|
||||||
length = 4;
|
length = 4;
|
||||||
@@ -1792,14 +1790,12 @@ static int sbp2_handle_status_write(struct hpsb_host *host, int nodeid,
|
|||||||
else
|
else
|
||||||
cmd = sbp2util_find_command_for_orb(lu, sb->ORB_offset_lo);
|
cmd = sbp2util_find_command_for_orb(lu, sb->ORB_offset_lo);
|
||||||
if (cmd) {
|
if (cmd) {
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev,
|
dma_sync_single_for_cpu(&hi->host->device, cmd->command_orb_dma,
|
||||||
cmd->command_orb_dma,
|
sizeof(struct sbp2_command_orb),
|
||||||
sizeof(struct sbp2_command_orb),
|
DMA_TO_DEVICE);
|
||||||
PCI_DMA_TODEVICE);
|
dma_sync_single_for_cpu(&hi->host->device, cmd->sge_dma,
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev,
|
sizeof(cmd->scatter_gather_element),
|
||||||
cmd->sge_dma,
|
DMA_BIDIRECTIONAL);
|
||||||
sizeof(cmd->scatter_gather_element),
|
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
|
||||||
/* Grab SCSI command pointers and check status. */
|
/* Grab SCSI command pointers and check status. */
|
||||||
/*
|
/*
|
||||||
* FIXME: If the src field in the status is 1, the ORB DMA must
|
* FIXME: If the src field in the status is 1, the ORB DMA must
|
||||||
@@ -1935,13 +1931,12 @@ static void sbp2scsi_complete_all_commands(struct sbp2_lu *lu, u32 status)
|
|||||||
while (!list_empty(&lu->cmd_orb_inuse)) {
|
while (!list_empty(&lu->cmd_orb_inuse)) {
|
||||||
lh = lu->cmd_orb_inuse.next;
|
lh = lu->cmd_orb_inuse.next;
|
||||||
cmd = list_entry(lh, struct sbp2_command_info, list);
|
cmd = list_entry(lh, struct sbp2_command_info, list);
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev,
|
dma_sync_single_for_cpu(&hi->host->device, cmd->command_orb_dma,
|
||||||
cmd->command_orb_dma,
|
sizeof(struct sbp2_command_orb),
|
||||||
sizeof(struct sbp2_command_orb),
|
DMA_TO_DEVICE);
|
||||||
PCI_DMA_TODEVICE);
|
dma_sync_single_for_cpu(&hi->host->device, cmd->sge_dma,
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev, cmd->sge_dma,
|
sizeof(cmd->scatter_gather_element),
|
||||||
sizeof(cmd->scatter_gather_element),
|
DMA_BIDIRECTIONAL);
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
|
||||||
sbp2util_mark_command_completed(lu, cmd);
|
sbp2util_mark_command_completed(lu, cmd);
|
||||||
if (cmd->Current_SCpnt) {
|
if (cmd->Current_SCpnt) {
|
||||||
cmd->Current_SCpnt->result = status << 16;
|
cmd->Current_SCpnt->result = status << 16;
|
||||||
@@ -2064,14 +2059,13 @@ static int sbp2scsi_abort(struct scsi_cmnd *SCpnt)
|
|||||||
spin_lock_irqsave(&lu->cmd_orb_lock, flags);
|
spin_lock_irqsave(&lu->cmd_orb_lock, flags);
|
||||||
cmd = sbp2util_find_command_for_SCpnt(lu, SCpnt);
|
cmd = sbp2util_find_command_for_SCpnt(lu, SCpnt);
|
||||||
if (cmd) {
|
if (cmd) {
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev,
|
dma_sync_single_for_cpu(&hi->host->device,
|
||||||
cmd->command_orb_dma,
|
cmd->command_orb_dma,
|
||||||
sizeof(struct sbp2_command_orb),
|
sizeof(struct sbp2_command_orb),
|
||||||
PCI_DMA_TODEVICE);
|
DMA_TO_DEVICE);
|
||||||
pci_dma_sync_single_for_cpu(hi->host->pdev,
|
dma_sync_single_for_cpu(&hi->host->device, cmd->sge_dma,
|
||||||
cmd->sge_dma,
|
|
||||||
sizeof(cmd->scatter_gather_element),
|
sizeof(cmd->scatter_gather_element),
|
||||||
PCI_DMA_BIDIRECTIONAL);
|
DMA_BIDIRECTIONAL);
|
||||||
sbp2util_mark_command_completed(lu, cmd);
|
sbp2util_mark_command_completed(lu, cmd);
|
||||||
if (cmd->Current_SCpnt) {
|
if (cmd->Current_SCpnt) {
|
||||||
cmd->Current_SCpnt->result = DID_ABORT << 16;
|
cmd->Current_SCpnt->result = DID_ABORT << 16;
|
||||||
|
@@ -263,7 +263,7 @@ struct sbp2_command_info {
|
|||||||
dma_addr_t cmd_dma;
|
dma_addr_t cmd_dma;
|
||||||
enum sbp2_dma_types dma_type;
|
enum sbp2_dma_types dma_type;
|
||||||
unsigned long dma_size;
|
unsigned long dma_size;
|
||||||
int dma_dir;
|
enum dma_data_direction dma_dir;
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Per FireWire host */
|
/* Per FireWire host */
|
||||||
|
Reference in New Issue
Block a user