WorkStruct: make allyesconfig
Fix up for make allyesconfig. Signed-Off-By: David Howells <dhowells@redhat.com>
This commit is contained in:
@@ -1018,9 +1018,10 @@ mptfc_init_host_attr(MPT_ADAPTER *ioc,int portnum)
|
||||
}
|
||||
|
||||
static void
|
||||
mptfc_setup_reset(void *arg)
|
||||
mptfc_setup_reset(struct work_struct *work)
|
||||
{
|
||||
MPT_ADAPTER *ioc = (MPT_ADAPTER *)arg;
|
||||
MPT_ADAPTER *ioc =
|
||||
container_of(work, MPT_ADAPTER, fc_setup_reset_work);
|
||||
u64 pn;
|
||||
struct mptfc_rport_info *ri;
|
||||
|
||||
@@ -1043,9 +1044,10 @@ mptfc_setup_reset(void *arg)
|
||||
}
|
||||
|
||||
static void
|
||||
mptfc_rescan_devices(void *arg)
|
||||
mptfc_rescan_devices(struct work_struct *work)
|
||||
{
|
||||
MPT_ADAPTER *ioc = (MPT_ADAPTER *)arg;
|
||||
MPT_ADAPTER *ioc =
|
||||
container_of(work, MPT_ADAPTER, fc_rescan_work);
|
||||
int ii;
|
||||
u64 pn;
|
||||
struct mptfc_rport_info *ri;
|
||||
@@ -1154,8 +1156,8 @@ mptfc_probe(struct pci_dev *pdev, const struct pci_device_id *id)
|
||||
}
|
||||
|
||||
spin_lock_init(&ioc->fc_rescan_work_lock);
|
||||
INIT_WORK(&ioc->fc_rescan_work, mptfc_rescan_devices,(void *)ioc);
|
||||
INIT_WORK(&ioc->fc_setup_reset_work, mptfc_setup_reset, (void *)ioc);
|
||||
INIT_WORK(&ioc->fc_rescan_work, mptfc_rescan_devices);
|
||||
INIT_WORK(&ioc->fc_setup_reset_work, mptfc_setup_reset);
|
||||
|
||||
spin_lock_irqsave(&ioc->FreeQlock, flags);
|
||||
|
||||
|
@@ -111,7 +111,8 @@ struct mpt_lan_priv {
|
||||
u32 total_received;
|
||||
struct net_device_stats stats; /* Per device statistics */
|
||||
|
||||
struct work_struct post_buckets_task;
|
||||
struct delayed_work post_buckets_task;
|
||||
struct net_device *dev;
|
||||
unsigned long post_buckets_active;
|
||||
};
|
||||
|
||||
@@ -132,7 +133,7 @@ static int lan_reply (MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf,
|
||||
static int mpt_lan_open(struct net_device *dev);
|
||||
static int mpt_lan_reset(struct net_device *dev);
|
||||
static int mpt_lan_close(struct net_device *dev);
|
||||
static void mpt_lan_post_receive_buckets(void *dev_id);
|
||||
static void mpt_lan_post_receive_buckets(struct mpt_lan_priv *priv);
|
||||
static void mpt_lan_wake_post_buckets_task(struct net_device *dev,
|
||||
int priority);
|
||||
static int mpt_lan_receive_post_turbo(struct net_device *dev, u32 tmsg);
|
||||
@@ -345,7 +346,7 @@ mpt_lan_ioc_reset(MPT_ADAPTER *ioc, int reset_phase)
|
||||
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = i;
|
||||
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
||||
} else {
|
||||
mpt_lan_post_receive_buckets(dev);
|
||||
mpt_lan_post_receive_buckets(priv);
|
||||
netif_wake_queue(dev);
|
||||
}
|
||||
|
||||
@@ -441,7 +442,7 @@ mpt_lan_open(struct net_device *dev)
|
||||
|
||||
dlprintk((KERN_INFO MYNAM "/lo: Finished initializing RcvCtl\n"));
|
||||
|
||||
mpt_lan_post_receive_buckets(dev);
|
||||
mpt_lan_post_receive_buckets(priv);
|
||||
printk(KERN_INFO MYNAM ": %s/%s: interface up & active\n",
|
||||
IOC_AND_NETDEV_NAMES_s_s(dev));
|
||||
|
||||
@@ -854,7 +855,7 @@ mpt_lan_wake_post_buckets_task(struct net_device *dev, int priority)
|
||||
|
||||
if (test_and_set_bit(0, &priv->post_buckets_active) == 0) {
|
||||
if (priority) {
|
||||
schedule_work(&priv->post_buckets_task);
|
||||
schedule_delayed_work(&priv->post_buckets_task, 0);
|
||||
} else {
|
||||
schedule_delayed_work(&priv->post_buckets_task, 1);
|
||||
dioprintk((KERN_INFO MYNAM ": post_buckets queued on "
|
||||
@@ -1188,10 +1189,9 @@ mpt_lan_receive_post_reply(struct net_device *dev,
|
||||
/* Simple SGE's only at the moment */
|
||||
|
||||
static void
|
||||
mpt_lan_post_receive_buckets(void *dev_id)
|
||||
mpt_lan_post_receive_buckets(struct mpt_lan_priv *priv)
|
||||
{
|
||||
struct net_device *dev = dev_id;
|
||||
struct mpt_lan_priv *priv = dev->priv;
|
||||
struct net_device *dev = priv->dev;
|
||||
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
||||
MPT_FRAME_HDR *mf;
|
||||
LANReceivePostRequest_t *pRecvReq;
|
||||
@@ -1335,6 +1335,13 @@ out:
|
||||
clear_bit(0, &priv->post_buckets_active);
|
||||
}
|
||||
|
||||
static void
|
||||
mpt_lan_post_receive_buckets_work(struct work_struct *work)
|
||||
{
|
||||
mpt_lan_post_receive_buckets(container_of(work, struct mpt_lan_priv,
|
||||
post_buckets_task.work));
|
||||
}
|
||||
|
||||
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
||||
static struct net_device *
|
||||
mpt_register_lan_device (MPT_ADAPTER *mpt_dev, int pnum)
|
||||
@@ -1350,11 +1357,13 @@ mpt_register_lan_device (MPT_ADAPTER *mpt_dev, int pnum)
|
||||
|
||||
priv = netdev_priv(dev);
|
||||
|
||||
priv->dev = dev;
|
||||
priv->mpt_dev = mpt_dev;
|
||||
priv->pnum = pnum;
|
||||
|
||||
memset(&priv->post_buckets_task, 0, sizeof(struct work_struct));
|
||||
INIT_WORK(&priv->post_buckets_task, mpt_lan_post_receive_buckets, dev);
|
||||
memset(&priv->post_buckets_task, 0, sizeof(priv->post_buckets_task));
|
||||
INIT_DELAYED_WORK(&priv->post_buckets_task,
|
||||
mpt_lan_post_receive_buckets_work);
|
||||
priv->post_buckets_active = 0;
|
||||
|
||||
dlprintk((KERN_INFO MYNAM "@%d: bucketlen = %d\n",
|
||||
|
@@ -2006,9 +2006,10 @@ __mptsas_discovery_work(MPT_ADAPTER *ioc)
|
||||
*(Mutex LOCKED)
|
||||
*/
|
||||
static void
|
||||
mptsas_discovery_work(void * arg)
|
||||
mptsas_discovery_work(struct work_struct *work)
|
||||
{
|
||||
struct mptsas_discovery_event *ev = arg;
|
||||
struct mptsas_discovery_event *ev =
|
||||
container_of(work, struct mptsas_discovery_event, work);
|
||||
MPT_ADAPTER *ioc = ev->ioc;
|
||||
|
||||
mutex_lock(&ioc->sas_discovery_mutex);
|
||||
@@ -2068,9 +2069,9 @@ mptsas_find_phyinfo_by_target(MPT_ADAPTER *ioc, u32 id)
|
||||
* Work queue thread to clear the persitency table
|
||||
*/
|
||||
static void
|
||||
mptsas_persist_clear_table(void * arg)
|
||||
mptsas_persist_clear_table(struct work_struct *work)
|
||||
{
|
||||
MPT_ADAPTER *ioc = (MPT_ADAPTER *)arg;
|
||||
MPT_ADAPTER *ioc = container_of(work, MPT_ADAPTER, sas_persist_task);
|
||||
|
||||
mptbase_sas_persist_operation(ioc, MPI_SAS_OP_CLEAR_NOT_PRESENT);
|
||||
}
|
||||
@@ -2093,9 +2094,10 @@ mptsas_reprobe_target(struct scsi_target *starget, int uld_attach)
|
||||
* Work queue thread to handle SAS hotplug events
|
||||
*/
|
||||
static void
|
||||
mptsas_hotplug_work(void *arg)
|
||||
mptsas_hotplug_work(struct work_struct *work)
|
||||
{
|
||||
struct mptsas_hotplug_event *ev = arg;
|
||||
struct mptsas_hotplug_event *ev =
|
||||
container_of(work, struct mptsas_hotplug_event, work);
|
||||
MPT_ADAPTER *ioc = ev->ioc;
|
||||
struct mptsas_phyinfo *phy_info;
|
||||
struct sas_rphy *rphy;
|
||||
@@ -2341,7 +2343,7 @@ mptsas_send_sas_event(MPT_ADAPTER *ioc,
|
||||
break;
|
||||
}
|
||||
|
||||
INIT_WORK(&ev->work, mptsas_hotplug_work, ev);
|
||||
INIT_WORK(&ev->work, mptsas_hotplug_work);
|
||||
ev->ioc = ioc;
|
||||
ev->handle = le16_to_cpu(sas_event_data->DevHandle);
|
||||
ev->parent_handle =
|
||||
@@ -2366,7 +2368,7 @@ mptsas_send_sas_event(MPT_ADAPTER *ioc,
|
||||
* Persistent table is full.
|
||||
*/
|
||||
INIT_WORK(&ioc->sas_persist_task,
|
||||
mptsas_persist_clear_table, (void *)ioc);
|
||||
mptsas_persist_clear_table);
|
||||
schedule_work(&ioc->sas_persist_task);
|
||||
break;
|
||||
case MPI_EVENT_SAS_DEV_STAT_RC_SMART_DATA:
|
||||
@@ -2395,7 +2397,7 @@ mptsas_send_raid_event(MPT_ADAPTER *ioc,
|
||||
return;
|
||||
}
|
||||
|
||||
INIT_WORK(&ev->work, mptsas_hotplug_work, ev);
|
||||
INIT_WORK(&ev->work, mptsas_hotplug_work);
|
||||
ev->ioc = ioc;
|
||||
ev->id = raid_event_data->VolumeID;
|
||||
ev->event_type = MPTSAS_IGNORE_EVENT;
|
||||
@@ -2474,7 +2476,7 @@ mptsas_send_discovery_event(MPT_ADAPTER *ioc,
|
||||
ev = kzalloc(sizeof(*ev), GFP_ATOMIC);
|
||||
if (!ev)
|
||||
return;
|
||||
INIT_WORK(&ev->work, mptsas_discovery_work, ev);
|
||||
INIT_WORK(&ev->work, mptsas_discovery_work);
|
||||
ev->ioc = ioc;
|
||||
schedule_work(&ev->work);
|
||||
};
|
||||
@@ -2511,8 +2513,7 @@ mptsas_event_process(MPT_ADAPTER *ioc, EventNotificationReply_t *reply)
|
||||
break;
|
||||
case MPI_EVENT_PERSISTENT_TABLE_FULL:
|
||||
INIT_WORK(&ioc->sas_persist_task,
|
||||
mptsas_persist_clear_table,
|
||||
(void *)ioc);
|
||||
mptsas_persist_clear_table);
|
||||
schedule_work(&ioc->sas_persist_task);
|
||||
break;
|
||||
case MPI_EVENT_SAS_DISCOVERY:
|
||||
|
@@ -646,9 +646,10 @@ struct work_queue_wrapper {
|
||||
int disk;
|
||||
};
|
||||
|
||||
static void mpt_work_wrapper(void *data)
|
||||
static void mpt_work_wrapper(struct work_struct *work)
|
||||
{
|
||||
struct work_queue_wrapper *wqw = (struct work_queue_wrapper *)data;
|
||||
struct work_queue_wrapper *wqw =
|
||||
container_of(work, struct work_queue_wrapper, work);
|
||||
struct _MPT_SCSI_HOST *hd = wqw->hd;
|
||||
struct Scsi_Host *shost = hd->ioc->sh;
|
||||
struct scsi_device *sdev;
|
||||
@@ -695,7 +696,7 @@ static void mpt_dv_raid(struct _MPT_SCSI_HOST *hd, int disk)
|
||||
disk);
|
||||
return;
|
||||
}
|
||||
INIT_WORK(&wqw->work, mpt_work_wrapper, wqw);
|
||||
INIT_WORK(&wqw->work, mpt_work_wrapper);
|
||||
wqw->hd = hd;
|
||||
wqw->disk = disk;
|
||||
|
||||
@@ -784,9 +785,10 @@ MODULE_DEVICE_TABLE(pci, mptspi_pci_table);
|
||||
* renegotiate for a given target
|
||||
*/
|
||||
static void
|
||||
mptspi_dv_renegotiate_work(void *data)
|
||||
mptspi_dv_renegotiate_work(struct work_struct *work)
|
||||
{
|
||||
struct work_queue_wrapper *wqw = (struct work_queue_wrapper *)data;
|
||||
struct work_queue_wrapper *wqw =
|
||||
container_of(work, struct work_queue_wrapper, work);
|
||||
struct _MPT_SCSI_HOST *hd = wqw->hd;
|
||||
struct scsi_device *sdev;
|
||||
|
||||
@@ -804,7 +806,7 @@ mptspi_dv_renegotiate(struct _MPT_SCSI_HOST *hd)
|
||||
if (!wqw)
|
||||
return;
|
||||
|
||||
INIT_WORK(&wqw->work, mptspi_dv_renegotiate_work, wqw);
|
||||
INIT_WORK(&wqw->work, mptspi_dv_renegotiate_work);
|
||||
wqw->hd = hd;
|
||||
|
||||
schedule_work(&wqw->work);
|
||||
|
@@ -232,7 +232,7 @@ int i2o_driver_dispatch(struct i2o_controller *c, u32 m)
|
||||
break;
|
||||
}
|
||||
|
||||
INIT_WORK(&evt->work, (void (*)(void *))drv->event, evt);
|
||||
INIT_WORK(&evt->work, drv->event);
|
||||
queue_work(drv->event_queue, &evt->work);
|
||||
return 1;
|
||||
}
|
||||
|
@@ -371,8 +371,10 @@ static int i2o_exec_remove(struct device *dev)
|
||||
* new LCT and if the buffer for the LCT was to small sends a LCT NOTIFY
|
||||
* again, otherwise send LCT NOTIFY to get informed on next LCT change.
|
||||
*/
|
||||
static void i2o_exec_lct_modified(struct i2o_exec_lct_notify_work *work)
|
||||
static void i2o_exec_lct_modified(struct work_struct *_work)
|
||||
{
|
||||
struct i2o_exec_lct_notify_work *work =
|
||||
container_of(_work, struct i2o_exec_lct_notify_work, work);
|
||||
u32 change_ind = 0;
|
||||
struct i2o_controller *c = work->c;
|
||||
|
||||
@@ -439,8 +441,7 @@ static int i2o_exec_reply(struct i2o_controller *c, u32 m,
|
||||
|
||||
work->c = c;
|
||||
|
||||
INIT_WORK(&work->work, (void (*)(void *))i2o_exec_lct_modified,
|
||||
work);
|
||||
INIT_WORK(&work->work, i2o_exec_lct_modified);
|
||||
queue_work(i2o_exec_driver.event_queue, &work->work);
|
||||
return 1;
|
||||
}
|
||||
@@ -460,13 +461,15 @@ static int i2o_exec_reply(struct i2o_controller *c, u32 m,
|
||||
|
||||
/**
|
||||
* i2o_exec_event - Event handling function
|
||||
* @evt: Event which occurs
|
||||
* @work: Work item in occurring event
|
||||
*
|
||||
* Handles events send by the Executive device. At the moment does not do
|
||||
* anything useful.
|
||||
*/
|
||||
static void i2o_exec_event(struct i2o_event *evt)
|
||||
static void i2o_exec_event(struct work_struct *work)
|
||||
{
|
||||
struct i2o_event *evt = container_of(work, struct i2o_event, work);
|
||||
|
||||
if (likely(evt->i2o_dev))
|
||||
osm_debug("Event received from device: %d\n",
|
||||
evt->i2o_dev->lct_data.tid);
|
||||
|
@@ -419,16 +419,18 @@ static int i2o_block_prep_req_fn(struct request_queue *q, struct request *req)
|
||||
|
||||
/**
|
||||
* i2o_block_delayed_request_fn - delayed request queue function
|
||||
* delayed_request: the delayed request with the queue to start
|
||||
* @work: the delayed request with the queue to start
|
||||
*
|
||||
* If the request queue is stopped for a disk, and there is no open
|
||||
* request, a new event is created, which calls this function to start
|
||||
* the queue after I2O_BLOCK_REQUEST_TIME. Otherwise the queue will never
|
||||
* be started again.
|
||||
*/
|
||||
static void i2o_block_delayed_request_fn(void *delayed_request)
|
||||
static void i2o_block_delayed_request_fn(struct work_struct *work)
|
||||
{
|
||||
struct i2o_block_delayed_request *dreq = delayed_request;
|
||||
struct i2o_block_delayed_request *dreq =
|
||||
container_of(work, struct i2o_block_delayed_request,
|
||||
work.work);
|
||||
struct request_queue *q = dreq->queue;
|
||||
unsigned long flags;
|
||||
|
||||
@@ -538,8 +540,9 @@ static int i2o_block_reply(struct i2o_controller *c, u32 m,
|
||||
return 1;
|
||||
};
|
||||
|
||||
static void i2o_block_event(struct i2o_event *evt)
|
||||
static void i2o_block_event(struct work_struct *work)
|
||||
{
|
||||
struct i2o_event *evt = container_of(work, struct i2o_event, work);
|
||||
osm_debug("event received\n");
|
||||
kfree(evt);
|
||||
};
|
||||
@@ -938,8 +941,8 @@ static void i2o_block_request_fn(struct request_queue *q)
|
||||
continue;
|
||||
|
||||
dreq->queue = q;
|
||||
INIT_WORK(&dreq->work, i2o_block_delayed_request_fn,
|
||||
dreq);
|
||||
INIT_DELAYED_WORK(&dreq->work,
|
||||
i2o_block_delayed_request_fn);
|
||||
|
||||
if (!queue_delayed_work(i2o_block_driver.event_queue,
|
||||
&dreq->work,
|
||||
|
@@ -96,7 +96,7 @@ struct i2o_block_request {
|
||||
|
||||
/* I2O Block device delayed request */
|
||||
struct i2o_block_delayed_request {
|
||||
struct work_struct work;
|
||||
struct delayed_work work;
|
||||
struct request_queue *queue;
|
||||
};
|
||||
|
||||
|
Reference in New Issue
Block a user