blkio: Fix compile errors
Fixes compile errors in blk-cgroup code for empty_time stat and a merge fix in CFQ. The first error was when CONFIG_DEBUG_CFQ_IOSCHED is not set. Signed-off-by: Divyesh Shah <dpshah@google.com> Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
This commit is contained in:
@@ -219,6 +219,33 @@ void blkiocg_update_avg_queue_size_stats(struct blkio_group *blkg)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(blkiocg_update_avg_queue_size_stats);
|
EXPORT_SYMBOL_GPL(blkiocg_update_avg_queue_size_stats);
|
||||||
|
|
||||||
|
void blkiocg_set_start_empty_time(struct blkio_group *blkg, bool ignore)
|
||||||
|
{
|
||||||
|
unsigned long flags;
|
||||||
|
struct blkio_group_stats *stats;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&blkg->stats_lock, flags);
|
||||||
|
stats = &blkg->stats;
|
||||||
|
|
||||||
|
if (stats->stat_arr[BLKIO_STAT_QUEUED][BLKIO_STAT_READ] ||
|
||||||
|
stats->stat_arr[BLKIO_STAT_QUEUED][BLKIO_STAT_WRITE]) {
|
||||||
|
spin_unlock_irqrestore(&blkg->stats_lock, flags);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* If ignore is set, we do not panic on the empty flag being set
|
||||||
|
* already. This is to avoid cases where there are superfluous timeslice
|
||||||
|
* complete events (for eg., forced_dispatch in CFQ) when no IOs are
|
||||||
|
* served which could result in triggering the empty check incorrectly.
|
||||||
|
*/
|
||||||
|
BUG_ON(!ignore && blkio_blkg_empty(stats));
|
||||||
|
stats->start_empty_time = sched_clock();
|
||||||
|
blkio_mark_blkg_empty(stats);
|
||||||
|
spin_unlock_irqrestore(&blkg->stats_lock, flags);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(blkiocg_set_start_empty_time);
|
||||||
|
|
||||||
void blkiocg_update_dequeue_stats(struct blkio_group *blkg,
|
void blkiocg_update_dequeue_stats(struct blkio_group *blkg,
|
||||||
unsigned long dequeue)
|
unsigned long dequeue)
|
||||||
{
|
{
|
||||||
@@ -268,33 +295,6 @@ void blkiocg_update_timeslice_used(struct blkio_group *blkg, unsigned long time)
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(blkiocg_update_timeslice_used);
|
EXPORT_SYMBOL_GPL(blkiocg_update_timeslice_used);
|
||||||
|
|
||||||
void blkiocg_set_start_empty_time(struct blkio_group *blkg, bool ignore)
|
|
||||||
{
|
|
||||||
unsigned long flags;
|
|
||||||
struct blkio_group_stats *stats;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&blkg->stats_lock, flags);
|
|
||||||
stats = &blkg->stats;
|
|
||||||
|
|
||||||
if (stats->stat_arr[BLKIO_STAT_QUEUED][BLKIO_STAT_READ] ||
|
|
||||||
stats->stat_arr[BLKIO_STAT_QUEUED][BLKIO_STAT_WRITE]) {
|
|
||||||
spin_unlock_irqrestore(&blkg->stats_lock, flags);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* If ignore is set, we do not panic on the empty flag being set
|
|
||||||
* already. This is to avoid cases where there are superfluous timeslice
|
|
||||||
* complete events (for eg., forced_dispatch in CFQ) when no IOs are
|
|
||||||
* served which could result in triggering the empty check incorrectly.
|
|
||||||
*/
|
|
||||||
BUG_ON(!ignore && blkio_blkg_empty(stats));
|
|
||||||
stats->start_empty_time = sched_clock();
|
|
||||||
blkio_mark_blkg_empty(stats);
|
|
||||||
spin_unlock_irqrestore(&blkg->stats_lock, flags);
|
|
||||||
}
|
|
||||||
EXPORT_SYMBOL_GPL(blkiocg_set_start_empty_time);
|
|
||||||
|
|
||||||
void blkiocg_update_dispatch_stats(struct blkio_group *blkg,
|
void blkiocg_update_dispatch_stats(struct blkio_group *blkg,
|
||||||
uint64_t bytes, bool direction, bool sync)
|
uint64_t bytes, bool direction, bool sync)
|
||||||
{
|
{
|
||||||
|
@@ -2231,7 +2231,7 @@ static int cfq_forced_dispatch(struct cfq_data *cfqd)
|
|||||||
int dispatched = 0;
|
int dispatched = 0;
|
||||||
|
|
||||||
/* Expire the timeslice of the current active queue first */
|
/* Expire the timeslice of the current active queue first */
|
||||||
cfq_slice_expired(cfqd, 0);
|
cfq_slice_expired(cfqd, 0, true);
|
||||||
while ((cfqq = cfq_get_next_queue_forced(cfqd)) != NULL) {
|
while ((cfqq = cfq_get_next_queue_forced(cfqd)) != NULL) {
|
||||||
__cfq_set_active_queue(cfqd, cfqq);
|
__cfq_set_active_queue(cfqd, cfqq);
|
||||||
dispatched += __cfq_forced_dispatch_cfqq(cfqq);
|
dispatched += __cfq_forced_dispatch_cfqq(cfqq);
|
||||||
|
Reference in New Issue
Block a user