forked from luck/tmp_suning_uos_patched
efd4b81abb
We need to flush the batch _before_ we check the number of samples,
otherwise we'll miss all of the batched samples.
Fixes: cf43e6b
("block: add scalable completion tracking of requests")
Signed-off-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Jens Axboe <axboe@fb.com>
257 lines
5.9 KiB
C
257 lines
5.9 KiB
C
/*
|
|
* Block stat tracking code
|
|
*
|
|
* Copyright (C) 2016 Jens Axboe
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/blk-mq.h>
|
|
|
|
#include "blk-stat.h"
|
|
#include "blk-mq.h"
|
|
|
|
static void blk_stat_flush_batch(struct blk_rq_stat *stat)
|
|
{
|
|
const s32 nr_batch = READ_ONCE(stat->nr_batch);
|
|
const s32 nr_samples = READ_ONCE(stat->nr_samples);
|
|
|
|
if (!nr_batch)
|
|
return;
|
|
if (!nr_samples)
|
|
stat->mean = div64_s64(stat->batch, nr_batch);
|
|
else {
|
|
stat->mean = div64_s64((stat->mean * nr_samples) +
|
|
stat->batch,
|
|
nr_batch + nr_samples);
|
|
}
|
|
|
|
stat->nr_samples += nr_batch;
|
|
stat->nr_batch = stat->batch = 0;
|
|
}
|
|
|
|
static void blk_stat_sum(struct blk_rq_stat *dst, struct blk_rq_stat *src)
|
|
{
|
|
blk_stat_flush_batch(src);
|
|
|
|
if (!src->nr_samples)
|
|
return;
|
|
|
|
dst->min = min(dst->min, src->min);
|
|
dst->max = max(dst->max, src->max);
|
|
|
|
if (!dst->nr_samples)
|
|
dst->mean = src->mean;
|
|
else {
|
|
dst->mean = div64_s64((src->mean * src->nr_samples) +
|
|
(dst->mean * dst->nr_samples),
|
|
dst->nr_samples + src->nr_samples);
|
|
}
|
|
dst->nr_samples += src->nr_samples;
|
|
}
|
|
|
|
static void blk_mq_stat_get(struct request_queue *q, struct blk_rq_stat *dst)
|
|
{
|
|
struct blk_mq_hw_ctx *hctx;
|
|
struct blk_mq_ctx *ctx;
|
|
uint64_t latest = 0;
|
|
int i, j, nr;
|
|
|
|
blk_stat_init(&dst[BLK_STAT_READ]);
|
|
blk_stat_init(&dst[BLK_STAT_WRITE]);
|
|
|
|
nr = 0;
|
|
do {
|
|
uint64_t newest = 0;
|
|
|
|
queue_for_each_hw_ctx(q, hctx, i) {
|
|
hctx_for_each_ctx(hctx, ctx, j) {
|
|
blk_stat_flush_batch(&ctx->stat[BLK_STAT_READ]);
|
|
blk_stat_flush_batch(&ctx->stat[BLK_STAT_WRITE]);
|
|
|
|
if (!ctx->stat[BLK_STAT_READ].nr_samples &&
|
|
!ctx->stat[BLK_STAT_WRITE].nr_samples)
|
|
continue;
|
|
if (ctx->stat[BLK_STAT_READ].time > newest)
|
|
newest = ctx->stat[BLK_STAT_READ].time;
|
|
if (ctx->stat[BLK_STAT_WRITE].time > newest)
|
|
newest = ctx->stat[BLK_STAT_WRITE].time;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* No samples
|
|
*/
|
|
if (!newest)
|
|
break;
|
|
|
|
if (newest > latest)
|
|
latest = newest;
|
|
|
|
queue_for_each_hw_ctx(q, hctx, i) {
|
|
hctx_for_each_ctx(hctx, ctx, j) {
|
|
if (ctx->stat[BLK_STAT_READ].time == newest) {
|
|
blk_stat_sum(&dst[BLK_STAT_READ],
|
|
&ctx->stat[BLK_STAT_READ]);
|
|
nr++;
|
|
}
|
|
if (ctx->stat[BLK_STAT_WRITE].time == newest) {
|
|
blk_stat_sum(&dst[BLK_STAT_WRITE],
|
|
&ctx->stat[BLK_STAT_WRITE]);
|
|
nr++;
|
|
}
|
|
}
|
|
}
|
|
/*
|
|
* If we race on finding an entry, just loop back again.
|
|
* Should be very rare.
|
|
*/
|
|
} while (!nr);
|
|
|
|
dst[BLK_STAT_READ].time = dst[BLK_STAT_WRITE].time = latest;
|
|
}
|
|
|
|
void blk_queue_stat_get(struct request_queue *q, struct blk_rq_stat *dst)
|
|
{
|
|
if (q->mq_ops)
|
|
blk_mq_stat_get(q, dst);
|
|
else {
|
|
blk_stat_flush_batch(&q->rq_stats[BLK_STAT_READ]);
|
|
blk_stat_flush_batch(&q->rq_stats[BLK_STAT_WRITE]);
|
|
memcpy(&dst[BLK_STAT_READ], &q->rq_stats[BLK_STAT_READ],
|
|
sizeof(struct blk_rq_stat));
|
|
memcpy(&dst[BLK_STAT_WRITE], &q->rq_stats[BLK_STAT_WRITE],
|
|
sizeof(struct blk_rq_stat));
|
|
}
|
|
}
|
|
|
|
void blk_hctx_stat_get(struct blk_mq_hw_ctx *hctx, struct blk_rq_stat *dst)
|
|
{
|
|
struct blk_mq_ctx *ctx;
|
|
unsigned int i, nr;
|
|
|
|
nr = 0;
|
|
do {
|
|
uint64_t newest = 0;
|
|
|
|
hctx_for_each_ctx(hctx, ctx, i) {
|
|
blk_stat_flush_batch(&ctx->stat[BLK_STAT_READ]);
|
|
blk_stat_flush_batch(&ctx->stat[BLK_STAT_WRITE]);
|
|
|
|
if (!ctx->stat[BLK_STAT_READ].nr_samples &&
|
|
!ctx->stat[BLK_STAT_WRITE].nr_samples)
|
|
continue;
|
|
|
|
if (ctx->stat[BLK_STAT_READ].time > newest)
|
|
newest = ctx->stat[BLK_STAT_READ].time;
|
|
if (ctx->stat[BLK_STAT_WRITE].time > newest)
|
|
newest = ctx->stat[BLK_STAT_WRITE].time;
|
|
}
|
|
|
|
if (!newest)
|
|
break;
|
|
|
|
hctx_for_each_ctx(hctx, ctx, i) {
|
|
if (ctx->stat[BLK_STAT_READ].time == newest) {
|
|
blk_stat_sum(&dst[BLK_STAT_READ],
|
|
&ctx->stat[BLK_STAT_READ]);
|
|
nr++;
|
|
}
|
|
if (ctx->stat[BLK_STAT_WRITE].time == newest) {
|
|
blk_stat_sum(&dst[BLK_STAT_WRITE],
|
|
&ctx->stat[BLK_STAT_WRITE]);
|
|
nr++;
|
|
}
|
|
}
|
|
/*
|
|
* If we race on finding an entry, just loop back again.
|
|
* Should be very rare, as the window is only updated
|
|
* occasionally
|
|
*/
|
|
} while (!nr);
|
|
}
|
|
|
|
static void __blk_stat_init(struct blk_rq_stat *stat, s64 time_now)
|
|
{
|
|
stat->min = -1ULL;
|
|
stat->max = stat->nr_samples = stat->mean = 0;
|
|
stat->batch = stat->nr_batch = 0;
|
|
stat->time = time_now & BLK_STAT_NSEC_MASK;
|
|
}
|
|
|
|
void blk_stat_init(struct blk_rq_stat *stat)
|
|
{
|
|
__blk_stat_init(stat, ktime_to_ns(ktime_get()));
|
|
}
|
|
|
|
static bool __blk_stat_is_current(struct blk_rq_stat *stat, s64 now)
|
|
{
|
|
return (now & BLK_STAT_NSEC_MASK) == (stat->time & BLK_STAT_NSEC_MASK);
|
|
}
|
|
|
|
bool blk_stat_is_current(struct blk_rq_stat *stat)
|
|
{
|
|
return __blk_stat_is_current(stat, ktime_to_ns(ktime_get()));
|
|
}
|
|
|
|
void blk_stat_add(struct blk_rq_stat *stat, struct request *rq)
|
|
{
|
|
s64 now, value;
|
|
|
|
now = __blk_stat_time(ktime_to_ns(ktime_get()));
|
|
if (now < blk_stat_time(&rq->issue_stat))
|
|
return;
|
|
|
|
if (!__blk_stat_is_current(stat, now))
|
|
__blk_stat_init(stat, now);
|
|
|
|
value = now - blk_stat_time(&rq->issue_stat);
|
|
if (value > stat->max)
|
|
stat->max = value;
|
|
if (value < stat->min)
|
|
stat->min = value;
|
|
|
|
if (stat->batch + value < stat->batch ||
|
|
stat->nr_batch + 1 == BLK_RQ_STAT_BATCH)
|
|
blk_stat_flush_batch(stat);
|
|
|
|
stat->batch += value;
|
|
stat->nr_batch++;
|
|
}
|
|
|
|
void blk_stat_clear(struct request_queue *q)
|
|
{
|
|
if (q->mq_ops) {
|
|
struct blk_mq_hw_ctx *hctx;
|
|
struct blk_mq_ctx *ctx;
|
|
int i, j;
|
|
|
|
queue_for_each_hw_ctx(q, hctx, i) {
|
|
hctx_for_each_ctx(hctx, ctx, j) {
|
|
blk_stat_init(&ctx->stat[BLK_STAT_READ]);
|
|
blk_stat_init(&ctx->stat[BLK_STAT_WRITE]);
|
|
}
|
|
}
|
|
} else {
|
|
blk_stat_init(&q->rq_stats[BLK_STAT_READ]);
|
|
blk_stat_init(&q->rq_stats[BLK_STAT_WRITE]);
|
|
}
|
|
}
|
|
|
|
void blk_stat_set_issue_time(struct blk_issue_stat *stat)
|
|
{
|
|
stat->time = (stat->time & BLK_STAT_MASK) |
|
|
(ktime_to_ns(ktime_get()) & BLK_STAT_TIME_MASK);
|
|
}
|
|
|
|
/*
|
|
* Enable stat tracking, return whether it was enabled
|
|
*/
|
|
bool blk_stat_enable(struct request_queue *q)
|
|
{
|
|
if (!test_bit(QUEUE_FLAG_STATS, &q->queue_flags)) {
|
|
set_bit(QUEUE_FLAG_STATS, &q->queue_flags);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|