forked from luck/tmp_suning_uos_patched
block: remove deadline __deadline manipulation helpers
No users left since the removal of the legacy request interface, we can remove all the magic bit stealing now and make it a normal field. But use WRITE_ONCE/READ_ONCE on the new deadline field, given that we don't seem to have any mechanism to guarantee a new value actually gets seen by other threads. Reviewed-by: Hannes Reinecke <hare@suse.com> Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
parent
8f4236d900
commit
079076b341
|
@ -325,7 +325,7 @@ static struct request *blk_mq_rq_ctx_init(struct blk_mq_alloc_data *data,
|
||||||
rq->special = NULL;
|
rq->special = NULL;
|
||||||
/* tag was already set */
|
/* tag was already set */
|
||||||
rq->extra_len = 0;
|
rq->extra_len = 0;
|
||||||
rq->__deadline = 0;
|
WRITE_ONCE(rq->deadline, 0);
|
||||||
|
|
||||||
rq->timeout = 0;
|
rq->timeout = 0;
|
||||||
|
|
||||||
|
@ -839,7 +839,7 @@ static bool blk_mq_req_expired(struct request *rq, unsigned long *next)
|
||||||
if (rq->rq_flags & RQF_TIMED_OUT)
|
if (rq->rq_flags & RQF_TIMED_OUT)
|
||||||
return false;
|
return false;
|
||||||
|
|
||||||
deadline = blk_rq_deadline(rq);
|
deadline = READ_ONCE(rq->deadline);
|
||||||
if (time_after_eq(jiffies, deadline))
|
if (time_after_eq(jiffies, deadline))
|
||||||
return true;
|
return true;
|
||||||
|
|
||||||
|
|
|
@ -84,7 +84,7 @@ void blk_abort_request(struct request *req)
|
||||||
* immediately and that scan sees the new timeout value.
|
* immediately and that scan sees the new timeout value.
|
||||||
* No need for fancy synchronizations.
|
* No need for fancy synchronizations.
|
||||||
*/
|
*/
|
||||||
blk_rq_set_deadline(req, jiffies);
|
WRITE_ONCE(req->deadline, jiffies);
|
||||||
kblockd_schedule_work(&req->q->timeout_work);
|
kblockd_schedule_work(&req->q->timeout_work);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(blk_abort_request);
|
EXPORT_SYMBOL_GPL(blk_abort_request);
|
||||||
|
@ -121,14 +121,16 @@ void blk_add_timer(struct request *req)
|
||||||
req->timeout = q->rq_timeout;
|
req->timeout = q->rq_timeout;
|
||||||
|
|
||||||
req->rq_flags &= ~RQF_TIMED_OUT;
|
req->rq_flags &= ~RQF_TIMED_OUT;
|
||||||
blk_rq_set_deadline(req, jiffies + req->timeout);
|
|
||||||
|
expiry = jiffies + req->timeout;
|
||||||
|
WRITE_ONCE(req->deadline, expiry);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* If the timer isn't already pending or this timeout is earlier
|
* If the timer isn't already pending or this timeout is earlier
|
||||||
* than an existing one, modify the timer. Round up to next nearest
|
* than an existing one, modify the timer. Round up to next nearest
|
||||||
* second.
|
* second.
|
||||||
*/
|
*/
|
||||||
expiry = blk_rq_timeout(round_jiffies_up(blk_rq_deadline(req)));
|
expiry = blk_rq_timeout(round_jiffies_up(expiry));
|
||||||
|
|
||||||
if (!timer_pending(&q->timeout) ||
|
if (!timer_pending(&q->timeout) ||
|
||||||
time_before(expiry, q->timeout.expires)) {
|
time_before(expiry, q->timeout.expires)) {
|
||||||
|
|
35
block/blk.h
35
block/blk.h
|
@ -238,26 +238,6 @@ void blk_account_io_start(struct request *req, bool new_io);
|
||||||
void blk_account_io_completion(struct request *req, unsigned int bytes);
|
void blk_account_io_completion(struct request *req, unsigned int bytes);
|
||||||
void blk_account_io_done(struct request *req, u64 now);
|
void blk_account_io_done(struct request *req, u64 now);
|
||||||
|
|
||||||
/*
|
|
||||||
* EH timer and IO completion will both attempt to 'grab' the request, make
|
|
||||||
* sure that only one of them succeeds. Steal the bottom bit of the
|
|
||||||
* __deadline field for this.
|
|
||||||
*/
|
|
||||||
static inline int blk_mark_rq_complete(struct request *rq)
|
|
||||||
{
|
|
||||||
return test_and_set_bit(0, &rq->__deadline);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline void blk_clear_rq_complete(struct request *rq)
|
|
||||||
{
|
|
||||||
clear_bit(0, &rq->__deadline);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline bool blk_rq_is_complete(struct request *rq)
|
|
||||||
{
|
|
||||||
return test_bit(0, &rq->__deadline);
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Internal elevator interface
|
* Internal elevator interface
|
||||||
*/
|
*/
|
||||||
|
@ -322,21 +302,6 @@ static inline void req_set_nomerge(struct request_queue *q, struct request *req)
|
||||||
q->last_merge = NULL;
|
q->last_merge = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* Steal a bit from this field for legacy IO path atomic IO marking. Note that
|
|
||||||
* setting the deadline clears the bottom bit, potentially clearing the
|
|
||||||
* completed bit. The user has to be OK with this (current ones are fine).
|
|
||||||
*/
|
|
||||||
static inline void blk_rq_set_deadline(struct request *rq, unsigned long time)
|
|
||||||
{
|
|
||||||
rq->__deadline = time & ~0x1UL;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline unsigned long blk_rq_deadline(struct request *rq)
|
|
||||||
{
|
|
||||||
return rq->__deadline & ~0x1UL;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Internal io_context interface
|
* Internal io_context interface
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -224,9 +224,7 @@ struct request {
|
||||||
refcount_t ref;
|
refcount_t ref;
|
||||||
|
|
||||||
unsigned int timeout;
|
unsigned int timeout;
|
||||||
|
unsigned long deadline;
|
||||||
/* access through blk_rq_set_deadline, blk_rq_deadline */
|
|
||||||
unsigned long __deadline;
|
|
||||||
|
|
||||||
union {
|
union {
|
||||||
struct __call_single_data csd;
|
struct __call_single_data csd;
|
||||||
|
|
Loading…
Reference in New Issue
Block a user