lightnvm: pblk: verify that cache read is still valid

When a read is directed to the cache, we risk that the lba has been
updated during the time we made the L2P table lookup and the time we are
actually reading form the cache. We intentionally not hold the L2P lock
not to block other threads.

While strict ordering is not a guarantee at this level (unless REQ_FLUSH
has been previously issued), we have experience that some databases that
have recently implemented direct I/O support, issue metadata reads very
close to the writes, without issuing a fsync in the middle. An easy way
to support them while they is to make an extra effort and check the L2P
map right before reading the cache.

Signed-off-by: Javier González <javier@cnexlabs.com>
Signed-off-by: Matias Bjørling <matias@cnexlabs.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
This commit is contained in:
Javier González 2017-06-30 17:56:42 +02:00 committed by Jens Axboe
parent b5e063a286
commit 076984669d
3 changed files with 22 additions and 6 deletions

View File

@ -150,6 +150,7 @@ static void clean_wctx(struct pblk_w_ctx *w_ctx)
/* Release flags on context. Protect from writes and reads */ /* Release flags on context. Protect from writes and reads */
smp_store_release(&w_ctx->flags, PBLK_WRITABLE_ENTRY); smp_store_release(&w_ctx->flags, PBLK_WRITABLE_ENTRY);
pblk_ppa_set_empty(&w_ctx->ppa); pblk_ppa_set_empty(&w_ctx->ppa);
w_ctx->lba = ADDR_EMPTY;
} }
#define pblk_rb_ring_count(head, tail, size) CIRC_CNT(head, tail, size) #define pblk_rb_ring_count(head, tail, size) CIRC_CNT(head, tail, size)
@ -656,15 +657,17 @@ unsigned int pblk_rb_read_to_bio(struct pblk_rb *rb, struct nvm_rq *rqd,
* be directed to disk. * be directed to disk.
*/ */
int pblk_rb_copy_to_bio(struct pblk_rb *rb, struct bio *bio, sector_t lba, int pblk_rb_copy_to_bio(struct pblk_rb *rb, struct bio *bio, sector_t lba,
u64 pos, int bio_iter) struct ppa_addr ppa, int bio_iter)
{ {
struct pblk *pblk = container_of(rb, struct pblk, rwb);
struct pblk_rb_entry *entry; struct pblk_rb_entry *entry;
struct pblk_w_ctx *w_ctx; struct pblk_w_ctx *w_ctx;
struct ppa_addr l2p_ppa;
u64 pos = pblk_addr_to_cacheline(ppa);
void *data; void *data;
int flags; int flags;
int ret = 1; int ret = 1;
spin_lock(&rb->w_lock);
#ifdef CONFIG_NVM_DEBUG #ifdef CONFIG_NVM_DEBUG
/* Caller must ensure that the access will not cause an overflow */ /* Caller must ensure that the access will not cause an overflow */
@ -674,8 +677,14 @@ int pblk_rb_copy_to_bio(struct pblk_rb *rb, struct bio *bio, sector_t lba,
w_ctx = &entry->w_ctx; w_ctx = &entry->w_ctx;
flags = READ_ONCE(w_ctx->flags); flags = READ_ONCE(w_ctx->flags);
spin_lock(&rb->w_lock);
spin_lock(&pblk->trans_lock);
l2p_ppa = pblk_trans_map_get(pblk, lba);
spin_unlock(&pblk->trans_lock);
/* Check if the entry has been overwritten or is scheduled to be */ /* Check if the entry has been overwritten or is scheduled to be */
if (w_ctx->lba != lba || flags & PBLK_WRITABLE_ENTRY) { if (!pblk_ppa_comp(l2p_ppa, ppa) || w_ctx->lba != lba ||
flags & PBLK_WRITABLE_ENTRY) {
ret = 0; ret = 0;
goto out; goto out;
} }

View File

@ -34,8 +34,7 @@ static int pblk_read_from_cache(struct pblk *pblk, struct bio *bio,
BUG_ON(!pblk_addr_in_cache(ppa)); BUG_ON(!pblk_addr_in_cache(ppa));
#endif #endif
return pblk_rb_copy_to_bio(&pblk->rwb, bio, lba, return pblk_rb_copy_to_bio(&pblk->rwb, bio, lba, ppa, bio_iter);
pblk_addr_to_cacheline(ppa), bio_iter);
} }
static void pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd, static void pblk_read_ppalist_rq(struct pblk *pblk, struct nvm_rq *rqd,

View File

@ -670,7 +670,7 @@ unsigned int pblk_rb_read_to_bio_list(struct pblk_rb *rb, struct bio *bio,
struct list_head *list, struct list_head *list,
unsigned int max); unsigned int max);
int pblk_rb_copy_to_bio(struct pblk_rb *rb, struct bio *bio, sector_t lba, int pblk_rb_copy_to_bio(struct pblk_rb *rb, struct bio *bio, sector_t lba,
u64 pos, int bio_iter); struct ppa_addr ppa, int bio_iter);
unsigned int pblk_rb_read_commit(struct pblk_rb *rb, unsigned int entries); unsigned int pblk_rb_read_commit(struct pblk_rb *rb, unsigned int entries);
unsigned int pblk_rb_sync_init(struct pblk_rb *rb, unsigned long *flags); unsigned int pblk_rb_sync_init(struct pblk_rb *rb, unsigned long *flags);
@ -1037,6 +1037,14 @@ static inline void pblk_ppa_set_empty(struct ppa_addr *ppa_addr)
ppa_addr->ppa = ADDR_EMPTY; ppa_addr->ppa = ADDR_EMPTY;
} }
static inline bool pblk_ppa_comp(struct ppa_addr lppa, struct ppa_addr rppa)
{
if (lppa.ppa == rppa.ppa)
return true;
return false;
}
static inline int pblk_addr_in_cache(struct ppa_addr ppa) static inline int pblk_addr_in_cache(struct ppa_addr ppa)
{ {
return (ppa.ppa != ADDR_EMPTY && ppa.c.is_cached); return (ppa.ppa != ADDR_EMPTY && ppa.c.is_cached);