[dm-devel] [PATCH 2/3] dm zoned: metadata: use refcount_t for dm zoned reference counters
John Pittman
jpittman at redhat.com
Thu Aug 23 22:54:41 UTC 2018
Sounds good Damien. Thanks for reviewing!
On Thu, Aug 23, 2018 at 6:12 PM, Damien Le Moal <Damien.LeMoal at wdc.com> wrote:
> John,
>
> On 2018/08/23 10:37, John Pittman wrote:
>> The API surrounding refcount_t should be used in place of atomic_t
>> when variables are being used as reference counters. This API can
>> prevent issues such as counter overflows and use-after-free
>> conditions. Within the dm zoned metadata stack, the atomic_t API
>> is used for mblk->ref and zone->refcount. Change these to use
>> refcount_t, avoiding the issues mentioned.
>>
>> Signed-off-by: John Pittman <jpittman at redhat.com>
>> ---
>> drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
>> drivers/md/dm-zoned.h | 2 +-
>> 2 files changed, 14 insertions(+), 13 deletions(-)
>>
>> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
>> index 969954915566..92e635749414 100644
>> --- a/drivers/md/dm-zoned-metadata.c
>> +++ b/drivers/md/dm-zoned-metadata.c
>> @@ -99,7 +99,7 @@ struct dmz_mblock {
>> struct rb_node node;
>> struct list_head link;
>> sector_t no;
>> - atomic_t ref;
>> + refcount_t ref;
>
> While reviewing your patch, I realized that this ref is always manipulated under
> the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
> refcount. An unsigned int would do as well and be faster. My fault.
>
> I will send a patch to go on top of yours to fix that.
>
> Otherwise:
>
> Reviewed-by: Damien Le Moal <damien.lemoal at wdc.com>
> Tested-by: Damien Le Moal <damien.lemoal at wdc.com>
>
> Thanks !
>
>
>> unsigned long state;
>> struct page *page;
>> void *data;
>> @@ -296,7 +296,7 @@ static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd,
>>
>> RB_CLEAR_NODE(&mblk->node);
>> INIT_LIST_HEAD(&mblk->link);
>> - atomic_set(&mblk->ref, 0);
>> + refcount_set(&mblk->ref, 0);
>> mblk->state = 0;
>> mblk->no = mblk_no;
>> mblk->data = page_address(mblk->page);
>> @@ -397,7 +397,7 @@ static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd,
>> return NULL;
>>
>> spin_lock(&zmd->mblk_lock);
>> - atomic_inc(&mblk->ref);
>> + refcount_inc(&mblk->ref);
>> set_bit(DMZ_META_READING, &mblk->state);
>> dmz_insert_mblock(zmd, mblk);
>> spin_unlock(&zmd->mblk_lock);
>> @@ -484,7 +484,7 @@ static void dmz_release_mblock(struct dmz_metadata *zmd,
>>
>> spin_lock(&zmd->mblk_lock);
>>
>> - if (atomic_dec_and_test(&mblk->ref)) {
>> + if (refcount_dec_and_test(&mblk->ref)) {
>> if (test_bit(DMZ_META_ERROR, &mblk->state)) {
>> rb_erase(&mblk->node, &zmd->mblk_rbtree);
>> dmz_free_mblock(zmd, mblk);
>> @@ -511,7 +511,8 @@ static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd,
>> mblk = dmz_lookup_mblock(zmd, mblk_no);
>> if (mblk) {
>> /* Cache hit: remove block from LRU list */
>> - if (atomic_inc_return(&mblk->ref) == 1 &&
>> + refcount_inc(&mblk->ref);
>> + if (refcount_read(&mblk->ref) == 1 &&
>> !test_bit(DMZ_META_DIRTY, &mblk->state))
>> list_del_init(&mblk->link);
>> }
>> @@ -753,7 +754,7 @@ int dmz_flush_metadata(struct dmz_metadata *zmd)
>>
>> spin_lock(&zmd->mblk_lock);
>> clear_bit(DMZ_META_DIRTY, &mblk->state);
>> - if (atomic_read(&mblk->ref) == 0)
>> + if (refcount_read(&mblk->ref) == 0)
>> list_add_tail(&mblk->link, &zmd->mblk_lru_list);
>> spin_unlock(&zmd->mblk_lock);
>> }
>> @@ -1048,7 +1049,7 @@ static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
>> }
>>
>> INIT_LIST_HEAD(&zone->link);
>> - atomic_set(&zone->refcount, 0);
>> + refcount_set(&zone->refcount, 0);
>> zone->chunk = DMZ_MAP_UNMAPPED;
>>
>> if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) {
>> @@ -1574,7 +1575,7 @@ struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd)
>> void dmz_activate_zone(struct dm_zone *zone)
>> {
>> set_bit(DMZ_ACTIVE, &zone->flags);
>> - atomic_inc(&zone->refcount);
>> + refcount_inc(&zone->refcount);
>> }
>>
>> /*
>> @@ -1585,7 +1586,7 @@ void dmz_activate_zone(struct dm_zone *zone)
>> */
>> void dmz_deactivate_zone(struct dm_zone *zone)
>> {
>> - if (atomic_dec_and_test(&zone->refcount)) {
>> + if (refcount_dec_and_test(&zone->refcount)) {
>> WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags));
>> clear_bit_unlock(DMZ_ACTIVE, &zone->flags);
>> smp_mb__after_atomic();
>> @@ -2308,7 +2309,7 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>> mblk = list_first_entry(&zmd->mblk_dirty_list,
>> struct dmz_mblock, link);
>> dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)",
>> - (u64)mblk->no, atomic_read(&mblk->ref));
>> + (u64)mblk->no, refcount_read(&mblk->ref));
>> list_del_init(&mblk->link);
>> rb_erase(&mblk->node, &zmd->mblk_rbtree);
>> dmz_free_mblock(zmd, mblk);
>> @@ -2326,8 +2327,8 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>> root = &zmd->mblk_rbtree;
>> rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
>> dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree",
>> - (u64)mblk->no, atomic_read(&mblk->ref));
>> - atomic_set(&mblk->ref, 0);
>> + (u64)mblk->no, refcount_read(&mblk->ref));
>> + refcount_set(&mblk->ref, 0);
>> dmz_free_mblock(zmd, mblk);
>> }
>>
>> diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h
>> index 12419f0bfe78..b7829a615d26 100644
>> --- a/drivers/md/dm-zoned.h
>> +++ b/drivers/md/dm-zoned.h
>> @@ -78,7 +78,7 @@ struct dm_zone {
>> unsigned long flags;
>>
>> /* Zone activation reference count */
>> - atomic_t refcount;
>> + refcount_t refcount;
>>
>> /* Zone write pointer block (relative to the zone start block) */
>> unsigned int wp_block;
>>
>
>
> --
> Damien Le Moal
> Western Digital Research
More information about the dm-devel
mailing list