[linux-lvm] Thin Pool Performance
zkabelac at redhat.com
Tue Apr 19 08:11:25 UTC 2016
Dne 19.4.2016 v 03:05 shankha napsal(a):
> Please allow me to describe our setup.
> 1) 8 SSDS with a raid5 on top of it. Let us call the raid device : dev_raid5
> 2) We create a Volume Group on dev_raid5
> 3) We create a thin pool occupying 100% of the volume group.
> We performed some experiments.
> Our random write operations dropped by half and there was significant
> reduction for
> other operations(sequential read, sequential write, random reads) as
> well compared to native raid5
> If you wish I can share the data with you.
> We then changed our configuration from one POOL to 4 POOLS and were able to
> get back to 80% of the performance (compared to native raid5).
> To us it seems that the lvm metadata operations are the bottleneck.
> Do you have any suggestions on how to get back the performance with lvm ?
> LVM version: 2.02.130(2)-RHEL7 (2015-12-01)
> Library version: 1.02.107-RHEL7 (2015-12-01)
Thanks for playing with thin-pool, however your report is largely incomplete.
We do not see you actual VG setup.
Please attach 'vgs/lvs' i.e. thin-pool zeroing (if you don't need it keep it
disabled), chunk size (use bigger chunks if you do not need snapshots), number
of simultaneously active thin volumes in single thin-pool (running hundreds of
loaded thinLV is going to loose battle on locking) , size of thin pool
metadata LV - is this LV located on separate device (you should not use RAID5
and what kind of workload you try on ?
More information about the linux-lvm