Re: [ANNOUNCE] fsperf: a simple fs/block performance testing framework
From: Josef Bacik
Date: Sun Oct 08 2017 - 22:25:21 EST
On Mon, Oct 09, 2017 at 11:51:37AM +1100, Dave Chinner wrote:
> On Fri, Oct 06, 2017 at 05:09:57PM -0400, Josef Bacik wrote:
> > Hello,
> >
> > One thing that comes up a lot every LSF is the fact that we have no general way
> > that we do performance testing. Every fs developer has a set of scripts or
> > things that they run with varying degrees of consistency, but nothing central
> > that we all use. I for one am getting tired of finding regressions when we are
> > deploying new kernels internally, so I wired this thing up to try and address
> > this need.
> >
> > We all hate convoluted setups, the more brain power we have to put in to setting
> > something up the less likely we are to use it, so I took the xfstests approach
> > of making it relatively simple to get running and relatively easy to add new
> > tests. For right now the only thing this framework does is run fio scripts. I
> > chose fio because it already gathers loads of performance data about it's runs.
> > We have everything we need there, latency, bandwidth, cpu time, and all broken
> > down by reads, writes, and trims. I figure most of us are familiar enough with
> > fio and how it works to make it relatively easy to add new tests to the
> > framework.
> >
> > I've posted my code up on github, you can get it here
> >
> > https://github.com/josefbacik/fsperf
> >
> > All (well most) of the results from fio are stored in a local sqlite database.
> > Right now the comparison stuff is very crude, it simply checks against the
> > previous run and it only checks a few of the keys by default. You can check
> > latency if you want, but while writing this stuff up it seemed that latency was
> > too variable from run to run to be useful in a "did my thing regress or improve"
> > sort of way.
> >
> > The configuration is brain dead simple, the README has examples. All you need
> > to do is make your local.cfg, run ./setup and then run ./fsperf and you are good
> > to go.
>
> Why re-invent the test infrastructure? Why not just make it a
> tests/perf subdir in fstests?
>
Probably should have led with that shouldn't I have? There's nothing keeping me
from doing it, but I didn't want to try and shoehorn in a python thing into
fstests. I need python to do the sqlite and the json parsing to dump into the
sqlite database.
Now if you (and others) are not opposed to this being dropped into tests/perf
then I'll work that up. But it's definitely going to need to be done in python.
I know you yourself have said you aren't opposed to using python in the past, so
if that's still the case then I can definitely wire it all up.
> > The plan is to add lots of workloads as we discover regressions and such. We
> > don't want anything that takes too long to run otherwise people won't run this,
> > so the existing tests don't take much longer than a few minutes each. I will be
> > adding some more comparison options so you can compare against averages of all
> > previous runs and such.
>
> Yup, that fits exactly into what fstests is for... :P
>
> Integrating into fstests means it will be immediately available to
> all fs developers, it'll run on everything that everyone already has
> setup for filesystem testing, and it will have familiar mkfs/mount
> option setup behaviour so there's no new hoops for everyone to jump
> through to run it...
>
TBF I specifically made it as easy as possible because I know we all hate trying
to learn new shit. I figured this was different enough to warrant a separate
project, especially since I'm going to add block device jobs so Jens can test
block layer things. If we all agree we'd rather see this in fstests then I'm
happy to do that too. Thanks,
Josef