Re: [PATCH 1/2] KVM: vmx, pmu: accept 0 for absent MSRs when host-initiated
From: Paolo Bonzini
Date: Thu Jun 16 2022 - 06:37:28 EST
On 6/15/22 20:52, Sean Christopherson wrote:
On Thu, Jun 02, 2022, Like Xu wrote:
I actually agree and understand the situation of maintainers/reviewers.
No one wants to maintain flawed code, especially in this community
where the majority of previous contributors disappeared after the code
was merged in. The existing heavy maintenance burden is already visible.
I don't think this is true. I think it's relatively rare for
contributors to disappear.
Thus we may have a maintainer/reviewers scalability issue. Due to missing
trust, competence or mastery of rules, most of the patches sent to the list
have no one to point out their flaws.
Then write tests and run the ones that already exist. Relying purely on reviewers
to detect flaws does not and cannot scale. I agree that we currently have a
scalability issue, but I have different views on how to improve things.
I have privately received many complaints about the indifference of our
community, which is distressing.
You're welcome to expand on these complaints. But I suspect that a lot
of these would come from people that have been told "review other
people's work", "write tests" and/or "you submitted a broken patch" before.
"Let's try to accept" is basically what I did for PEBS and LBR, both of
which I merged basically out of guilt after a little-more-than-cursory
review. It turns out that both of them were broken in ways that weren't
subtle at all; and as a result, other work already queued to 5.19 had to
be bumped to 5.20.
Honestly I should have complained and un-merged them right after seeing
the msr.flat failure. Or perhaps I should have just said "write tests
and then I'll consider the series", but I "tried to accept" and we can
already see it was a failure.
Obviously, "try to accept" is not a 100% commitment and it will fail with high
probability, but such a stance (along with standard clarifications and requirements)
from reviewers and maintainers will make the contributors more concerned,
attract potential volunteers, and focus the efforts of our nominated reviewers.
If it "fails with high probability", all that happened was a waste of
time for everyone involved. Including the submitter who has waited for
weeks for a reviews only to be told "test X fails".
I completely agree on needing better transparency for the lifecycle of patches
going through the KVM tree. First and foremost, there need to be formal, documented
rules for the "official" kvm/* branches, e.g. everything in kvm/queue passes ABC
tests, everything in kvm/next also passes XYZ tests. That would also be a good
place to document expectations, how things works, etc...
Agreed. I think this is a more general problem with Linux development
and I will propose this for maintainer summit.
But again, the relationship between contributors and maintainers should
be of mutual benefit. Rules help contributors, but contributors should
themselves behave and not throw broken patches at maintainers. And
again, guess what the best way is to tell maintainers your patch is not
broken? Include a test. It shows that you are paying attention.
I fully realize that writing tests is not glamorous, and that some of KVM's tooling
and infrastructure is lacking,
I wouldn't say lacking. Sure it's complicated, but between selftests
and kvm-unit-tests the tools *are* there. selftests that allow you to
test migration at an instruction boundary, for example, are not that
hard to write and were very important for features such as nested state
and AMX. They're not perfect, but they go a long way towards giving
confidence in the code; and it's easier to catch weird ioctl policies
from reviewing comprehensive tests than from reviewing the actual KVM code.
We're not talking of something like SEV or TDX here, we're talking about
very boring MSR emulation and only slightly less boring PMU passthrough.
Paolo