[PATCH v11] drm: Add initial ci/ subdirectory
Maxime Ripard
mripard at kernel.org
Thu Sep 14 02:54:28 PDT 2023
Hi,
On Tue, Sep 12, 2023 at 02:16:41PM +0100, Daniel Stone wrote:
> Hopefully less mangled formatting this time: turns out Thunderbird +
> plain text is utterly unreadable, so that's one less MUA that is
> actually usable to send email to kernel lists without getting shouted
> at.
Sorry if it felt that way, it definitely wasn't my intention to shout at
you. Email is indeed kind of a pain to deal with, and I wanted to keep
the discussion going.
> On Mon, 11 Sept 2023 at 15:46, Maxime Ripard <mripard at kernel.org> wrote:
> > On Mon, Sep 11, 2023 at 03:30:55PM +0200, Michel Dänzer wrote:
> > > > There's in 6.6-rc1 around 240 reported flaky tests. None of them have
> > > > any context. That new series hads a few dozens too, without any context
> > > > either. And there's no mention about that being a plan, or a patch
> > > > adding a new policy for all tests going forward.
> > >
> > > That does sound bad, would need to be raised in review.
> > >
> > > > Any concern I raised were met with a giant "it worked on Mesa" handwave
> > >
> > > Lessons learned from years of experience with big real-world CI
> > > systems like this are hardly "handwaving".
> >
> > Your (and others) experience certainly isn't. It is valuable, welcome,
> > and very much appreciated.
> >
> > However, my questions and concerns being ignored time and time again
> > about things like what is the process is going to be like, what is going
> > to be tested, who is going to be maintaining that test list, how that
> > interacts with stable, how we can possibly audit the flaky tests list,
> > etc. have felt like they were being handwaived away.
>
> Sorry it ended up coming across like that. It wasn't the intent.
>
> > I'm not saying that because I disagree, I still do on some, but that's
> > fine to some extent. However, most of these issues are not so much an
> > infrastructure issue, but a community issue. And I don't even expect a
> > perfect solution right now, unlike what you seem to think. But I do
> > expect some kind of plan instead of just ignoring that problem.
> >
> > Like, I had to ask the MT8173 question 3 times in order to get an
> > answer, and I'm still not sure what is going to be done to address that
> > particular issue.
> >
> > So, I'm sorry, but I certainly feel like it here.
>
> I don't quite see the same picture from your side though. For example,
> my reading of what you've said is that flaky tests are utterly
> unacceptable, as are partial runs, and we shouldn't pretend otherwise.
> With your concrete example (which is really helpful, so thanks), what
> happens to the MT8173 hdmi-inject test? Do we skip all MT8173 testing
> until it's perfect, or does MT8173 testing always fail because that
> test does?
It's not clear to me why that test is even running in the first place?
There's been some confusion on my side here about what we're going to
test with this. You've mentioned Mesa and GPUs before, but that's a KMS
test so there must be more to it.
Either way, it's a relevant test so I guess why not. It turns out that
the test is indeed flaky, I guess we could add it to the flaky tests
list.
BUT
I want to have every opportunity to fix whatever that failure is.
So:
- Is the test broken? If so, we should report it to IGT dev and remove
it from the test suite.
- If not, is that test failure have been reported to the driver author?
- If no answer/fix, we can add it to the flaky tests list, but do we
have some way to reproduce the test failure?
The last part is especially critical. Looking at the list itself, I have
no idea what board, kernel version, configuration, or what the failure
rate was. Assuming I spend some time looking at the infra to find the
board and configuration, how many times do I have to run the tests to
expect to reproduce the failure (and thus consider it fixed if it
doesn't occur anymore).
Like, with that board and test, if my first 100 runs of the test work
fine, is it reasonable for me to consider it fixed, or is it only
supposed to happen once every 1000 runs?
So, ideally, having some (mandatory) metadata in the test lists with a
link to the bug report, the board (DT name?) it happened with, the
version and configuration it was first seen with, and an approximation
of the failure rate for every flaky test list.
I understand that it's probably difficult to get that after the fact on
the tests that were already merged, but I'd really like to get that
enforced for every new test going forward.
That should hopefully get us in a much better position to fix some of
those tests issues. And failing that, I can't see how that's
sustainable.
And Mesa does show what I'm talking about:
$ find -name *-flakes.txt | xargs git diff --stat e58a10af640ba58b6001f5c5ad750b782547da76
src/amd/ci/deqp-radv-stoney-aco-flakes.txt | 18 +++++
src/broadcom/ci/deqp-v3d-rpi4-flakes.txt | 2 +
src/broadcom/ci/deqp-v3dv-rpi4-flakes.txt | 15 ++++
src/broadcom/ci/deqp-vc4-rpi3-flakes.txt | 30 ++++++++
src/broadcom/ci/piglit-v3d-rpi4-flakes.txt | 14 ++++
src/broadcom/ci/piglit-vc4-rpi3-flakes.txt | 5 ++
src/freedreno/ci/deqp-freedreno-a307-flakes.txt | 6 ++
src/freedreno/ci/deqp-freedreno-a530-flakes.txt | 31 +++++++++
src/freedreno/ci/deqp-freedreno-a630-bypass-flakes.txt | 11 +++
src/freedreno/ci/deqp-freedreno-a630-flakes.txt | 109 ++++++++++++++++++++++++++++++
src/gallium/drivers/freedreno/ci/piglit-freedreno-a530-flakes.txt | 21 ++++++
src/gallium/drivers/freedreno/ci/piglit-freedreno-a630-flakes.txt | 31 +++++++++
src/gallium/drivers/llvmpipe/ci/deqp-llvmpipe-flakes.txt | 4 ++
src/gallium/drivers/panfrost/ci/deqp-panfrost-g52-flakes.txt | 6 ++
src/gallium/drivers/panfrost/ci/deqp-panfrost-t720-flakes.txt | 4 ++
src/gallium/drivers/panfrost/ci/deqp-panfrost-t760-flakes.txt | 1 +
src/gallium/drivers/panfrost/ci/deqp-panfrost-t860-flakes.txt | 1 +
src/gallium/drivers/radeonsi/ci/piglit-radeonsi-stoney-flakes.txt | 1 +
src/gallium/drivers/softpipe/ci/deqp-softpipe-flakes.txt | 2 +
src/gallium/drivers/virgl/ci/deqp-virgl-gl-flakes.txt | 3 +
src/gallium/drivers/zink/ci/deqp-zink-lvp-flakes.txt | 0
src/gallium/drivers/zink/ci/piglit-zink-lvp-flakes.txt | 0
src/gallium/frontends/lavapipe/ci/deqp-lvp-flakes.txt | 1 +
23 files changed, 316 insertions(+)
In the history of Mesa, there's never been a single test removed from a
flaky test list.
> Both have their downsides. Not doing any testing has the obvious
> downside, and means that the driver can get worse until it gets
> perfect. Always marking the test as failed means that the test results
> are useless: if failure is expected, then red is good. I mean, say
> you're contributing a patch to fix some documentation or add a helper
> to common code which only v3d uses. The test results come back, and
> your branch is failing tests on MT8173, specifically the
> hdmi-inject at 4k test. What then? Either as a senior contributor you
> 'know' that's the case, or as a casual contributor you get told 'oh
> yeah, don't worry about the test results, they always fail'. Both lead
> to the same outcome, which is that no-one pays any attention to the
> results, and they get worse.
I believe the above would achieve a better result for us kernel folks,
while still providing enough flexibility for you.
> What we do agree on is that yes, those tests should absolutely be
> fixed, and not just swept under the rug. Part of this is having
> maintainers actually meaningfully own their test results. For example,
> I'm looking at the expectation lists for the Intel gen in my laptop,
> and I'm seeing a lot of breakage in blending tests, as well as
> dual-display fails which include the resolution of my external
> display.
I just bought a 7900XT and I'm pretty much in the same situation, so I
definitely feel what you're saying :)
However...
> I'd expect the Intel driver maintainers to look at them, get them
> fixed, and gradually prune those xfails/flakes down towards zero.
>
> If the maintainers don't own it though, then it's not going to get
> fixed. And we are exactly where we are today: broken plane blending
> and 1440p on KBL, broken EDID injection on MT8173, and broken atomic
> commits on stoney. Without stronger action from the maintainers (e.g.
> throwing i915 out of the tree until it has 100% pass 100% of the
> time), adding testing isn't making the situation better or worse in
> and of itself. What it _is_ doing though, is giving really clear
> documentation of the status of each driver, and backing that up by
> verifying it.
... I haven't reported any of the failures yet so I don't really expect
them to be fixed by themselves. That's on me.
> Only maintainers can actually fix the drivers (or the tests tbf). But
> doing the testing does let us be really clear to everyone what the
> actual state is, and that way people can make informed decisions too.
> And the only way we're going to drive the test rate down is by the
> subsystem maintainers enforcing it.
Just FYI, I'm not on the other side of the fence there, I'd really like
to have some kind of validation. I talked about it at XDC some years
ago, and discussed it several people at length over the years. So I'm
definitely not in the CI-is-bad camp.
> Does that make sense on where I'm (and I think a lot of others are)
> coming from?
That makes sense from your perspective, but it's not clear to me how you
can expect maintainers to own the tests if they were never involved in
the process.
They are not in Cc of the flaky tests patches, they are not reported
that the bug is failing, how can they own that process if we never
reached out and involved them?
We're all overworked, you can't expect them to just look at the flaky
test list every now and then and figure it out.
Maxime
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 228 bytes
Desc: not available
URL: <http://lists.infradead.org/pipermail/linux-mediatek/attachments/20230914/65a54134/attachment.sig>
More information about the Linux-mediatek
mailing list