- From: Dael Jackson <daelcss@gmail.com>
- Date: Tue, 3 Dec 2019 18:12:42 -0500
- To: www-style@w3.org
- Cc: public-test-infra@w3.org
=========================================
These are the official CSSWG minutes.
Unless you're correcting the minutes,
Please respond by starting a new thread
with an appropriate subject line.
=========================================
Joint Meeting with WPT
----------------------
- Everyone was appreciative for all the work done to improve
documentation on WPT.
- There is still a gap for manual testing. One follow up is to see
if there's a way to convert tests run elsewhere to show on WPT
as a temporary solution.
- Work is being done to associate labels with tests which will
hopefully solve for the open issues of MAY/SHOULD vs MUST tests
and being able to pull tests by spec instead of by directory.
===== FULL MINUTES BELOW =======
Agenda: https://wiki.csswg.org/planning/tpac-2019#agenda
Present:
Rachel Andrew, Fronteers
Rossen Atanassov, Microsoft
L. David Baron, Mozilla
Amelia Bellamy-Royds, Invited Expert
Christian Biesinger, Google
Brian Birtles, Invited Expert
Oriol Brufau, Igalia
Tantek Çelik, Mozilla
Emilio Cobos, Mozilla
Emil A Eklund, Google
Elika Etemad, Invited Expert
Jihye Hong, LGE
Koji Ishii, Google
Sanket Joshi, Microsoft
Brian Kardell, Igalia and Open JS Foundation
Eugene Kashida, Salesforce, Observer
Rune Lillesveen, Google
Chris Lilley, W3C
Peter Linss, Invited Expert
Stanton Marcum, Amazon
Myles C. Maxfield, Apple
Cameron McCormack, Mozilla
Nat McCully, Adobe
Chase Phillips, Google
Florian Rivoal, Invited Expert
Dominik Röttsches, Google
Devin Rousso, Apple
Peter Rushforth, Natural Resources Canada, Invited Expert, Observer
Hiroshi Sakakibara, Beyond Perspective Solutions(BPS)
Simon Sapin, Mozilla
Dirk Schulze, Adobe
Markus Stange, Mozilla
Alan Stearns, Adobe
Aleks Totic, Google
Lea Verou, Invited Expert
Matt Woodrow, Mozilla
Fuqiao Xue, W3C
Scribe: birtles
Joint Meeting with WPT
======================
<astearns> foolip presenting
https://docs.google.com/presentation/d/1F1II3lx0krPBIqAzgU9YFIENC-UBEnZtN796hye_bWw/edit?usp=sharing
archived at https://lists.w3.org/Archives/Public/www-archive/2019Oct/att-0001/wpt-review-2019.pdf
<astearns> +1 to wpt documentation improvements. It's much much
better!
fantasai: Thank you for updating the docs
fantasai: It's great to have everything up front
fantasai: It would be nice to have a navigation system like a tab
bar or footer
fantasai: to connect them all
fantasai: It would be easier to get from wpt.fyi to the docs and back
fantasai: Second comment: do we have any plans for testing printing
or paginated modes?
jgraham: We have talked about it a bit
fantasai: Do we have any way of helping people find and run and
report results for manual tests
fantasai: There are some things that are not automated and are
unlikely to be automated in the near future
foolip: I know Florian does a lot of manual tests
foolip: The answer is no
foolip: and it doesn't report results in the right format
foolip: There are not so many building blocks missing
fantasai: Support for manual testing can close the gaps for the types
of the tests for which we don't have infrastructure
fantasai: e.g. printing and media queries
fantasai: I don't know how we'll test media queries
fantasai: a person has to run those
fantasai: e.g. plugging in a monitor
fantasai: We don't have to run these constantly, but need to run at
least once
fantasai: and maybe a little more frequently, e.g. once a year
foolip: Would it be acceptable if you have to use wpt to run the
test?
foolip: locally wpt run
foolip: ./wpt run --manual would be less work than a website approach
fantasai: It would be nice if you can load the test on a website
fantasai: but the important thing would be to find the manual tests
and then compile a result
fantasai: If you can accept data from Peter's system into wpt
reporting system
fantasai: that might be a way
foolip: Just converting them would be straightforward I guess
foolip: but do you want to keep maintaining that?
plinss: I would be happy for it to be replaced
plinss: once its functions are covered
plinss: It tracks what git revision of each test is used
jgraham: Perhaps we can follow up later
fantasai: Minimum requirement is to get a list of manual tests then
provide a report of results
foolip: Would it be fine to get all the manual tests in a specific
directory?
fantasai: Yes
florian: If the system is designed so that you have to run all of
them, that wouldn't be so convenient
florian: There are thousands of tests in writing-modes for example
foolip: If there's no subdirectories then we can maybe use chunking
jgraham: The requirements to run this probably need to be considered
in more depth
jgraham: There are data model considerations with regards to wpt.fyi
jgraham: There are lots of technical details
florian: I think it helps breaking the problem down because
eventually we want to replace and retire Peter's system
florian: even working out how to display manual tests is an
interesting problem
florian: there might be conflicting reports
florian: Using an existing data source and then working out how to
display them later
astearns: We do need to work out how to fix this
astearns: I want to second the joy at the improved documentation
astearns: I went through the docs to submit a reftest recently and
everything was there
astearns: when I submitted a PR, however, something broken and I
didn't know what to do
astearns: I could see the test results, I submitted another patch
and it passed
astearns: but I don't know how to fix it the first time
astearns: The first two checks were red and I didn't know what to do
about it
astearns: I didn't want to bother anyone
foolip: Please bother us
foolip: We need to fix this
jgraham: Some of this have failed for reasons out of our control
jgraham: e.g. GitHub actions being beta
jgraham: sometimes it takes quite a bit of experience to work out
what went wrong
jgraham: There are infrastructure problems but there are also things
we can do better
dbaron: Is pinging you on #testing the way to reach you?
jgraham: Yes
foolip: Or on the PR
jgraham: IRC is generally more likely to get attention
jgraham: Can get lots on GitHub email
zcorpan: It depends a bit on what it is
zcorpan: If it's something you see repeatedly
zcorpan: it might be a system bug we need to fix
zcorpan: otherwise ping one of us on the PR / IRC
florian: In addition to the manual testing, we need 2 more things to
retire Peter's tool
florian: One is MUST tests vs SHOULD/MAY
florian: important for exiting CR
florian: The other is that wpt.fyi groups report by directory
florian: but using the metadata pulling results by spec link, not
just directory would be helpful
florian: then we can retire Peter's tool
jgraham: We are working on associating labels with tests
jgraham: I think that would cover these cases
jgraham: If you want that information to be in the test files rather
than in metadata
jgraham: then you could have a bot to check they are in sync
florian: Currently that data is in the test, the flags meta
florian: so we need to sync that with labels
jgraham: That labelling system is quite limited but it is being
worked on
jgraham: We could have a script to check they are synced
foolip: If you can search for not labeled tests is that fine?
florian: If there is a button to click that's better, but as long as
it's do-able
florian: One more question about fuzzy tests
florian: For non-fuzzy reftests we often have problems where one
side is done with Ahem and one is not
florian: but the box is quite large so many pixels are affected
florian: is that covered by fuzzy?
jgraham: There are two parameters in fuzzy reftests -- one is the
total number of pixels you allow to differ
jgraham: One is the amount of you allow in difference per color
channel
jgraham: so in this case you would allow a very small range in
difference per color channel
jgraham: If you specify one number it means exactly this number must
be difference, but you can specify a range, e.g. from 0
jgraham: The other thing is that in Gecko we can have meta files
that layer over the wpt tests so we can apply different
fuzzy parameter for Gecko only
jgraham: e.g. we know our form controls differ
florian: For the color channel parameter, it sounds like what I want
to do
florian: but getting the right numbers might be difficult
florian: If we can find a way to make that easier that would be good
jgraham: The screenshot tool could do that
dbaron: The gecko reftest failure tells you that -- the failure range
dbaron: makes it easy to know what numbers to use
AmeliaBR: For SVG I have been planning to work around this for SVG
reftests
AmeliaBR: by adding a stroke overline to cover the anti-aliased part
AmeliaBR: Not the ideal solution
AmeliaBR: but you can block off the regions where you don't care
about differences
emilio: Thank you for all your work
emilio: The tests that keep going to Gecko's internal test suite are
crashtests
emilio: Is there any way we can convert them to wpt in the future
jgraham: I think for crashtests the thinking is that it's normally a
very browser-specific thing
jgraham: There might be cases where that's not true
jgraham: if we think it's worth doing it's quite do-able
emilio: I can try to see how many of the Gecko crashtests affect
other browsers
smfr: Thanks for all your hard work
smfr: getting Safari's stuff working right is great
smfr: I want to confirm that if a run fails in the middle there's a
clear differentiation between failed tests and not run
foolip: If it fails in the middle we don't get results...
smfr: I'm interested in cases where a test doesn't not run... it
should show up as a fail
foolip: A missing test is better than a fail
smfr: Just following up on AmeliaBR's point
smfr: Sometimes there are differences we do care about in a test
with fuzzy parts
smfr: A number is not sufficient, sometimes you want to obscure the
irrelevant part
jgraham: The fuzzy thing is a useful thing, it works in Gecko
jgraham: and when running it locally it will spit out the numbers
you need to annotate them
dbaron: The fuzzy number notation is very useful when they are exact
numbers
dbaron: since if you get any other kind of number you get an
unexpected pass
majidvp: Thank you for adding testdriver annotation
majidvp: For writing scrolling tests, we often need to inject input
majidvp: you support pointer actions, what about mouse
foolip: If pointer actions supports it we can do it
jgraham: If it's not in webdriver we can add a feature request for
that
Received on Tuesday, 3 December 2019 23:13:37 UTC