- From: Dael Jackson <daelcss@gmail.com>
- Date: Tue, 3 Dec 2019 18:12:42 -0500
- To: www-style@w3.org
- Cc: public-test-infra@w3.org
========================================= These are the official CSSWG minutes. Unless you're correcting the minutes, Please respond by starting a new thread with an appropriate subject line. ========================================= Joint Meeting with WPT ---------------------- - Everyone was appreciative for all the work done to improve documentation on WPT. - There is still a gap for manual testing. One follow up is to see if there's a way to convert tests run elsewhere to show on WPT as a temporary solution. - Work is being done to associate labels with tests which will hopefully solve for the open issues of MAY/SHOULD vs MUST tests and being able to pull tests by spec instead of by directory. ===== FULL MINUTES BELOW ======= Agenda: https://wiki.csswg.org/planning/tpac-2019#agenda Present: Rachel Andrew, Fronteers Rossen Atanassov, Microsoft L. David Baron, Mozilla Amelia Bellamy-Royds, Invited Expert Christian Biesinger, Google Brian Birtles, Invited Expert Oriol Brufau, Igalia Tantek Çelik, Mozilla Emilio Cobos, Mozilla Emil A Eklund, Google Elika Etemad, Invited Expert Jihye Hong, LGE Koji Ishii, Google Sanket Joshi, Microsoft Brian Kardell, Igalia and Open JS Foundation Eugene Kashida, Salesforce, Observer Rune Lillesveen, Google Chris Lilley, W3C Peter Linss, Invited Expert Stanton Marcum, Amazon Myles C. Maxfield, Apple Cameron McCormack, Mozilla Nat McCully, Adobe Chase Phillips, Google Florian Rivoal, Invited Expert Dominik Röttsches, Google Devin Rousso, Apple Peter Rushforth, Natural Resources Canada, Invited Expert, Observer Hiroshi Sakakibara, Beyond Perspective Solutions(BPS) Simon Sapin, Mozilla Dirk Schulze, Adobe Markus Stange, Mozilla Alan Stearns, Adobe Aleks Totic, Google Lea Verou, Invited Expert Matt Woodrow, Mozilla Fuqiao Xue, W3C Scribe: birtles Joint Meeting with WPT ====================== <astearns> foolip presenting https://docs.google.com/presentation/d/1F1II3lx0krPBIqAzgU9YFIENC-UBEnZtN796hye_bWw/edit?usp=sharing archived at https://lists.w3.org/Archives/Public/www-archive/2019Oct/att-0001/wpt-review-2019.pdf <astearns> +1 to wpt documentation improvements. It's much much better! fantasai: Thank you for updating the docs fantasai: It's great to have everything up front fantasai: It would be nice to have a navigation system like a tab bar or footer fantasai: to connect them all fantasai: It would be easier to get from wpt.fyi to the docs and back fantasai: Second comment: do we have any plans for testing printing or paginated modes? jgraham: We have talked about it a bit fantasai: Do we have any way of helping people find and run and report results for manual tests fantasai: There are some things that are not automated and are unlikely to be automated in the near future foolip: I know Florian does a lot of manual tests foolip: The answer is no foolip: and it doesn't report results in the right format foolip: There are not so many building blocks missing fantasai: Support for manual testing can close the gaps for the types of the tests for which we don't have infrastructure fantasai: e.g. printing and media queries fantasai: I don't know how we'll test media queries fantasai: a person has to run those fantasai: e.g. plugging in a monitor fantasai: We don't have to run these constantly, but need to run at least once fantasai: and maybe a little more frequently, e.g. once a year foolip: Would it be acceptable if you have to use wpt to run the test? foolip: locally wpt run foolip: ./wpt run --manual would be less work than a website approach fantasai: It would be nice if you can load the test on a website fantasai: but the important thing would be to find the manual tests and then compile a result fantasai: If you can accept data from Peter's system into wpt reporting system fantasai: that might be a way foolip: Just converting them would be straightforward I guess foolip: but do you want to keep maintaining that? plinss: I would be happy for it to be replaced plinss: once its functions are covered plinss: It tracks what git revision of each test is used jgraham: Perhaps we can follow up later fantasai: Minimum requirement is to get a list of manual tests then provide a report of results foolip: Would it be fine to get all the manual tests in a specific directory? fantasai: Yes florian: If the system is designed so that you have to run all of them, that wouldn't be so convenient florian: There are thousands of tests in writing-modes for example foolip: If there's no subdirectories then we can maybe use chunking jgraham: The requirements to run this probably need to be considered in more depth jgraham: There are data model considerations with regards to wpt.fyi jgraham: There are lots of technical details florian: I think it helps breaking the problem down because eventually we want to replace and retire Peter's system florian: even working out how to display manual tests is an interesting problem florian: there might be conflicting reports florian: Using an existing data source and then working out how to display them later astearns: We do need to work out how to fix this astearns: I want to second the joy at the improved documentation astearns: I went through the docs to submit a reftest recently and everything was there astearns: when I submitted a PR, however, something broken and I didn't know what to do astearns: I could see the test results, I submitted another patch and it passed astearns: but I don't know how to fix it the first time astearns: The first two checks were red and I didn't know what to do about it astearns: I didn't want to bother anyone foolip: Please bother us foolip: We need to fix this jgraham: Some of this have failed for reasons out of our control jgraham: e.g. GitHub actions being beta jgraham: sometimes it takes quite a bit of experience to work out what went wrong jgraham: There are infrastructure problems but there are also things we can do better dbaron: Is pinging you on #testing the way to reach you? jgraham: Yes foolip: Or on the PR jgraham: IRC is generally more likely to get attention jgraham: Can get lots on GitHub email zcorpan: It depends a bit on what it is zcorpan: If it's something you see repeatedly zcorpan: it might be a system bug we need to fix zcorpan: otherwise ping one of us on the PR / IRC florian: In addition to the manual testing, we need 2 more things to retire Peter's tool florian: One is MUST tests vs SHOULD/MAY florian: important for exiting CR florian: The other is that wpt.fyi groups report by directory florian: but using the metadata pulling results by spec link, not just directory would be helpful florian: then we can retire Peter's tool jgraham: We are working on associating labels with tests jgraham: I think that would cover these cases jgraham: If you want that information to be in the test files rather than in metadata jgraham: then you could have a bot to check they are in sync florian: Currently that data is in the test, the flags meta florian: so we need to sync that with labels jgraham: That labelling system is quite limited but it is being worked on jgraham: We could have a script to check they are synced foolip: If you can search for not labeled tests is that fine? florian: If there is a button to click that's better, but as long as it's do-able florian: One more question about fuzzy tests florian: For non-fuzzy reftests we often have problems where one side is done with Ahem and one is not florian: but the box is quite large so many pixels are affected florian: is that covered by fuzzy? jgraham: There are two parameters in fuzzy reftests -- one is the total number of pixels you allow to differ jgraham: One is the amount of you allow in difference per color channel jgraham: so in this case you would allow a very small range in difference per color channel jgraham: If you specify one number it means exactly this number must be difference, but you can specify a range, e.g. from 0 jgraham: The other thing is that in Gecko we can have meta files that layer over the wpt tests so we can apply different fuzzy parameter for Gecko only jgraham: e.g. we know our form controls differ florian: For the color channel parameter, it sounds like what I want to do florian: but getting the right numbers might be difficult florian: If we can find a way to make that easier that would be good jgraham: The screenshot tool could do that dbaron: The gecko reftest failure tells you that -- the failure range dbaron: makes it easy to know what numbers to use AmeliaBR: For SVG I have been planning to work around this for SVG reftests AmeliaBR: by adding a stroke overline to cover the anti-aliased part AmeliaBR: Not the ideal solution AmeliaBR: but you can block off the regions where you don't care about differences emilio: Thank you for all your work emilio: The tests that keep going to Gecko's internal test suite are crashtests emilio: Is there any way we can convert them to wpt in the future jgraham: I think for crashtests the thinking is that it's normally a very browser-specific thing jgraham: There might be cases where that's not true jgraham: if we think it's worth doing it's quite do-able emilio: I can try to see how many of the Gecko crashtests affect other browsers smfr: Thanks for all your hard work smfr: getting Safari's stuff working right is great smfr: I want to confirm that if a run fails in the middle there's a clear differentiation between failed tests and not run foolip: If it fails in the middle we don't get results... smfr: I'm interested in cases where a test doesn't not run... it should show up as a fail foolip: A missing test is better than a fail smfr: Just following up on AmeliaBR's point smfr: Sometimes there are differences we do care about in a test with fuzzy parts smfr: A number is not sufficient, sometimes you want to obscure the irrelevant part jgraham: The fuzzy thing is a useful thing, it works in Gecko jgraham: and when running it locally it will spit out the numbers you need to annotate them dbaron: The fuzzy number notation is very useful when they are exact numbers dbaron: since if you get any other kind of number you get an unexpected pass majidvp: Thank you for adding testdriver annotation majidvp: For writing scrolling tests, we often need to inject input majidvp: you support pointer actions, what about mouse foolip: If pointer actions supports it we can do it jgraham: If it's not in webdriver we can add a feature request for that
Received on Tuesday, 3 December 2019 23:13:38 UTC