- From: Jeanne Spellman <jspellman@spellmanconsulting.com>
- Date: Tue, 25 Jun 2019 20:30:55 -0400
- To: Silver Task Force <public-silver@w3.org>
- Message-ID: <d1ffc9ba-77e6-38a4-ff4b-87241edc2e65@spellmanconsulting.com>
Formatted version of minutes:
https://www.w3.org/2019/06/25-silver-minutes.html
Text of minutes:
[1]W3C
[1] http://www.w3.org/
- DRAFT -
Silver Community Group Teleconference
25 Jun 2019
Attendees
Present
Lauriat, jeanne, Makoto, JF, Cyborg, AngelaAccessForAll
Regrets
Denis, Luis, KIm
Chair
Shawn, Jeanne
Scribe
Cyborg
Contents
* [2]Topics
1. [3]Schedule (U.S. holiday)
2. [4]Review milestones & timeline
3. [5]Review current conformance state
* [6]Summary of Action Items
* [7]Summary of Resolutions
__________________________________________________________
<jeanne> s//zakim, order agenda 1,4,2,3
<jeanne> rrsgent, make minutes
<jeanne> scribe: Cyborg
Schedule (U.S. holiday)
Shawn: U.S. holiday on July 4, Canada on July 1
... next week is a wash - after this Friday, meet again on July
9 (Tues)
<JF> +1 to July 9
<jeanne> +1
Jeanne: we need to talk about milestones and timelines from
meeting with Alastair, some agreement needed from this group
before it goes in the Charter
<Zakim> JF, you wanted to comment on @ - Tests
Review milestones & timeline
Jeanne: welcome back Kelsey (spelling?) - very excited to get
her back, a hard and prolific worker
Kelsey Callister: was at Baylor University, looking for work in
UX
Jeanne: AGWG meeting with chairs on Monday - Alastair did a
week 5 timeline to get to CR
<jeanne>
[8]https://docs.google.com/spreadsheets/d/1X5tc7HJ0jIY_u3bLj7iT
bxO1fxCeX1-SigpYE42Vz90/edit?ts=5d10d93c#gid=0
[8] https://docs.google.com/spreadsheets/d/1X5tc7HJ0jIY_u3bLj7iTbxO1fxCeX1-SigpYE42Vz90/edit?ts=5d10d93c#gid=0
<jeanne> By week schedule
<jeanne> Milestones for Silver
[9]https://www.w3.org/WAI/GL/task-forces/silver/wiki/Major_Mile
stones_for_Silver
[9] https://www.w3.org/WAI/GL/task-forces/silver/wiki/Major_Milestones_for_Silver
Jeanne: updated Silver milestones based on that
... rather than editor's draft for Charter, will publish
something we can get comments on with guidelines, methods,
tests, so people can see how conformance would work with real
examples - in October
... public editor's draft in Jan 2020
... developing new Silver content in March 2020, with next
CSUN, and a year doing that. then candidate recommendations,
maintenance and responding to comments
... a year for responding to comments.
Alastair: a couple of assumptions and explanations. it is in 3
columns: content is self-explanatory, conformance model on
left, based on idea that we would have different people working
on different things. we do need sample content to test sample
conformance with. guidelines, methods to test conformance
model. once we get past first editor's draft, it is difficult
to say how long things will take, may involve working group
past that point.
... this was a first pass to get something down in detail, and
easier to answer questions now.
<Zakim> alastairc, you wanted to mention a few of the
assumptions in the timeline
John: looking at Nov 25 and Dec 2, 2019 - don't believe there
is enough time for that, we don't even have a model at this
point. there are discussions and drafts and people interested
in getting involved.
... what does Jeanne mean about cutting off work, if cognitive
walkthroughs not completed are they jettisoned?
Jeanne: we would work on guidance same as AGWG, in short number
at time, complete them, editor's draft, heartbeat drafts,
anything that has not made it into working draft would be
postponed to next version. next version starts right after CR,
so they can start on that work immediately even before version
goes to rec.
John: appreciate year to deal with comments, but that is very
aggressive. migration from 1.0 to 2.0 was 18 months.
... crystal ball is cloudy.
Alastair: it will become clearer. we have 400 odd reasonable
techniques in 2.x we could migrate those to Silver (6 months to
a year).
... finger in the wind, but best detection we have available.
John: under promise, but over deliver.
Alastair: we have that discussion lined up in the next meeting.
Jeanne: we shouldn't promise any Silver dates beyond length of
Charter
John: in Wiki page, depending on length of Charter, we have
Silver rec in 2022.
Jeanne: if candidate recommendation goes longer, it goes
longer.
John: agrees that we get to candidate recommendation and then
some flexibility after that.
Kelsey: who are we anticipating comments from?
Jeanne: general public, W3C. 2.0 got 1000 public comments. we
don't know what it will look like, but it could be a lot.
Review current conformance state
Jeanne: let's talk about conformance. take a look at what we
currently have on conformance. a lot of people are repeating
work we already did. some creative new ideas. would like to
give everyone a chance to get caught up on work that has been
done.
... 2 major phases working on conformance. a year ago last
summer, subgroup did basic work on structure of conformance and
in sept to november, once we had IA solid, and did more work on
conformance, bringing the two together. here's what we have
already done and to recap some of the proposals which have come
out this week. one of the things Cybele and I worked on this
weekend, boiling down 35 pages of emails on work we did last
year
... to turn it into a digestible summary
<jeanne>
[10]https://docs.google.com/document/d/1wklZRJAIPzdp2RmRKZcVsyR
dXpgFbqFF6i7gzCRqldc/edit#heading=h.sevi88jq0fiq
[10] https://docs.google.com/document/d/1wklZRJAIPzdp2RmRKZcVsyRdXpgFbqFF6i7gzCRqldc/edit#heading=h.sevi88jq0fiq
Jeanne: put a lot of work to keep Wiki main page up to date to
always find things, including a lot of the old work. but this
is the latest of the conformance design, nothing really
changed, but polished and organized and easier to read
... so people not involved with Silver can get caught up in
broad strokes
... set series of goals for conformance, coming out of design
sprint - score cards and rubrics, solving problem of
substantially meets, and where people with large sites can show
Silver conformance
... access supported, flexible method of claiming conformance.
... number of issues still outstanding. but let's focus on
point system that is ...
<jeanne> How do we set up a point scoring system that will be
transparent, fair, and motivate or reward organizations to do
more?
Jeanne: how do we maintain system that is current and protected
from gaming? migrating? methodologies? lots of issues. plus
others raised this week.
... in November, we put together IA and conformance prototype
would work with it. flattening structure of 2.x to guidelines
and methods. methods includes tests, examples, instructions.
tagging engine to find more easily and API to extract info for
own purposes
... migrating to Silver - WCAG principles become tags,
guidelines and SC to guidelines, tech specific criteria will
move to methods, techniques will move to methods, and
understanding becomes part of the Guideline Explainer.
... A, AA, AAA levels deleted, Silver conformance overall for
product or project, not specific SC or guideline
... auto and manual testing, rubrics and distance from mean,
task completion, etc. Jeanne walking through the document.
... scoring system - when I wrote this, we had rough work and
alternate proposals. from John and Bruce. will update that.
... levels not by SC, but overall for project or product as
defined by org
John: I'll let you finish
Jeanne: what Cybele and I did this weekend, work we did on
points system, bringing it up to date to show how point system
can work
<jeanne>
[11]https://docs.google.com/spreadsheets/d/1sOQ6odaK43pV4VfHSFP
XAV7Ry1KlcGDxbyWy2Vb1d-s/edit#gid=0
[11] https://docs.google.com/spreadsheets/d/1sOQ6odaK43pV4VfHSFPXAV7Ry1KlcGDxbyWy2Vb1d-s/edit#gid=0
Jeanne: this is not something that the general public would
see, this is done in background. public info, not hiding it,
but not in the face of users. but on legal and regulatory side,
there are people who are very interested in ensuring that
system is transparent and fair, what goes on at particular
levels are not as transparent and fair as what people are
asking for
... a lot of work from last summer is how do we set up a point
system that is transparent and fair
<jeanne>
[12]https://docs.google.com/document/d/1ccKlaPMaVvazbSqMPgttvMe
sy9D0KAjGY01pAQES2K0/edit#
[12] https://docs.google.com/document/d/1ccKlaPMaVvazbSqMPgttvMesy9D0KAjGY01pAQES2K0/edit
Jeanne: here is the explainer for the spreadsheet
... the first thing we started with is to rank user needs, took
headings content which is most complete starting guideline. we
ranked user needs, and immediately ran into a problem, which is
that all of them were critical. 3 is most important, 2 middle,
1 low. ways to move that to guideline points. one thing added
is to look at reputation points, but we didn't work that in
detail yet. but where we said we had a lot of power and were
clear how to do this
... was when we got to individual methods. when I was trying to
move the spreadsheet we did last summer into this one, took 4
tries. how do we express this. formulas. factors. many
repetitions. realization was that every score has three
components: information from guideline, information from
method, and information from test (info is really math factor)
<Kelsey> What are reputation points...would that be a score for
the organization itself, rather than the digital product we're
scoring?
Jeanne: the guideline needs the most work. when we look at the
methods, 3 aspects: ease of implementation (vs hard),
effectiveness (does it work for all the user needs - can be
full or partial and related to quality), and does it allow
customization (bonus)
... method for customizing headings could be on browser side,
Wayne Dick did presentation on how people with low vision use
customized spreadsheets to remove white space around headings,
that would be a browser method
John: as content author, no control over what browser I use
Jeanne: headings method for customization would not apply to
content author, I will add that...
... we had a number of tests, proof of concept, to put numbers
in here. took Bruce Bailey's idea that instead of numbering
things as 1 worth 1, we use order of magnitude change between
them, to amplify differences and make them more visible
... walking through spreadsheet...
John: lots of concerns with proposed scoring mechanism,
rewarding methods as opposed to outcomes.
Jeanne: correct you, as result of tests, which are result of
outcomes
Shawn: effectiveness is incorrect, as it is written and
proposed, it looks like associated with methods
Jeanne: div with ARIA may not be correct
Shawn: the point around it is less than for that example than
that example is for the first place
John: for screen reader user, not capturing if semantic
structure correct
Jeanne: that is quality test, rubric that we did 2 weeks ago
John: how will that have an impact on scoring
... scoring seems to happen at page level, how do we bring that
up to a master score. how can I get a 72, instead of a 100 or 1
... what if it works for some people but not all?
... addressing language of page is easy to do, but language of
content inside of a page is harder to do.
... two SC, linked, one is hard to do, one easy to do, what is
relationship between them in terms of points?
... we need to look at this in a wholistic way, when we assign
points, where did 100 come from? how do we make that
determination?
Jeanne: these are all really good questions, we need to work on
that. i'm trying to bring you up to date on where we are.
John: I wasn't part of the work that happened, but the focus
seems to be on methods and that is not the right thing to be
testing and scoring.
Jeanne: we will discuss that in the conformance group
... we will address proposals today and Friday
... we spent a lot of time last summer to look at rubrics and
formulas in order to look at ranking SC, but everything we
tried, 35 pages of emails, any time we said this piece of
guidance is more important than this, we ended up with a bias
against the group that wasn't as important.
<Zakim> Lauriat, you wanted to touch on the point system and
progress so far.
Jeanne: we tried to rank by priority of user need, but there is
always some group for whom that is critical. and we did
experimenting with it here, we did not see a way to rank needs
because it wasn't fair to people with some disabilities.
... given the regulatory need for this, we couldn't rank needs.
<JF> From Jeanne's roll-up of comments and previous work:
Techniques tell you if a particular solution was used, out of a
possible infinite number of ways to do it, a technique
describes one solution. That's fantastic if you're a developer
and want to know common ways to do something correct. It's not
useful if you want to know if something conformanc, because not
using the technique doesn't actually tell you if something is
non-conformant. Rules do ju
Alastair: there will be some kick-off for the conformance work,
some people continue what we've worked on so far, if people
have alternate idea, run with that for a few weeks and see what
we are going to get to
<JF> (cont.) Rules do just that. If a rule is applicable it
will tell you if (part of) an accessibility requirement is
non-conformant.
<alastairc> Cybelle: Will we be able to get to google doc
explainer? Have some concerns that we should test our
assumptions before people work on various things.
<JF> A HUGE +1 for Cybele's p[oint re: assumptions
<Lauriat> +1
Kelsey: with outcome of not being able to rank criteria based
on user needs, how do we avoid that? how is that practical at
an organizational level?
... is that a roadblock?
Jeanne: having a filter or tag of priorities or things, using
easy checks as a guide - here are the things you can do first
because they are easy to implement, not about impact on
disability, they have a big impact and are easy to do.
Kelsey: thought the new scoring is motivator for updates, how
does that align with all criteria being equally prioritized
<Zakim> alastairc, you wanted to suggest conformance group
starts with requirements overview, and then persues 2(ish)
different models.
Jeanne: trying to get rid of inherent bias, but when you look
at the report, this disability best served by A are people with
no vision. low vision is AA, so says people with no vision are
higher priority than people with low vision. that is the bias
we are working against. trying to take what we did last summer,
and move it from november
... where we got to in November in conformance.
Kelsey: that no vision vs low vision helpful to understand
Jeanne: kick off on Friday who is working on what guideline. if
you know you are working on a guideline, to please work on it,
we NEED that information for conformance, we DESPERATELY need
the info at the guidance, test, method level
... to have real content to test
John: lots of assumptions here, and a lot concern me, questions
raised that have not been directly answered.
Jeanne: once you see previously work laid out, you see
assumptions, you realize how much we need data - guidelines,
methods and tests, in order to test assumptions. instead of
just debating
... this is why writing guidelines is PARAMOUNT
... to test assumptions
Received on Wednesday, 26 June 2019 00:31:21 UTC