W3C home > Mailing lists > Public > public-silver@w3.org > April 2019

Re: Heuristic testing

From: Detlev Fischer <detlev.fischer@testkreis.de>
Date: Thu, 25 Apr 2019 12:07:52 +0200
To: public-silver@w3.org
Message-ID: <b314983f-118e-f59e-a3d7-67e412e4008e@testkreis.de>
Hi Alastair, Charles, list,

I still feel uneasy about including methods used by an organisation at 
the design )or re-design) stage in a conformance evaluation. There are 
several problems with that:

 1. Evaluators will often not have the domain knowledge to assess, say,
    whether a grouping of navigation items (e.g. products) works well
    for the target audience (think a chemicals supplier)
 2. An expert might arrive at as a good a navigation structure as a
    group that went through a card-sorting excercise (if one were to
    carry out user testing to assess the quality of the result) - why
    should the fact that the structure was arrived in card sorting lead
    to a higher score if what counts is the accessibility/usability of
    the site for the end user?
 3. The fact that changes were made as a result of testing is the back
    story of the site being conformance-evaluated - for the user it has
    no impact on the actual user experience of the content  used. So why
    should it appear in a conformance result? (I do not mind - actually
    welcome - it those measures appear in another kind of rubic that may
    be labeled "accessible organisational processes", "proactive
    organisation", "company digs accessibility" or whatever).


I think there must be a clear separation of a conformance score that can 
be arrived at ideally by any external evaluator based on published 
techniques and common tools, without the need for domain knowledge and 
without access to company internals, and something that for want of a 
better word I will call a 'proactivity score', which is derived from 
insight into the organisation's internal processes. These may be shown 
as "stacking up" - conformance leading to "bronze" and the 'proactivity 
score' adding points for "silver" and finally "gold" - but I would 
personally prefer a side-by-side presentation to make it clear that this 
is addressing different aspects - site properties on the one hand, 
organisational properties on the other.

Just to be clear: I welcome extending the scope of conformance testing 
to things that are 'hard(er) to measure' but still do not rely on 
knowledge of company internals. These aspects might include things such 
as "proximity of related information" or "Concise navigation structure" 
(e.g., not more than x elements per level, consistent display of 
hierarchy in nested structures and of process steps in processes, etc - 
things that may often enter assessments in things like 3.2.3 Consistent 
Navigation already, but may not be explicitly measured). For these, the 
challenge will be to find a way to integrate measurement scales with the 
current PASS/FAIL approach.

Detlev

Am 25.04.2019 um 11:30 schrieb Alastair Campbell:
>
> Hi Charles,
>
> Thank you for the extra context, and I hope it didn’t come across as 
> negative. I do appreciate the thought-process, I was just worried 
> about how it might be taken.
>
> Taking a little step back to consider the various UX/User Centered 
> Design methods, I’ve long been of the opinion that:
>
>   * UCD is good for optimising for the majority of people within a
>     particular context / domain.
>   * Accessibility guidelines (so far) have been good for ensuring that
>     interface works for the most people possible.
>
> In our work UX tends to lead accessibility, so you define a good 
> solution for the task, then make sure it is as robust & accessible as 
> possible. (They aren’t separate, but iterative. Oh, and obviously 
> people with disabilities are part of the user-research, but we work 
> out the task first, then the interface.)
>
> Where the UCD methods shine is dealing with the context of the 
> problem, and getting out of your own mindset & assumptions.
>
> That means they are method to get to a more optimal solution, but not 
> a way to /compare/ solutions. That’s a really tough problem as the 
> context matters hugely, which is something that world-wide guidelines 
> cannot take account of.
>
> As a quick example of ‘context’ differences, the main UX problems you 
> work on in e-commerce are Information Architecture based, such as how 
> to display 10,000 products in a way that people can navigate to what 
> they want. Whereas something like web-based email is much more of an 
> interface problem.
>
> > The general idea would simply be to encourage practices that go 
> beyond the minimum, but not require them.
>
> In that context I can see at least one way forward then, where there 
> are a set of guidelines oriented around usability/IA that are 
> process-based.
>
> For example, the guideline could be (quick hypothetical example):
>
>   * Users can understand and use navigation elements which have more
>     than 10 options.
>
> The method(s) would be /process/ based, like ISO 27001 where you 
> essentially self-mark but have to show improvement each year.
>
> For example:
>
>   * Conduct a card-sorting exercise to establish the best groupings
>     and terms for the navigation.
>   * Conduct a menu test to optimise the terms used in the navigation.
>   * Conduct a heuristic evaluation of the navigation’s placement and
>     design.
>
> The ‘conformance’ for each of these is that you record that this 
> method has been used, and perhaps what changes you made as a result, 
> or even /that/ you made changes as a result.
>
> Then Silver is not trying to define a ‘good’ or replicable result 
> across the multitude of different websites, but provide a way of 
> scoring higher for organisations following best-practice UCD. In the 
> context of ‘going above the baseline’, that makes sense to me.
>
> I think it also helps to have these tasks as methods under particular 
> guidelines, rather than as an overall methodology for testing all the 
> guidelines. Then they could mix with some baseline methods from WCAG 
> 2.x as well, with these methods there for higher scoring.
>
> Cheers,
>
> Alastair
>

-- 
Detlev Fischer
Testkreis
Werderstr. 34, 20144 Hamburg

Mobil +49 (0)157 57 57 57 45

http://www.testkreis.de
Beratung, Tests und Schulungen für barrierefreie Websites
Received on Thursday, 25 April 2019 10:08:32 UTC

This archive was generated by hypermail 2.4.0 : Friday, 17 January 2020 17:23:58 UTC