- From: John C Klensin <john+w3c@jck.com>
- Date: Fri, 13 Sep 2013 06:47:13 -0400
- To: John Daggett <jdaggett@mozilla.com>, Richard Ishida <ishida@w3.org>
- cc: W3C Style <www-style@w3.org>, www International <www-international@w3.org>
--On Thursday, September 12, 2013 18:45 -0700 John Daggett
<jdaggett@mozilla.com> wrote:
>
> Richard Ishida wrote:
>
>> 4.5. Character range: the unicode-range descriptor
>> http://www.w3.org/TR/2013/WD-css-fonts-3-20130711/#unicode-ra
>> nge-desc
>>
>> 'Each <urange> value is a UNICODE-RANGE token made up of a
>> "U+" or "u+" prefix followed by a codepoint range'. The U+
>> is not always needed before every codepoint value (eg. in a
>> range).
>>
>> Why do we need the U+/u+ ? It would be easier to just use
>> bare hex codepoints, especially for ranges, where U+ is only
>> used at the start anyway.
>
> As Tab has already pointed out, the unicode range syntax was
> part of CSS 2.1 syntax and the descriptor itself is already
> supported by multiple implementations so it's not appropriate
> to make a change like this at this point.
After thinking about this a bit more, there is another reason.
U+[N[N]]NNNN rather clearly identifies a Unicode code point --
independent of the particular encoding/representation -- in
general practice. By contrast, "0x...." and its syntactic
equivalents takes us back into the question of whether it is a
Unicode code point or, e.g., UTF-16 or hexified UTF-8. So
there is also a slight argument for U+.... on grounds of clarity
and precision.
john
Received on Friday, 13 September 2013 11:29:36 UTC