- From: Dave Lewis <dave.lewis@cs.tcd.ie>
- Date: Thu, 01 Nov 2012 15:34:51 +0000
- To: public-multilingualweb-lt@w3.org
- Message-ID: <5092969B.5070507@cs.tcd.ie>
Hi Marcis, You are correct, we did not consider in depth this differenatiation when first considering the text analysis annotation data category, and we can't apply several text analytics annotations to the same text because of complete overides. The structure you indicate may be a bit complex to parse in cases where only one or a subset of the scores are needed. But I also agree that we don't want to have to add lots of new data categories to address this. Another question, as raised in the meeting today, is whether for disambiguation the confidence score might be better as a part of the disambiguation data categoty? We need to address that, but it doesn't help with the need you indicate for expressing confidence for the terminology and perhaps domain data categories. Another approach is to generalise the different confidence scores in a single data category along the lines sketched out below, effectively merging I think this would still work OK with its-tool _if_ we can refer to individual ITS attributes rather then just complete data categories as specified at present. cheers, Dave Processing Confidence [strawman] Definition The *Processing Confidence *data category is used to communicate the self reported confidence of the accuracy of an annotation. It is not intended to serve as a comparison metric between different engines providing specific annotations, but solely for providing self-reported confidence by the specific system that produced the annotations. This data category does NOT aim to establish any sort of correlation between the self-reported confidence and either human evaluation of usefulness of the output. This data category can be used for several purposes, including, but not limited to: * Automated prioritisation of annotations for further processing based on a quality threshold. * ·Providing readers, translators, post-editors, reviewers and proofreaders of machine translated text with self-reported relative accuracy prediction. * Automated prioritising of raw machine translated text for further processing based on empirically set thresholds. * Providing readers, translators, post-editors, reviewers and proofreaders of machine translated text with self-reported relative accuracy prediction. * MT confidence scores can be displayed e.g.on websites machine translated on the fly, by simple web-based translation editors or Computer Aided Translation (CAT) tools. This data category provides a Confidence Score as a rational number from the interval <0;1> or as percentage (0-100%) with up to 2 decimal digits. The confidence score can be displayed on websites on the fly, by simple content management systems. Confidence MAY be displayed for human consumers as segment annotation or as color-coded font or background. Implementation For elements, the data category informationinherits <http://www.w3.org/International/multilingualweb/lt/drafts/its20/its20.html#def-inheritance>to the annotation of textual content of the element,/including/child elements, but/excluding/attributes.. LOCAL: At least one of the following: ·A mtConf attribute with a value that is a rational number in the interval 0 to 1. ·A disambigConf attribute with a value that is a rational number in the interval 0 to 1. ·A domainConf attribute with a value that is a rational number in the interval 0 to 1. ·A termConf attribute with a value that is a rational number in the interval 0 to 1. On 01/11/2012 10:02, Mārcis Pinnis wrote: > Hi everyone, > > We had a short discussion with Dave about the different confidence scores and we came up to an understanding that the current version might have issues reflecting various kinds of confidences. > > For instance, if we have Named Entity Recognition, Term annotation, then also translation and possible other processes that might affect quality, then: > 1) The text analysis confidence does not allow distinguishing between multiple confidences for one tagged element (for instance, 0.1 - termAnnotator1; 0.5 termAnnotator2; 0.7 - namedEntityRecogniser1, etc.) > 2) Having different confidence namings will make the standard quite heavy (do we really know all the types of confidences that people might like to add/use? Maybe we should not restrict to the ones specified?) > > One solution that I was thinking could be having something like agentConfidence that may contain a complex structure ("0.1|0.5|0.7|etc.") and then have somewhere specified how to understand this sequence. > > Any thoughts? Ideas? > > Best regards, > Mārcis ;o) > ________________________________________ > No: Dave Lewis [dave.lewis@cs.tcd.ie] > Nosūtīts: ceturtdiena, 2012. gada 1. novembrī 10:16 > Kam: public-multilingualweb-lt@w3.org > Tēma: [ACTION-256]: Compile and circulate itsTool examples togehter with proposal text > > Hi, > The attached file contains a revision to Yves its-tools proposal. I have > also added revised text for mt confidence score and text analysis > annotation revised . Finally, I have also added at the end the examples > from these data categories, showing how the engine identification > information (which I've now removed from the data category), can be > repalced using the its-tools element. > > There are comments that need to be addressed in the document. We can go > through these quickly if we wish in the meeting. > > cheers, > Dave > > > > > >
Received on Thursday, 1 November 2012 15:35:27 UTC