RE: de-identification text for Wednesday's call

Dan,

Thank you for being open to further discussion (hopefully not considered 'simply arguing').

I can't speak to your experience with companies you've worked for in the past but I can say that a company that is truly committed to Accountability (as a Privacy concept) takes technical, operational, and administrative controls very seriously (documentation, design reviews, testing, auditing, etc.).  It appears you're seeking a way to capture bad actors for not living up to their promises (say they de-identify and then due to a poorly designed de-identification process or operational controls they allow re-identification to occur - or perhaps more likely that data reserved for a Permitted Use is used for a non-permitted use) - is that correct?

If that is correct, I don't have an easy answer for you as this same issue exists for almost all company privacy promises that involve internal management of data.  I would recommend we take the broader topic of 'external scrutiny of internal data practices' to a different thread/working group.  It's my assumption that if a W3C DNT standard emerges that does invite voluntary implementation by industry, that trade associations will step in to include this in their Codes of Conduct and regular audits of member compliance.  I don't see an immediate solution that would give EFF access to every online company's backend systems to self-review their practices for compliance.

I share your desire to provide meaningful guidance to implementers of de-identification and do this in a way as to not limit innovation in this area and not lock companies into solutions that harm their ability to operate as a business.  I don't believe your non-normative text strikes this balance and offered to come back to the group next week with a few alternate examples that better find this balance.  If you're open to that approach, we can open an action item for me to deliver next week.

Thank you,
Shane

From: Dan Auerbach [mailto:dan@eff.org]
Sent: Tuesday, April 02, 2013 3:59 PM
To: public-tracking@w3.org
Subject: Re: de-identification text for Wednesday's call

Shane,

Thanks for your response, and I suppose we can talk more tomorrow, though we should probably try to identify a path forward as opposed to simply arguing. For example, I would welcome any empirical investigation into the efficacy of various approaches, including operational ones.

I think the spectrum is wider than you suggest and ranges from no change at all to deletion of data. In practice I think deleting data is easy and for many actors the most straight forward way to comply with DNT. My position is far less aggressive than insisting on deleting data. On what I would consider to be the "lighter" end of the spectrum, there are solutions like yours that espouse an undefined mix of technical data manipulation measures and operational controls. Having worked in industry, I'm very skeptical of the value of operational controls, and think that in practice there will be very little accountability. If you'd like to provide more detailed background about the operational controls you have in mind or empirical evidence of the efficacy of such controls, I am certainly willing to listen.

I agree that we shouldn't be too prescriptive about particular methods of de-identification, but I think it's important for the standard to come a clear conclusion about the place in the spectrum that counts as good enough. I don't think that saying, "there's a spectrum, and anything even on the lighter side is OK because companies will be held accountable" is the right answer. My view is that we need to be specific enough to define the contours of what is acceptable, without being too prescriptive, and that's the balance that I hoped to strike with my normative and non-normative text.

On 04/02/2013 02:36 PM, Shane Wiley wrote:
Dan - my apologies - I used those terms to represent one end of a spectrum.  Let's step away from those particular terms to represent the most conservative side of the spectrum and instead speak to the range of possible points of application of consumer protection with respect to de-identification.

On one side of the spectrum we have a possible solution that is completely based on administrative and operational controls (as you pointed out earlier) and on the other side of the spectrum is a possible solution that completely relies on technology alone ("and provides for leeway as anonymization technology improves").  In the middle is a solution that is a mix and allows organizations the latitude to devise a solution that matches their ability to mitigate risk and maintain some value in data - while not being overly prescriptive on the specific technology employed.  All points on the spectrum focus on the desired outcome - that de-identified data not later be re-identified.

While we both agree on the outcome (I believe), we disagree on the best route to get there.  I would suggest we move away from prescriptive elements in the text and provide organizations the flexibility to innovate and build solutions they believe meets the goal.  As long as a company is stating they will meet the goal, you have all of the necessary commitment to hold them accountable to that outcome.

Evidence:  I believe the perceived harm is that de-identified data can later be re-identified, fair?  If yes, then any solution in the spectrum serves that purposes but forces companies most likely to the middle or more conservative side of the spectrum to mitigate their risks without being overly prescriptive.

Industry implementation of a standard that is too weak will be worse for users than having no standard at all, since it will provide only the guise of protection:  Compared to today's world, even this "in the middle" solution to de-identification GREATLY moves the needle in the correct direction - I hope you agree.
As a user, I'd prefer having no DNT at all to a de-identification approach that I would consider too soft. At least with no DNT, I have no false expectation that my data isn't being collected and retained.


  Will there be bad actors or companies that overstate their ability to maintain a de-identified state and fail?  Absolutely.  This approach ensures they will be appropriately held accountable and at the same time allow good actors an attractive (voluntary) option to enhance consumer privacy.

Your concerns:  I do not mean to dismiss your concerns and hopefully have done a better job this time of fairly representing a shared goal and spectrum of options to achieve that goal.  I believe our best path forward is to agree on the goal and avoid prescriptive requirements on exactly how to achieve it.  While we each have our own views into what the rest of the Working Group holds out as priorities (which can only be proven through a vote which the W3C likes to avoid), I hope this compromise position both represents a meaningful advance for consumer protection for consumer advocates and provides enough flexibility for voluntary adoption by industry.

- Shane

From: Dan Auerbach [mailto:dan@eff.org]
Sent: Tuesday, April 02, 2013 1:39 PM
To: public-tracking@w3.org<mailto:public-tracking@w3.org>
Subject: Re: de-identification text for Wednesday's call

Shane,

Labeling my view as "totalitarian" or "absolutist" is inaccurate and not appreciated. My approach allows lots of leeway as anonymization technology improves, and takes into account that perfect anonymization is impossible. It seems to me that you plan to label anything "totalitarian" that suggests that keeping raw logs is not a viable approach in general. To be clear, I don't even want to suggest that keeping raw logs is *always* prohibited, just that it should be in the normal case where records have relatively high entropy. I'm also happy to discuss examples in great detail, but in my experience you have been unwilling to engage in discussing the nitty gritty details of what anonymization should entail.

On 04/02/2013 01:09 PM, Shane Wiley wrote:
Dan,

As with HIPPA, I believe differentiated treatment of internal and external datasets is appropriate as this changes the risk profile of re-identification - again, the root of the conversation being a "risk-based" approach versus a totalitarian approach as you suggest.  My solution meets the perceived consumer harm in this case
Do you have evidence for this claim?




- yours of course does as well but goes far too far over the top of what is actually needed.  If your goal is to create a compromise end-point that will likely be implemented by industry then my recommended approach gets us there.
Any compromise must provide meaningful protection to users. Industry implementation of a standard that is too weak will be worse for users than having no standard at all, since it will provide only the guise of protection.



  If you'd like to instead stand by absolutist approaches, that is of course your prerogative and we'll have those removed through the standard W3C process.  I'm simply trying to save everyone some time and get to a meaningful outcome quickly.
Glibly dismissing my concerns is not a way to gain allies, or move the W3C process forward. I don't think your view is as universally shared as you seem to think that it is.





- Shane

From: Dan Auerbach [mailto:dan@eff.org]
Sent: Tuesday, April 02, 2013 12:21 PM
To: public-tracking@w3.org<mailto:public-tracking@w3.org>
Subject: Re: de-identification text for Wednesday's call

Shane,

Why hash at all in this case? If you are relying on operational and administrative controls, you might as well just pledge not to look up the cookie when you receive it. If you are rotating (and discarding) salts frequently, then it will have a positive effect, but otherwise I don't think hashing provides any benefit here.

But this is an aside to our main disagreement about the larger issue about the role that operational and administrative controls should play. I agree that they should play a role, but only after de-identification of data has been achieved. If the result of a DNT:1 request is business as usual, with minor scrubbing and the caveat that only 4000 engineers at a large corporation get default access to a specially marked database instead of 10000, then that will not be a successful standard. (Of course I welcome more detailed information about operational and administrative controls.)

One last point I wanted to make is that of course the data sets I mentioned refer to public data. We don't have access to internal corporate data sets. There are laws in place to protect the pilfering of that data, so of course no-one is going to steal data then publish an academic paper about it, effectively painting a big target on themselves for federal prosecutors and corporate legal teams. In light of this, the right empirical question to ask is: of large publicly available data sets that contain user data and are somewhat akin to log data, how often are there successful re-identification or attribute disclosure attacks? Can you point to any public data sets where such an attack has not been found?

If your argument is instead that public data should be treated differently from non-public data, then I'd suggest that this is out of scope for the DNT conversation. DNT is about giving users the choice to opt out of tracking by companies, which must entail meaningfully curbing data collection and retention by that company, not merely a request that a company not make public its collected data. (Indeed, in addition to being an excessively weak demand by the user, this would in some cases be a vacuous request, since making that information public is already prohibited by law.) The de-identification question exists within the scope of what the companies themselves can do with the data -- is the data de-identified with respect to the entity that collected the data?

Best,
Dan

On 04/02/2013 11:03 AM, Shane Wiley wrote:
Dan,

Once the one-way hash is applied (and other elements of record appropriately cleansed) the data is moved to a system that is not allowed to be accessed externally.  Its these operational and administrative controls that are essential to ensure de-identified data is not re-identified at some later time.  I believe you're looking only at the technical merits which is only seeing a small portion of the overall solution.

- Shane

From: Dan Auerbach [mailto:dan@eff.org]
Sent: Tuesday, April 02, 2013 10:59 AM
To: public-tracking@w3.org<mailto:public-tracking@w3.org>
Subject: Re: de-identification text for Wednesday's call

On 04/02/2013 08:50 AM, Shane Wiley wrote:
once the one-way hash function has been applied the data is never again able to be accessed in real-time to modify the user's experience.
I think I'm confused, can you explain this more? How is this possible? If you are just hashing a cookie string, your web server receives a request that includes a cookie string, you hash that cookie string (which is in incredibly fast operation), match the hashed cookie against the stored data, and return personalized results.

Or are you salting the hash differently for every request, or combining the cookie with an ephemeral piece of data (the timestamp) before hashing and then throwing away the timestamp?

Thanks for clarifying, apologies if I'm just being dense.

Dan






--

Dan Auerbach

Staff Technologist

Electronic Frontier Foundation

dan@eff.org<mailto:dan@eff.org>

415 436 9333 x134






--

Dan Auerbach

Staff Technologist

Electronic Frontier Foundation

dan@eff.org<mailto:dan@eff.org>

415 436 9333 x134





--

Dan Auerbach

Staff Technologist

Electronic Frontier Foundation

dan@eff.org<mailto:dan@eff.org>

415 436 9333 x134




--

Dan Auerbach

Staff Technologist

Electronic Frontier Foundation

dan@eff.org<mailto:dan@eff.org>

415 436 9333 x134

Received on Wednesday, 3 April 2013 05:16:22 UTC