- From: <meetings@w3c-ccg.org>
 - Date: Fri, 20 Jun 2025 18:04:23 -0400
 - To: public-credentials@w3.org
 - Message-ID: <CA+ChqYebnnk1-QOjgsScc5Q4Oj-DwOw2DopD7fn5xYSSLcNAzw@mail.gmail.com>
 
W3C Data Integrity Community Group Meeting Summary - 2025/06/20 *Topics Covered:* - *Community Updates:* Significant progress reported on a lightweight self-asserted skill credential tool (demo upcoming). Broadening adoption of verifiable credentials noted, particularly within the UN and World Bank for vital records and social services in developing nations (Open CRVS platform mentioned). Collaboration opportunity identified between this tool and the UN/World Bank initiative. - *Quantum-Safe Crypto Suites:* A pull request was merged, incorporating multiple quantum-safe crypto suites (MLDSA, hash-based, Falcon, and SQI). Next steps include renaming, addressing data duplication, and finalizing spec markup. August presentation planned. - *Everlasting Unlinkability (Anna's Paper & BBS Pseudonyms):* Discussion centered on a new paper proposing "everlasting anonymous rate-limited tokens," offering a similar but not identical solution to the group's goal of everlasting unlinkability in pseudonyms. The existing vector of secrets approach was deemed the most practical for current standardization efforts due to the complexity and lack of standardization of alternative cryptographic techniques presented in the paper. Further work needed to determine acceptable vector size based on threat modeling. Progress on getting BBS pseudonyms into Last Call was also discussed. - *Query over Credentials (Circuit-Based Implementation):* Collaboration with the Berkeley cryptography team on a circuit-based implementation for querying over credentials was discussed. Key design decisions regarding max input size, query complexity, and data types/operations needed to be prioritized for production. A consensus emerged to focus on transforming input data for efficiency with existing cryptographic mechanisms rather than inventing new signature schemes. Concerns were raised about the potential for unintentional privacy leakage through overly correlated queries. *Key Points:* - There's significant and expanding real-world adoption of verifiable credentials, particularly in social services and vital records management. - While promising, advanced cryptographic techniques for everlasting unlinkability (like those in Anna's paper) introduce complexities and require further standardization work before being suitable for immediate integration. The existing vector of secrets approach remains the most viable short-term solution. - For query-over-credentials, the focus should be on efficient data transformation to utilize existing cryptographic infrastructure rather than developing novel signature schemes. Careful consideration of query complexity and attribute correlation is crucial for maintaining privacy. - Defining acceptable limits for the size of the vector of secrets and clarifying the number of joins and attributes needed in queries are essential next steps. Text: https://meet.w3c-ccg.org/archives/w3c-ccg-data-integrity-2025-06-20.md Video: https://meet.w3c-ccg.org/archives/w3c-ccg-data-integrity-2025-06-20.mp4 *Data Integrity - 2025/06/20 09:58 EDT - Transcript* *Attendees* Geun-Hyung Kim, Greg Bernstein, Hiroyuki Sano, Jesse Wright, John's Notetaker, Manu Sporny, Parth Bhatt, Phillip Long, Ted Thibodeau Jr *Transcript* Manu Sporny: All right, let's go ahead and get started. I think it might be a little bit of a light call today. We've got multiple people out on paid time off vacation. but we want to keep kind of the momentum up. So not much of an agenda today. Apologies for that. I've been traveling did not get back until earlier this week. but I think we've got a couple of items to touch base on today. Greg Bernstein: Perfect. Manu Sporny: One of them is I think Greg you finished a preliminary analysis on Anna's paper on the everlasting u unlinkability portion of the vector of secrets thing. they tried a different approach from the one that we had before based on some kind of really fresh new research done by Anna Lassinskia and friends. So we'll get a little bit of an update on that. Manu Sporny: I've got a small update on the quantum safe crypto suites. the pull request that will put in that got merged down and I've got some questions kind of on next steps there. and then Jesse, I know that you asked a couple of questions that I still have not been able to get around to answering in the W3C data integrity Slack channel. And so I thought we might use some of the time today to talk about that. so that's the proposed agenda. are there any other changes or updates to the agenda? Anything additional that people would like to discuss today? if not, let's go ahead and kind of jump into things. Are there any community updates? Manu Sporny: anything else that we should be aware of that's happening anything of that nature. And the expectation is it's the summer so things are going to slow down quite a bit. I will note that I've been contacted by a number of people deploying verifiable credentials in areas that I was completely unaware that there was movement. but the United Nations is starting to verifiable credentials for things like marriage certificates in developing nation primarily using open- source software for vital records type stuff. 00:05:00 Manu Sporny: So places like Nepal Cambodia Zambia and some other places are starting to deploy verifiable credentials for kind of core social services things. So I unfortunately don't have the link on me but every day they're new kind of interesting uses of verifiable credentials and they're using data integrity as a part of that. the projects are kind of multifaceted at the UN it's World Bank as well. Manu Sporny: So, World Bank, United Nations, economic development funds, out of Germany, I forget where else, Sweden, forget where else, but they're covering vital records. they're covering farm worker, kind of like farm output in, these nations, food reserves. they're tracking benefits that individuals would receive including education benefits like paying for schools and textbooks and things like that. as well as disability benefits if someone has a disability of some kind. Manu Sporny: and so it's been interesting to kind of see and they're also looking for solutions that not only work in digital form but paper form as well. So the verifiable credential barcode stuff and the ultra compact signatures and things like that are of interest. So, that's kind of a heads up that the work is kind of broadening and there's some interesting new groups. and these are not small groups. The meeting groups are 40 to 50 people on a call at any given time. So, it's a large effort that I was completely unaware of. and of course, I'm unfortunately forgetting the name of the effort right now. Manu Sporny: I'll try to find a link for that later on. Any other updates from anyone else? Phillip Long: Manner, this is Phil. Manu Sporny: Hey Phil Phillip Long: Just as an update for everyone, we've been working for a while on a lightweight self asserted skill credential tool that essentially is launched into your browser and allows you to create a self- asserted credential and then request through a recommendation pathway for an endorsement to that work is basically finishing up. We're doing a final demo of it on this coming Wednesday to the funders which have been primarily Walmart and through the T3 US Chamber of Commerce Foundation. Phillip Long: in addition to that there has been a version of it that has been built for small businesses which allows for an employment verification with a quas two-factor approach. it's pretty flimsy but it works. and then the create and then a credential for performance review for volunteering and of course skills as well that a supervisor is endorsing on that has been that from the creation of the credential by the individual employee. Phillip Long: And finally, a P for building a verifiable credential-based resume. that allows you to have a narrative section that has links into other credential elements of the resume. so that you have a story you can tell backed up by individual C VCs that are associated with the data behind that and shows it as a PDF if you wish but also creates what we're calling a compound credential similar to a CLRV2 so that it verifies by typical verification methods and… 00:10:00 Phillip Long: that's almost all done so we are going to demonstrate that at bad sum in July. Thanks, Manu Sporny: That's excellent news,… Manu Sporny: Really excited to see those things progress as well. some overlap with the kind of self asserted workforce skills or the kind of witnessed workforce skills thing. this is one of the projects that looks like it's going to be adopting verifiable credentials. Manu Sporny: is open CRVS I forget it's something vital records but it's an open source platform and this is what I wanted to kind of point out Phil is the way they have approached this source product is that it is a tool set that a government or region county state federal would give its population and then you effectively onboard people in the community to witness certain events. Manu Sporny: These are births and deaths and marriages and things like that, and so this kind of endorsement process It's very decentralized. it starts with somebody going out to a village to collect data on a birth. They do a bunch of interviewing for example, the child's name from the parents. they talk to potentially the midwife to get other kind of vitals information if the birth didn't happen in a hospital. Phillip Long: What? Yep. Manu Sporny: And then they kind of submit it to the country's vital records or the county's whatever level the vital record services are run they submit it there and then you have kind of government officials endorse that. Manu Sporny: And so I thought that that was really interesting because it's aligned with kind of what you were talking about as well, this endorsement style workflow and they started back in 2015. They ran the first set of pilots in 2017 through 2019. And the field feedback that they got in these, countries where this stuff was deployed was that it was very effective. meaning they had a 50% gain in recorded births between,… Phillip Long: Right. Manu Sporny: not having the system in place and having the system in place. and so I mean, 50% jump in, the number of, events that you're recording is significant. and I would imagine that it would kind of apply to workforce training skills as well, right? Manu Sporny: Because again a lot of the workforce and training stuff that happens in some of these countries is I mean quite literally village and family based. Manu Sporny: It doesn't necessarily go through community colleges and universities and things like that. So, yeah, there's Yep. Phillip Long: Right. Do… Phillip Long: if you have a contact for me to follow up and I'd love to give them access to our repo and they can take whatever they want and vice versa. Manu Sporny: Yeah, that sounds good. I'm just u reaching out to them. I'll follow up an email and put you in touch. Phillip Long: Thank you. Manu Sporny: But here's the website if you want to copy it. Phillip Long: Yeah, I've already opened it on my machine. Manu Sporny: Okay, great. Manu Sporny: And I'll put it and… Phillip Long: Got it. Manu Sporny: they're just one of the platforms. Manu Sporny: there, 15 or 20 other platforms that are looking at this, but there they are adopting kind of verifiable credentials for it. And so, it feels like it's very complimentary to the stuff that you're doing in the education workforce training sector. Phillip Long: And there's a variant of this that I think is likely to be funded which is a registry for notaries that the notaries can endorse the identity of a person… Phillip Long: who is claiming various things and that could be used as a way of strengthening the credibility of the individual self- selfisssued credentials. Manu Sporny: Yep. Manu Sporny: Absolutely. okay, good stuff. any other community updates? Any other comments? Manu Sporny: all Let's go ahead and jump into the agenda. I don't have any particular order, but maybe here, I'll do it. Quantum safe crypto suite updates, then Greg, your updates on the everlasting unlinkability, and then Jesse, maybe we can tackle your questions towards the tail end of the call. any objections to that order? real quick, the, Quantum Safe, Crypto Suites, had a pull request that, Will had put in in April. We finally merged it a month and a half later. but a lot of it just had to do with kind of getting the commit order correct. 00:15:00 Manu Sporny: so we have the quantum safe crypto suites kind of broken up right now. The naming's a bit awkward. We need to fix that. There's quite a bit of duplication of data, but you can see here that we've got MLDDSA stateless hashbased one Falcon and SQI. I think we do need to rename those. but for each one like MLDDSA we've got the create proof we have the instantiation algorithms fully documented here. Manu Sporny: We've got, what you do for the proof values, creating verifying the serialization, of the verification of the proof, that sort of thing. so, that's good. I think there's kind of common transformations or common algorithms are broken out for transformation hashing a proof configuration that sort of thing. so it was a good step forward. Manu Sporny: the proof representations are also provided here and big signatures because theyre key formats are here. We haven't picked the varant u identifiers yet except for MLDDSA. and then their e values that we're going to have to kind of update here as and large it's in there. We've got some spec markup bugs that need to be cleaned up, but I think we're good other than that. Okay, so that's moving forward. that's good. Manu Sporny: Unfortunately u Andrea and Yarm were going to present on the quantum safe crypto suites last Tuesday, but unfortunately there's a mixup and they ended up on the wrong channel unfortunately. so they're scheduled for an August presentation on this stuff. okay, I think that's it until we get the next set of PRs in. things that we need to do here. Select the multi key things. select the crypto suite names and then a little more clean up of the spec text. Manu Sporny: Anything else on the quantum safe crypto suites before we move on? if not, the second item up is the updates on the everlasting unlinkability stuff that Greg has been working Greg, over to you for the update. Greg Bernstein: Thanks. Greg Bernstein: So, recently published, oops, let me drop a link to the paper. There's a new paper published by Anna Stefano Taro, University of Washington, Brown University and University of Washington, plus their grad students and, some folks in Germany. Greg Bernstein: by the way, I stick that link into Here's the chat. So, this is a paper on what they call everlasting anonymous rate limited limiting tokens. and this is a similar problem, but not exactly the same as ours. Greg Bernstein: What they'd like to do, and this is a privacy pass type of thing, is an issue where would hand like a holder, a user, they call them, something to generate up to K tokens that they can use N at a time, meaning, you could use it in kind of quickly batches of N uses. but not more than that within a specific period of time. 00:20:00 Greg Bernstein: So, that's an extra thing on top of what we'd like to do with pseudonyms where we'd like to allow people up to pay uses where they never could lose having this everlasting anonymity, be anonymous for up to K uses even in the presence of a quantum computer that could break discrete log or whatever. ever. Okay, so that's the everlasting meaning that even if a quantum computer comes up, you still can preserve up to some number of times that you use the thing. they use some interesting techniques, from ZKP and other places. Greg Bernstein: one of okay on the same thing let me just give a quick update with BBS so BBS the workg groupoup chairs are pinging some of the reviewers because the questions from the cryptographic review have been answered a while back but we never got feedback from the reviewers and so over at CFRG the workg group chairs Oops is that the wrong times out. Manu Sporny: The site's down. Unfortunately, Greg, I just checked. Yeah. Greg Bernstein: I left my browser up on the page, so I didn't realize it gone down. Greg Bernstein: so progress is being made towards getting the reviewers to doublech checkck the feedback and see if the corrections or their questions have been answered on the base so some progress on getting BBS. The goal is to get in the last call and all that stuff. on pseudonyms. That's why we're looking at this paper because we were asked to get something better in case in terms of everlasting unlinkability with pseudonyms. And the approach that we've taken is to use a vector of pseudonyms of secrets to help do that. Greg Bernstein: That's the same approach in this everlasting anonymous rate limited token paper. And so we're pursuing similar goals. they have an extra complication but they're not tying it to a full set of signature suite properties. the paper mentions like a BBS scheme but they use a different scheme tie it with it up with a different signature scheme. So right now it's not directly applicable to BBS but a technique like this could be. Greg Bernstein: The other thing that they bring up which could be useful in the future with BBS or other schemes is in the ZKP community there's a difference between what are known as snarks and starks. So these are types of zero knowledge proofs. One requires what's known as a common reference string or these parameters that have to be set up in such a way that you have to have trust or in how these parameters got established. Greg Bernstein: the secrets behind them need to be either securely kept or completely destroyed and people not know where they have any way of getting that otherwise they could forge things which is very important obviously in the cryptocurrency world. I was under the impression that we could only use the stark types which are called transparent setup. They don't require that kind of thing. However, an issuer in our point of view could actually set up some of those things and we could use some of those techniques. So, for example, there's one technique that does what's called a commitment like we're committing to this vector of secrets. Okay. 00:25:00 Greg Bernstein: and that's part of the pseudonym thing. And that, takes a certain amount of work to prove that. And who does the work and how long does it take to verify can vary depending on the schemes. for example the length of the proof and say our case grows with how long the vector of secrets. So if we wanted to have a thousand used pseudonym so you could use it in a thousand different contexts. Greg Bernstein: not That's a thousand different contexts create with different verifiers. our proof would be proportional or exactly like a thousand scalar elements. So that's 32 bytes each scalar element. So it's 32k. Okay. And one of these other cases that involve these parameters that snarks use they're commitment schemes where that proof would be like a 100 bytes and the verifier only has to do what we call a couple cryptographic pairings. Greg Bernstein: a rather efficient type of thing. Okay, so that's what this paper gets at. It says, hey, in this setting, we could use some techniques that, we hadn't been thinking about and we can reduce the size of the proof and the verifiers work. In our case though, unfortunately, everything's got to compromise. nothing is perfect. The work for the holder actually increases, which we don't want to do in our case right now. the holder is probably going to be a mobile device or things like that. Greg Bernstein: and we may be wanting a larger number of uses than their paper is calling for. I mean, what we wanted was like if we get kind of a breakthrough and they're able to have a million uses, then it's worth it to go for this more complicated technique. That doesn't mean it's still not worth it, but this paper doesn't completely get us someplace where we kind of want to go for pseudonyms, but it does help open the door. Okay, so that it's a good paper to kind of review. They have interesting techniques. Unfortunately, it is a cryptographic research paper which makes it hard to get out all these kind of different messages. It's like, okay, how are they doing that? What are the disadvantages? Greg Bernstein: right now with verifiable credentials, if we had to bring in one of these snark rather than Stark like techniques, it's like getting a huge public key that we'd have to keep somewhere in addition to the issuers's regular public key. So some of these things plus the fact it didn't help the holder made me in discussions with the other folks on BBS pseudonym say okay this is some interesting stuff this could be some stuff for optimate help us maybe do an optimization in the future but it doesn't hand us a solution right now that wouldn't require a lot of new crypto Greg Bernstein: photography being standardized too cuz all those techniques I don't know 10 to 15 years old which not too old not too new but none of which are standardized KCG commitments another type of signature a technique for hardening a PRF which could prove very useful. but for the immediate concern of pseudonyms, we're not quite there. For the future though, and some of the stuff we've talked with Ying Tong about, there's a lot of interesting things here and could open up the things for the future. 00:30:00 Greg Bernstein: So, that's kind of a quick kind of little down in the mud view of it. Like I said, may open up some eyes about some things we could do, but nothing comes quite for free. any questions or that on BBS pseudonyms where we're going with them? Manu Manu Sporny: Yeah. Yeah, that was great, Greg. Thank you for doing that, in-depth analysis and the implementation. I guess the takeaway here is the vector of secrets that we were using before is still probably what we want to standardize because this doesn't necessarily give us any major advantages and it comes with the added burden of un unstandardized cryptography we'd have to go into an even larger kind of cryptography different cryptographic technique standardization thing in ITF Manu Sporny: and we really, don't have the time to do that in this current iteration. Were there any and I know that, the burden on the holder was increased. Do you know by what magnitude? I mean, it feels like everything else kind of stayed the same. Manu Sporny: The burden on the holder was increased, but by how much? I mean are we talking poor I'm and I don't know what the parameters are but for a vector of a thousand things it bumps the holder calculation up by multiple seconds five six seconds. Greg Bernstein: The way R ours is proportional to the length of the vector K. Manu Sporny: What? Mhm. Greg Bernstein: So that we've been calling it N or the L. First of all let's establish that the paper and the protein pseudonyms are the same and they say we're going to use a vector of secrets. That's the way to keep this everlasting anonymity. Okay. Greg Bernstein: What they do is they apply more advanced cryptographic techniques. And their problem is slightly more complicated. Now I said some constant times K. Theirs is K time log squared of K. So for a thousand what's log of a thousands I figure it's three or if you do it in base two it's 10 so log squar you take that squared that could either be 100 times worse or nine times worse depending on the constants involved so it's a factor of something it's not a square of it or something Greg Bernstein: But it's a log squared. I mean according to them. now they have another approach in here which is still more advanced than ours. it scales as K for the user. the verifier goes to nearly nothing. But we have this extra set of stuff to carry around known as this common reference string which grows as about four three to two to three times the length of the vector itself. Greg Bernstein: You have to have these constants. You have to throw around these that can't be generated any other way. So good news we all say you got to use a vector. Greg Bernstein: Go ahead mono. Yeah. Plus,… Manu Sporny: The CRS ends up becoming for a vector of a thousand,… Manu Sporny: you're saying it ends up becoming four times the vector of a thousand 4 kilobytes plus. Greg Bernstein: I think Yeah,… Manu Sporny: So, we're talking about adding Greg Bernstein: because it's actually group elements and… 00:35:00 Manu Sporny: Mhm. Yeah. Greg Bernstein: those are 48 bytes each. So, you'd have 2,000 group elements. So, yeah, you're getting up to 50 to 100K. Greg Bernstein: And I was going, if this was really really great, but what I'm would like one is that shifts the burden off the holder prover a bit, I mean, it's great. I mean, it made it really nice for the verifier because, I tried this commitment scheme that was based and, that worked the way you said. It's like,… Greg Bernstein: yeah, made it a little worse for the holder. but yeah, the verifier was real easy or… Manu Sporny: Yeah, there's a significant speed speed size tradeoff I guess is… Manu Sporny: what we're saying here is and… Greg Bernstein: or who you put the work on too. Greg Bernstein: And that's… Manu Sporny: Yeah. Mhm. Greg Bernstein: where there's all this work happening, in the ZKP community, Before they really were all pushing to reduce verifier, but Leo and some of those schemes sacrificed a little bit of size for making it easier on the prover. So these are nifty things to keep an eye on. And this is some of that stuff Ying Tong was with these things called polomial commitments. Manu Sporny: Mhm. Okay. Greg Bernstein: However, none of those are standardized yet. And so if we want something sooner than later for BBS, just going with, a straightforward vector commitment based on what are known as Person commitments. Greg Bernstein: This is the kind of stuff that goes into blind BBS and underneath even in BBS signatures. it has the same amount of work for the prover, or a little less. And yes, the verifier is going to have about that amount of work. Yeah. Manu Sporny: So it sounds like then back to the previous vector of secrets better approach which sounds good. and then the analysis that we still haven't done is what is an acceptable level for that vector size based on your threat model we still need to do that. I think Wes Smith might take a shot at doing that. we definitely need to at least document that in the BBS specification so people know… Manu Sporny: what to set that vector to if they're concerned around, everlasting anonymity Greg Bernstein: Yeah, we definitely need to eluc set there's something to do in the BBS CFRG pseudonym spec to discuss the issues and… Greg Bernstein: then try and… Manu Sporny: Yep. Okay. Greg Bernstein: take that the next step in the rifble credential the base cryptographic spec we can say it depends on how many colluding these guys and this but it would be nice at the verifiable credential level to try and be a little more helpful right I mean as a raw cryptographic tool we can say make your choice based on these ideas. Greg Bernstein: It'd be great if we can do more at the VC level in the spec. Manu Sporny: Yeah. Yeah. Greg Bernstein: If we can Yep. Manu Sporny: Yeah. Yeah. Yeah. Yeah. No. I'm wondering, what we would say in each spec. but, that's a PR that can be worked out later. Greg Bernstein: That's yeah that we had had discussion with the BBS folks because there's only one or so issue left to be thought up through on the calculation procedure. Greg Bernstein: Otherwise we're ready to do PRs and update the CFRG pseudonym spec and… Greg Bernstein: get that in shapes then roll out the changes. Manu Sporny: And… Manu Sporny: then for timeline, Greg, you said that, CFRG, reviewers, the chairs jumped in and… Greg Bernstein: The chairs are asking the reviewers to … Manu Sporny: asked for. Greg Bernstein: can you please look at the comments that were used to adjust, the base BBS spec and… Manu Sporny: Mhm. Right. Greg Bernstein: what the original authors of BBS did to fix that or accommodate explain blah blah blah. 00:40:00 Manu Sporny: But the I guess this review hasn't started on pseudonyms or… Manu Sporny: or blind yet. Is that right? Okay. Yeah. Greg Bernstein: No, no,… Greg Bernstein: we have not asked because … Manu Sporny: It needs to be in a shape that Okay. Greg Bernstein: yeah, we're going to update pseudonyms. Blinds has not changed. I don't think anybody has a problem with it,… Greg Bernstein: but I think we need to double check if we have all the security and privacy sections written. I can't remember if we did or not, but Yep. Manu Sporny: And… Manu Sporny: and then I think we'll probably have to do another pass on getting Anna and Sorrow and those folks to write into the CFRG to say that they've looked at the spec and that sort of thing. So, I think we'll have to just collect all the previous names that wrote in and… Manu Sporny: and have them do another set of input for blind and pseudonyms. Greg Bernstein: Sounds good. Manu Sporny: All anything else on that, Greg, before we move on? Greg Bernstein: We're getting there. And there's cool techniques to look for the future,… Greg Bernstein: but there's also the practicalities that let's get this current pseudonym going out the door. Manu Sporny: Yep. Manu Sporny: Thank you for continuing to work on that, g. Much appreciated always. all with that, let's move on to our last item. Jesse, I tried to write some of these down. you had asked a number of kind of questions in the Slack channel. I just kind of put them here. do you want to start here or do you have any background that you'd want to provide before we jump into the questions? Jesse Wright: Yeah, I can give context which is Kristoff Braw who wrote the other paper that we talked about two weeks ago and I are collaborating with the Berkeley cryptography team and we're doing a circuit based implementation for query over credentials. There are a number of design decisions that need to be made there which will affect the trade-off between performance and how quantum safe the proof is in terms of whether you can forge the proof in terms of whether you can disclose knowledge from the proof. Jesse Wright: So I would like to get an understanding of the priorities for production so that we can design towards those priorities. Manu Sporny: Okay, sounds great. so I can provide an opinion on these. I'd love to hear other people's opinions. I'll try to give some backing reasoning. Manu Sporny: So, one of the things is max input size and I think the input you're measuring as a triple or a quad coming in. Is that typically identification cards have 35 fields on Driver's license typically has 35 triples associated with it minus the type information. let's say it's 40. Jesse Wright: Yeah. Heat. Manu Sporny: one of the things we can look for are regular processes that require you to provide a set of documents to get another document. So again looking at kind of a driver's license you have to provide two to three other documents to prove your identity. so that holds for things like getting a passport, getting a driver's license, opening a bank account. the general thing is two to three forms of other ID utility bill and statement educational bank account information birth certificate things like that. Manu Sporny: And each one of those documents could have roughly anywhere from 10 to 40 triples or quads. that said the other thing to keep in mind is in the future we do want to massively reduce the amount of data each one of those credentials has. over time and I'm talking 10 to 20 years we would expect credentials to contain less and less PII and more and more general statements like the person is over the age of 21, not the person's birthday is this exact date. but for now I'd say let's say we're talking hundreds of triples at most in a full calculation. 00:45:00 Manu Sporny: So we're talking about between 10 20 to 40 triples per verifiable credential and… Manu Sporny: you're providing between three to four at most verifiable credentials in this proof. yeah I think so I think that's probably a acceptable approximation to start with. Jesse Wright: So at most even it's about 150 160. Jesse Wright: Okay. Awesome. Manu Sporny: The only other the place where you might not see that are things like supply chain credentials which might be like a shipping manifest which could have thousands of entries on it in 2,000 3,000 triples but I don't think that's a primary use case yet and they may issue much reduced credentials for those use cases. Manu Sporny: One other thing, and, Greg, this is a bit of a sidebar, but one other thing that we might consider is when doing these BBS signatures or… Greg Bernstein: Yep. Manu Sporny: BBS credentials, that the issuer would actually sign a much smaller subset of the credential, right? So, let's say it's a shipping manifest with 2,000 triples in it, but the BBS signature only covers maybe 20 of them, And so if you're going to use BBS and the reason for this is to optimize the signature size in BBS so that you're not having to hide 2,000 pieces of information and… Greg Bernstein: Mhm. Manu Sporny: thus growing the size of the signature by 200 because of all of the commitments you have to include. So, this is just a random thought, Greg, but we might want to think about that. for example, the initial base proof only includes 20 of the 2,000 triples in BBS because the issuer has decided that… Manu Sporny: if you're going to do a BBS disclosure of this, really, you're only going to be disclosing these 20 fields. and for example,… Greg Bernstein: Yeah, that's yeah and… Manu Sporny: every single item in the shipment is not something you're going to want to selectively disclose. Greg Bernstein: we have time to put in some of those recommend those discuss that in the section of the BBS VC spec. Manu Sporny: Yeah, that's right. the only thing we would need is a list of the JSON pointer paths or whatever that were actually signed over in the base so going back to Jesse, your question is like, we may be able to tune this in time by basically saying, yeah, sure,… Manu Sporny: there are thousands of triples here that could be mixed together, but really we're minimizing it down to 50 or 100 so that we never hit the problem of having thousands of trip having to deal with thousands of triples, if that makes sense. Jesse Wright: Yeah. And in that line of… Jesse Wright: since we're talking about signatures, the inclination in this work is to use new signature mechanisms where so that there's a way of signing a vector of triples such that you can extract a subset of that vector to pull into the zero knowledge circuit. Manu Sporny: Yeah, exactly. Jesse Wright: I don't know that I'm not kind of specializing on this bit. that is not a commitment scheme as far as I understand or BBS plus. Is it just totally out of the question to be going and inventing new signature schemes for this piece of work as a proof of concept? Jesse Wright: I.e. they'll never be accepted or just Yes. Manu Sporny: It depends… Manu Sporny: what you mean by signature scheme and invent So if there is math that explains what's being done very clearly and it's very broadly accepted then technically you're not inventing anything new it already exists. Manu Sporny: and as far as signature schemes, what we're talking about here are really data transformations into a useful form where we can potentially just use an existing signature scheme like ECDSA whatever. So I think what we're largely talking about is transformation mechanisms in those we can invent any new transformation mechanism right and then we transform the data into a form where then the transform data as input into a cryptographic circuit or… 00:50:00 Manu Sporny: a digital signature mechanism becomes far more efficient. go ahead Greg you've got your hand up. Greg Bernstein: Yeah, one of the prime examples of that Mano is talking about is we have two different selective disclosure schemes. Greg Bernstein: ones based on traditional ECDSA and in that case we're using ECDSA per quad per triple and actually for the credential itself come up with an ephemeral key and we sign each of the triples separately and those all become a set of signatures versus BBS where it's one signature Greg Bernstein: over the whole thing and we do selective disclosure that way. do you need a new signature scheme or… Greg Bernstein: do you want to apply it in a different way and that's what the flexibility we have with credentials is change the way you transform it but don't try and invent brand new cryptography use that cryptography in a different way. Jesse Wright: Yeah. … Jesse Wright: I've put a link to the paper that we were looking at because it's better than me trying to explain it. and it's look up arguments that and actually is this your lab manu or no different matter. so I've not read This is the Berkeley folks that have come to us and proposed this paper. I think it is in the data transformation camp. Jesse Wright: I would need to confirm. Okay. Manu Sporny: So to be clear creating an actual digital signal al algorithm is completely off the table like we do not want to do that that is a decade of work right to get it into production. so we want to reuse existing I mean BBS is a stretch as it is right and it's been around for 20 years so we definitely don't want to go down that path. What we want to do in arguably this new ZK stuff, the Lihero stuff is in that class of new cryptography. the only reason we're riding that wave is… Jesse Wright: Nothing. Manu Sporny: because Google is really excited about it. And when people hear that Google's working on it, they think it's going to go much faster than it typically goes, which it could maybe, right? Manu Sporny: and a lot of the cryptography community is really excited about the na ZK Stark based approaches and so there's a lot of desire to push that stuff forward faster than the other raphic circuit stuff is bleeding edge new cryptography but it has this benefit of we might be able to use really relatively ancient cryptography techn technology ECDSA but in a new way that allows us to use these very deployed infrastructure for doing ECDSA signatures HSA hardware security modules and things like that but use them in a way that provides Manu Sporny: unlinkability, which is what the Lihiero, ZK stuff is So, hopefully that answers your question. it's new signature schemes we really want to stay away from. But if we can transform input data into a form that makes it efficient for these Snark, ZK Starks or, traditional crypto, we can completely do that, because that's way easier to prove that it is much easier to prove a transformation algorithm than it is arbitrary new cryptography. Manu Sporny: Did that answer your question? Jesse Wright: Yeah, got it. Jesse Wright: Yes. Thank you. Manu Sporny: All right. Greg Bernstein: But those transformation methods can be quite elaborate and… Greg Bernstein: wellestablished nobody's going to complain about a Merkel tree these other hashing schemes or… Greg Bernstein: how you and most of these commitment scheme things have pretty good roots to them too. So there's a lot that can be done. 00:55:00 Manu Sporny: All right. Manu Sporny: Any other opinions on max input size? can anyone else think of a use case where we might be going above a hundreds hundreds of triples or… Manu Sporny: All right. second question, query complexity. How many joins do we expect? By joins I'm presuming you mean we have multiple different types of verifiable credentials and we're trying to join the data in each one of those to together Jesse Wright: No, I mean in a sparkle context. Jesse Wright: text each if you're doing pattern matching. So if you're getting all triples that match the pattern s name o let's go s a person and… Jesse Wright: then s age x that would be a join of two triples Manu Sporny: Got it. Manu Sporny: I would imagine those stay below 10. but that's just kind of a gut reaction cuz usually when you're trying to do I mean these verifiable credentials are usually used to determine authorization access to a particular resource or social benefit or something of that nature right so if we look at something like age verification it's probably between 1 to three the age thing is the core one but you might also want to make sure that it's a person you're talking about or they have some one or two other qualities that you're looking for. Manu Sporny: So for age, let's say it's three at most joins. but for some other use cases there have been talks of being authorized for credit or a loan or something like that where you would provide much more information are you currently employed? what's your salary in some kind of large bucketed range what tax bracket are you in things like that and those types of joins could be upwards of 10 to maybe at most 20 I'd imagine because you share enough of that and you're pretty strongly identified right your dem Manu Sporny: demographic information narrows you vastly down from,… Manu Sporny: 100 million people down to one in 100,000, right? any other thoughts from folks like… Jesse Wright: Yeah, brilliant. Jesse Wright: The circuit. Manu Sporny: how many attributes are we looking for in these queries? can you give us an idea of how w wildly these kind of change do the joins result in exponential growth in computation or Jesse Wright: So that you design the circuit for a given max size of joins because as far the circuit is … Jesse Wright: what is it an set of arithmetic constraints that you're proving to be true. Manu Sporny: Mhm. Jesse Wright: So in designing the circuit there is an upper bound that you are placing on the number of joins that you can do that may not like you could generate new circuits because generally the query that you're asking is knowledge to both parties. So there is a world in which you generate new circuits when you hit a use case where you've got more triples and that's not really an issue. Manu Sporny: H. Mhm. Jesse Wright: But for now what is the max bound that we want to put in place for the circuits that we're building. I don't have an idea of the computational cost that comes with making those larger at the moment. Manu Sporny: Okay, that's fine. That makes sense. my gut feel on this is that the more the larger the join set the more damaging the privacy the query ends up becoming to the point where it's like why are you doing this in an unlinkable way to a degree right me meaning if the join set is pulling 20 attributes 01:00:00 Manu Sporny: on the individual. sure, cryptographically it might be unlinkable,… Manu Sporny: but statistically, you're looking at some fairly strongly identifying set of attributes. Jesse Wright: That depends though,… Jesse Wright: so it might be that you're pulling attributes that are known to be attributes on everyone in a population. We know that everyone has parents. We know that everyone has an age. We know all of these things exist as attributes on individuals. Jesse Wright: And then for narrowing down to if you're eligible for a clinical study, you need to have some weird combination of these attributes that holds that. Manu Sporny: Mhm. Absolutely. Jesse Wright: I'm kind of b**** here, but at least if at some of these attributes hold query rather than all of these attributes hold query,… Manu Sporny: Yeah. Yeah. Jesse Wright: that might be Manu Sporny: Yeah. Yeah. Yeah. Yes. I completely agreed. it's highly dependent on the use case and the attributes coming in from the credential and all that stuff. Manu Sporny: The general thing we tend to get concerned about is there are going to be people that use this technology that do not have the training to understand how correlatable certain claims and attributes and things are and they will inevitably put for example a date of birth and a zip code into one of these unlinkable queries and… Jesse Wright: Brilliant. Manu Sporny: that you as we know is highly correlating, data. Even though it seems fairly benign, it's not. And, you can get down to a handful of 10 individuals in a population of 150,000 by just doing zip code and Birth date. okay. Manu Sporny: So I think we're expecting certainly no more than 20 but on average I would imagine we're looking at five to 10 10 at most but I'd say averaging five. what provenence is required in production? yeah. Jesse Wright: Before that one, there was a second part to the join question… Manu Sporny: String comparison. Jesse Wright: which is what on filters and other functional operations, what data types and what are we needing to support and what operations are we needing to support? So in states whatever Yeah. Manu Sporny: So you mean like string and string comparison? Manu Sporny: I think yeah ideally range proofs if we can get there but those tend to make things fairly complex which means integer range calculations. I think dates and datetimes fit into integer range calculations. So we could probably coersse date times. We'd need to know that it's a date time. but then for example, no it's a day time but convert it to milliseconds seconds since the epoch or something like that. what else? I mean classes but that's probably just a sub type of direct string comparison. Manu Sporny: me meaning it would be good to know that this is a person or… Manu Sporny: a vehicle or something like that. But I think that just is a subtype of string comparison Jesse Wright: that wouldn't even be a data type comparison… Jesse Wright: So if you want to know that there is something that is a subclass of vehicle in your data set, you disclose in the inputs what a vehicle is as a string and then in the circuit show that there is a thing that is a sub like you don't need to do string comparisons of those. Jesse Wright: You can reveal what the terms are and then prove properties hold about those terms. Manu Sporny: Got it. Manu Sporny: Okay. Yep. Jesse Wright: Poorly explained, but I'm glad you got it. 01:05:00 Manu Sporny: Yep. Yep. No, that makes sense. I have completely lost track of the time. My apologies. We're way over. Jesse, we can either cover these other ones the next time we meet or I'll try to shoot out some quick, responses to you today on these, and… Manu Sporny: then we can discuss next. Jesse Wright: If you can shoot out responses,… Jesse Wright: that would be useful. Manu Sporny: All right. we'll do. Jesse Wright: Thank you. Manu Sporny: We're at time. Jesse. Really appreciate you looking into this as thanks everyone for their time today. Sorry I went way over. but appreciate have a wonderful weekend and we will meet again next week. Thanks all. Bye. Jesse Wright: Have a good weekend everyone. Hey Meeting ended after 01:05:50 👋 *This editable transcript was computer generated and might contain errors. People can also change the text after it was created.*
Received on Friday, 20 June 2025 22:04:33 UTC