Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:00
A quick note before we
0:02
begin today, due to
0:04
some equipment failure, the
0:07
first 15 minutes of
0:09
this episode sound worse
0:11
than you would normally
0:14
expect from our show.
0:16
We think the episode's great
0:18
and still worth bringing to
0:21
you and it gets better
0:23
later, but as you'll hear,
0:26
it's not your headphones,
0:28
it's us. Yeah, so we last talked
0:30
probably over a year ago at
0:32
this point, which is pretty heavily
0:34
focused on what you guys were
0:37
building. We'll get into that a
0:39
little bit later, but the main
0:41
reason I wanted to have you on
0:43
is it feels like we're sort of
0:45
at a turning point with zero
0:47
knowledge systems. There was a huge
0:49
amount of hype from 2018 kind of
0:52
through 2022. It feels like we got
0:54
a little quieter in the last few
0:56
years, but a lot of stuff's actually
0:59
launched and usable. And on Solana, Zero
1:01
Knowledge has actually had a little bit
1:03
of a resurgence. It's actually become a
1:05
chain that has a ton of different
1:08
ZK primitives, a lot of different companies
1:10
building with zero knowledge technology, whereas from
1:12
state compression to privacy layers, to confidential
1:14
computing, all sorts of other types of
1:17
other types of applications. But at the same
1:19
time, a lot of the big name zero
1:21
knowledge players that big on the technology,
1:23
that big on L2 is that we're
1:25
going to be using it, have really
1:28
struggled to sort of find differentiation and
1:30
a reason for people to sort of
1:32
care about what the technology is built
1:34
on how it actually works. So it's
1:36
a whole bunch of stuff I want
1:39
to talk through today, but sort of just
1:41
start off, like, give us a run through
1:43
of like, at the end of the day, like...
1:45
How does zero knowledge technology
1:48
actually work in private tests?
1:50
Yeah, sure. So there's essentially,
1:52
you can think of zero knowledge
1:54
technology as this big term with
1:57
different kinds of technologies. and there's
1:59
zero knowledge proofs, which is the
2:01
one big traditional way you're mostly
2:03
referring to for zero knowledge technology,
2:05
but there's also a secure multi-party
2:08
computation and fully homomorphic encryption. Those
2:10
three basically form what we think
2:12
of as zero knowledge technology. Now
2:14
the security and different properties that
2:17
zero knowledge technology enables come from
2:19
cryptography. at the end of the
2:21
day, right? So there's a lot
2:23
of math, a lot of cool
2:25
math, I have to say, and
2:28
cryptography involved, but the history of
2:30
CK really is that in the
2:32
80s, the first kind of zero
2:34
knowledge proofs have been designed, and
2:36
back in the day those were
2:39
interactive proving protocols. So a zero
2:41
knowledge proof usually involves two parties.
2:43
Me and an example, the prober
2:45
and you, Austin, the verifier, and
2:47
Professor knowledge proof, I'm able to
2:50
convince you of the truth of
2:52
some statement. without having to share
2:54
the statement itself, simply just communicating
2:56
the validity and the truth of
2:59
some expression. And back in the
3:01
day in the 80s, those were
3:03
interactive protocols, so we had to
3:05
perform some complex computation with each
3:07
other to generate a proof. Over
3:10
time, those have evolved into non-interactive
3:12
proving skills. So, 100% makes sense.
3:14
Sort of the hand or butt,
3:16
right? We sort of have systems
3:18
that do some version of this
3:21
today, right? I can give you
3:23
a installer from a closed source
3:25
software provider and it includes a
3:27
certificate that I can verify that
3:29
no one's monkeyed with the code
3:32
without actually looking at the code
3:34
myself because I see this sort
3:36
of certificate that's said, yes, the
3:38
check sum matches or you can
3:41
publish a check some somewhere and
3:43
and I don't actually have to
3:45
see the source code to verify.
3:47
that something is what it says
3:49
it is we have wallets that
3:52
have public private key encryption where
3:54
i can verify that you have
3:56
signed a transaction without having your
3:58
private key so how is this
4:00
similar work different from a lot
4:03
of these encryption methods were familiar
4:05
with today that are used to
4:07
provide some amount of verifiability. Yes,
4:09
so I think, especially in the
4:11
context of what you were referring
4:14
to with some science certificate
4:16
for some software provided by
4:19
a vendor, that is really
4:21
a case of some trusted interaction
4:23
because at the end of the
4:26
day, your source code is supplied
4:28
by some entity. right? And so you
4:30
have this remaining trust of okay I
4:32
trust this entity and I know that
4:34
they've sent it to me so it's
4:36
sort of like an encrypted channel if
4:38
you will right because they've signed this
4:40
the software you can trust to install
4:42
it software because you have a contract
4:44
with their company right and initially were
4:46
to exploit you then you could sue
4:48
them in court. So I think what's
4:50
key and that's something that we're trying
4:52
to achieve with block trains in general
4:54
with Solana with what beer building but
4:57
any kind of distributed trust
4:59
the systems is removing single points
5:01
of failure, removing trusted entities. So
5:03
verifiable compute is an important aspect
5:05
for that. And that's one of
5:07
the properties that zero knowledge proofs
5:09
and other cryptography schemes can provide.
5:12
So the term zero knowledge proof
5:14
can be a bit misleading because
5:16
a lot of usage of zero
5:18
knowledge proof actually circles around scalability
5:20
and verifiable compute mostly. But there's
5:22
different kinds of properties for a
5:24
zero knowledge proof. One is the
5:27
sukendness of the expression. So that
5:29
basically expresses that there is
5:31
some asymmetry. There's someone generating
5:33
the zero knowledge proof, the
5:35
prover and someone verifying it.
5:37
And there's a computational and
5:40
informational asymmetry between those two
5:42
players. So you can express
5:44
a very complex computation in
5:46
a zero-notish proof, which in
5:48
the case of CK-snarks, as
5:50
this constant, always the same
5:53
complexity. of evaluating and verifying
5:55
the serial solution. And that's
5:57
where scalability comes to play. That's where
5:59
CK came. used as a scaling solution
6:01
because at the end of the
6:04
day can outsource computation to some
6:06
third party and simply verified. Yeah,
6:08
so I don't want to get
6:10
too familiar with the math because
6:12
you're going to blow past me
6:14
pretty quickly. But the concept of
6:17
I am trying to verify something
6:19
without knowing its contents, right? In
6:21
the traditional world, the only way
6:23
that's done is the third party
6:25
vouchers for it. So KYC is
6:28
a great example, right. My accountant.
6:30
vouches that I am an accredited
6:32
investor, I don't have to actually
6:34
show you my bank statements, you
6:36
trust him because of his accreditation.
6:38
How are you able to, like
6:41
at a high level, prove the
6:43
contents of something without revealing what
6:45
that is? So I think there's
6:47
something that can blow your mind
6:49
even more, right? So still knowledge
6:51
rules, just generating this proof and
6:54
submitting it to a third party.
6:56
I would have to sort of
6:58
go into the weeds of it,
7:00
but I think as a concept
7:02
it's easier to understand. So I
7:05
have some secret that I can
7:07
do some mathematical permutations, right? And
7:09
submit a proof of the operations
7:11
that I have performed, right? So
7:13
you know the algorithm that I've
7:15
executed. I know the algorithm that
7:18
I've executed. So we both have
7:20
to share knowledge of the function
7:22
that should be executed, the kind
7:24
of permutation, right? The more... Yeah,
7:26
disturbing concept if we will is
7:29
what we are actually doing. And
7:31
that involves not just me having
7:33
secrets and running something, but instead
7:35
allowing a network of nodes to
7:37
run something over secrets. The network
7:39
doesn't know those secrets. And they
7:42
can generate verifiable output, right? Which
7:44
is even more crazy concept if
7:46
you think about it, right? So
7:48
I think at the end of
7:50
the day, this technology... enables use
7:53
cases that we haven't seen before.
7:55
All of those three protocols enable
7:57
new use cases in a distributed
7:59
set. that make those
8:01
systems way more resilient and
8:03
performant at the same time. Yeah,
8:06
so I want to talk about
8:08
some of the dynamics you
8:10
mentioned around compute face symmetry. So
8:12
most encryption methods we use
8:14
nowadays, most security methods that we
8:16
use nowadays are effectively free
8:18
to verify. Right, signature
8:20
verifications on a Solana
8:23
transaction is basically a short
8:25
of a time as you can get it to be,
8:27
you know, me validating a certificate is valid
8:29
for a website or any other type of
8:31
encryption. It's not a
8:33
computationally intensive task like encrypt my
8:35
messages. So what is it about zero
8:37
knowledge systems that, you know, if
8:39
you talk to folks about zero knowledge
8:42
systems, it feels a little bit
8:44
like nuclear fusion. Like, oh, it's always
8:46
just a few years away from
8:48
mass applicability. So what is it actually
8:50
that requires what not from a math
8:52
side, but what components of the
8:55
system require compute? And why has that
8:57
been such a barrier historically? So
8:59
one of the key problems
9:02
really is that you're
9:04
operating on a completely different
9:06
computer. If you run your
9:08
normal signature verification, those
9:10
algorithms usually are optimized for
9:12
some binary operations that
9:14
your computer can perform. Right.
9:16
And those things are
9:18
realized. There's hardware acceleration. Exactly.
9:20
Yeah. And then you're
9:22
basically operating over numbers and
9:25
bits with zero knowledge proofs
9:27
and all of that cryptography,
9:29
where you want to represent
9:31
some computation that a normal computer can
9:33
operate in a cryptographic way. You
9:35
need to transmute that into
9:37
a different representation and that
9:40
representation involves so -called elliptic curves,
9:42
different functions and finite fields.
9:44
And that makes things much
9:46
more complex because computations now
9:48
are being represented as so
9:51
-called circuits, arithmetic circuits, for
9:53
example, of which you can
9:55
think similarly to fuses in an
9:57
actual chip, right? Yeah. Actual
9:59
circuits. signals. And it's not those
10:01
signals you know from a computer,
10:04
but instead it's finite field signals,
10:06
if you will, that run in
10:08
those fuses. So there's complexities involved
10:10
in representing operations that on a
10:12
normal hardware are efficient in this
10:14
virtual circuit format. So that's really
10:17
one of those problems, where there's
10:19
significant overhead associated. And there's hardware
10:21
acceleration for those things as well,
10:23
but it takes time like that.
10:25
Yeah, I guess that was sort
10:28
of, the scale problem feels like
10:30
it's, it should be more solvable
10:32
than it seems to be. Like,
10:34
between FPGAs and between like the
10:36
way, like, you look at something
10:39
like ACIC acceleration and Bitcoin, and
10:41
that sort of revolutionized what was
10:43
possible in terms of hat-tray there.
10:45
It's not like the Bitcoin folks
10:47
had crazy super corporations organizing venture
10:49
rounds to build ACICS and some
10:52
that did eventually. But with the
10:54
amount of money that's gone into
10:56
ZK research over the years and
10:58
the amount of money that's gone
11:00
into funding, ZK crypto companies, like,
11:03
why is it today that we
11:05
either don't have hardware that's capable
11:07
of doing these things? Like, how
11:09
much of the breakdown is like
11:11
research versus like no one's actually
11:14
just built it yet? That's the
11:16
exact question. I think it goes
11:18
into the direction of why we
11:20
at RQ initially, we're working with
11:22
zero knowledge group. back in the
11:24
day we were called elusive and
11:27
we were doing on-chain privacy with
11:29
zero knowledge proofs but eventually we
11:31
pivoted and I would rather call
11:33
it evolved into using secure multi-party
11:35
computation instead because for us it
11:38
was some limitation that could not
11:40
be overcome yes associated with CK
11:42
and what what we've noticed I
11:44
think is that there's quite a
11:46
lot of teams that also came
11:49
across those limitations and especially on
11:51
the confidentiality and an encryption front
11:53
does just this big limitation of
11:55
you just being able to provide
11:57
privacy for one party that interacts.
12:00
in this approval verifier system, but you're
12:02
not able to build smart contract
12:04
systems, you're not able to aggregate
12:06
data. And so that's one of
12:09
those big limitations that we came
12:11
across. And that's why we decided
12:13
to go down a different cryptographic
12:16
road. Yeah. So talk to me a little bit
12:18
about how that new system works and
12:20
how it differs from some of the
12:23
other zero knowledge technologies we've been
12:25
talking about. So MPZ, as similarly
12:27
to CKK, has actually... they have
12:29
evolved in a parallel fashion basically
12:31
also starting in the 80s. I
12:33
think the fascinating thing about MPC
12:35
really is that it allows a
12:38
set of participants to jointly compute
12:40
a function without having to share
12:42
the individual inputs. And MPC in
12:44
our space mostly gets associated with
12:46
some designer wallet management, right? Right,
12:49
right. I'm breaking my wallet into
12:51
a few pieces and then multiple
12:53
compute puts it together to sign
12:56
a transaction. But MPC really. at
12:58
its core allows a set of
13:00
nodes to jointly run a predefined
13:02
function an algorithm and have individual
13:05
inputs and keep those inputs private
13:08
by producing a shared output.
13:10
And what we're essentially making use
13:12
of also with the research that
13:14
we're doing is with new trust
13:17
models that yeah, have become possible
13:19
with those kind of systems. Because
13:21
initially it was really only honest
13:23
majority which I guess you're familiar
13:26
with, right, but for the folks
13:28
listening. Most Byzantine fault tolerance, practical
13:30
business tolerance systems have this honest
13:33
majority as trust assumption, right? So
13:35
you have some qualified majority that
13:37
has to behave honestly, has to
13:40
behave according to the protocol. And
13:42
yeah, you've seen a breakthrough in
13:44
those kinds of protocols that are
13:46
now allows for dishonest majority,
13:48
which basically inverts it. Only
13:50
one single honest participant required
13:52
to guarantee different properties for
13:54
those computations. So I think
13:56
it's worth diving a little
13:58
bit more into. how that works
14:00
because the blockchain is always and
14:03
most of these systems have always
14:05
had like an objective truth problem
14:07
right where it's like unless you're
14:09
like fairly Catholic like objective truth
14:12
is pretty unknowable right in the
14:14
sense of like you need an
14:16
external reference point to base something
14:18
as being true which is where
14:20
the honest majority assumption comes from
14:22
that you can only like it's the
14:24
whole thing about like you know
14:27
Ethereum classic is actually the real
14:29
Ethereum. Right, and like the thing
14:31
we all call Ethereum is actually
14:33
a fork, right, of Ethereum, but
14:35
we all have socially decided that
14:37
this is the main one, the
14:39
other one's classic. But like, from
14:41
the computer's perspective, there's no way to
14:43
know which one is quote unquote real
14:45
Ethereum. Yeah, I think that's actually
14:47
an excellent question and most people
14:50
don't don't ask me that question.
14:52
So it's, I think even, even
14:54
the smart question to ask, right,
14:57
because how could... dishonest majority function,
14:59
right, as a concept, because how
15:01
could... Right. Just this one person
15:03
be correct, right, if there's
15:05
only one honest participant. So
15:08
that's also where the connection
15:10
between Arquium and Solana comes
15:12
to play action, because what
15:14
we're saying essentially is Arquium
15:16
is a purely computational network
15:18
that supplies computing capabilities
15:21
that Solana itself cannot
15:23
facilitate. confidential computing,
15:25
generalized confidential computing basically. And
15:27
so we make use of
15:30
this connection with Solana. So
15:32
we have a program on
15:34
Solana that functions basically as
15:37
entry point as mempool for
15:39
computations happening within algorithm.
15:41
And so Solana sort of
15:43
becomes this true player, if we will,
15:46
where there's just the consensus
15:48
over what computation should be
15:50
performed. And so if... Arquium
15:52
with an Arquium, a computational
15:54
cluster, doesn't run a computation
15:56
as it's tasked to do
15:59
on Solana. the corresponding party within
16:01
the cluster, it could be all
16:03
of those nodes or just one
16:05
or zero nodes, right? If anything
16:07
goes wrong, anything doesn't go according
16:09
to protocol, they get punished for
16:11
that. So that's really where
16:13
this connection comes into play
16:15
and where actually blockchain
16:17
forms a system that can
16:19
overcome limitations from very
16:21
theoretical papers. Because now with
16:23
Solana, it becomes possible
16:25
to overcome problems that have
16:27
always existed within those
16:30
papers. Yeah, that's fascinating because
16:32
there's this like trend right
16:34
now, which is like, oh, anything that
16:36
is built on top of something
16:38
should probably become its own L1 or
16:40
at least think about it, right?
16:42
And that usually all you're getting
16:44
is some form of like economic
16:46
security or network effect or something
16:49
along those lines. But this is
16:51
like a real case where the
16:53
security model would be worse if
16:55
it didn't have a blockchain underneath
16:57
it. Exactly. Yeah. And it wouldn't
16:59
make any logical sense to then
17:01
also extend it into an L2
17:04
infrastructure. I really feel like this
17:06
setup will be the future to
17:08
have this big consensus mechanism and
17:10
then have Archeom with
17:12
this honest majority which can
17:14
perform specific kind of executions
17:16
in an asynchronous way but
17:19
doesn't have to find consensus
17:21
overstate. Yeah, that's
17:23
really interesting. So
17:26
from a
17:28
practical tools perspective, what does
17:30
this allow folks to do
17:32
either on or off blockchain
17:34
that they really can't do
17:36
today? So without using
17:38
the term healthcare records.
17:40
Yeah. Yeah.
17:43
So it's funny
17:45
because I think breakpoint
17:47
has thrown me and
17:49
my entire team that
17:51
basically everyone craves some
17:53
form of confidential computing
17:56
within their applications. Yeah. The
18:00
big problems why all of those CK
18:02
infrastructures haven't so far succeeded is that
18:04
they force developers to learn some new
18:07
paradigms, learn some new technology, actually need
18:09
to learn what a circuit is and
18:11
all those kinds of things. We for
18:13
example abstract all of that away from
18:16
developers. So we just say you want
18:18
confidentiality as a developer. So all you
18:20
have to do is you. mark the
18:22
functions and the state within your program
18:25
that should be confidential as confidential. And
18:27
then you let Arquium process that and
18:29
settle it back for you, right? So
18:31
that's what we're trying to achieve. And
18:34
so what folks can have is they
18:36
can have any arbitrary rust program essentially
18:38
be executed in a confidential way. And
18:40
so there's very tangible use cases. I
18:43
think one straightforward use case and has
18:45
been amazing to see, yeah, folks on
18:47
Solana start building that with us in
18:49
our private test net, private markets, right?
18:52
So dark pools, for example, since 30
18:54
to 40% of daily US spot volume
18:56
happens in dark pools, right? And so
18:58
there needs to be an answer for
19:01
that on the blockchain as well. A
19:03
lot of folks I talk with from
19:05
the threat file space actually often say,
19:08
yeah, you know, JP Morgan is actually
19:10
running a private ledger, right? Yes. But
19:12
they can't move to some public ledger
19:14
because of confidentiality assurance. So I think
19:17
that will be a significant step also
19:19
in regards to stable coin payments to
19:21
further enhance their adoption. And that's something
19:23
barrier. closely working together with core folks
19:26
in the Solana space because Solana has
19:28
the confidential talking program which yeah it's
19:30
only for transfers you can't run a
19:32
market on it exactly has some limitations
19:35
right and so by extending that to
19:37
allow smart contracts programs to have confidential
19:39
balances have confidential interactions between parties that
19:41
makes it so powerful and that will
19:44
allow some confidential lending protocol right and
19:46
will allow for some form of a
19:48
unified liquidity for all
19:50
those folks to
19:53
build applications with that,
19:55
but I think
19:57
DeFi is only the
19:59
first part. We've
20:02
seen a lot of
20:04
folks in Deepin
20:06
as well, be very
20:08
interested in this
20:11
kind of technology, so
20:13
that ranges from
20:15
people just wanting to
20:17
replace some trusted server infrastructure
20:20
with something trustless that
20:22
is being executed on the
20:24
blockchain. So, for example,
20:26
have some proprietary algorithm that
20:28
does some action
20:30
within a Deepin. Have
20:32
that be executed now in
20:34
a verifiable and trustless
20:36
way together with their program.
20:38
And on the more
20:41
complex side of things, I
20:43
think data aggregation and
20:45
also federated learning and machine
20:47
learning is something. My
20:49
take on
20:52
AI XCrypto has mostly
20:54
been that I think
20:56
AI agents sort of are
20:58
glorified bots and most
21:00
of it is fully artificial,
21:02
which makes sense, I
21:05
guess, but it's just the
21:07
superficial forced marriage. But
21:09
I think confidential computing can
21:11
actually be a logical fit
21:13
because we have very
21:15
highly sensitive data that could
21:18
benefit all of us, could
21:20
benefit humanity as a whole.
21:22
If we would find a way
21:24
to safely aggregate that into
21:26
models and then have a
21:28
way to, yeah, for example,
21:30
monetize this data rise, so
21:32
it could be encrypted sensitive
21:34
data that's collected on your
21:36
phone by some Deepin project,
21:38
for example, that then uses decentralized
21:40
confidential computing to train a
21:42
model that in itself remains
21:44
encrypted. So it's stored on the blockchain
21:46
as an encrypted model. And so
21:49
whoever wants to use that model has
21:51
to pay some smart contract, which then
21:53
distributes those funds back to the
21:55
people who gave their data to the
21:57
model. They actually never had to
21:59
give it. it remains their ownership, but
22:01
they provided encrypted data that formed
22:04
that model. So there can be
22:06
new economic incentives to produce more
22:08
high quality data. Yeah, and I
22:11
think down the line then, what
22:13
that will lead to is a
22:16
utopian future, where there's safe HGI
22:18
that runs on Solana, for example,
22:20
right? Because you can have safe
22:23
access control and the most high
22:25
quality data being encrypted. So I
22:28
think that's something where blockchain can
22:30
actually enhance AI, I think. The
22:32
Darkfills piece you talked about is
22:35
really interesting because there's a lot
22:37
of stuff in this space that
22:39
feels like it's transparency first. Yes.
22:42
Right. And there are a lot
22:44
of situations where transparency isn't necessarily
22:47
useful or necessary for a market
22:49
to function. I guess like I
22:51
think a lot about OTC desks,
22:54
right? How there are a huge
22:56
percentage of volume and there's... Literally
22:59
zero. I mean you can see
23:01
the movement on chain but like
23:03
other than that like it's all
23:06
settled and wire transfers after the
23:08
fact and you can't really see
23:10
where price is moving and those
23:13
sorts of things. Do you is
23:15
this technology fast enough today that
23:18
you can really run like a
23:20
central limit order book in a
23:22
confidential compute environment or are we
23:25
really still talking more like maker
23:27
and taker like OTC style systems?
23:30
So there's different approaches and what
23:32
we've seen is teams for example
23:34
using FGE fully homomorphic encryption attempting
23:37
to do that and yeah that's
23:39
not a market with sufficient sufficient
23:41
throughput but that's one of the
23:44
reasons why we decided to use
23:46
secure multi-party computation yeah because that
23:49
yields fast performance now moving from
23:51
an honest majority to a dishonest
23:53
majority trust assumption makes things a
23:56
bit slower because you want to
23:58
operate in the smallest possible trust
24:01
assumption, right? So what to be...
24:03
using is a so-called MPC protocol
24:05
in the pre-processing model, which means
24:08
that when you have a computation,
24:10
order matching, for example, that computation
24:12
now split into two phases. One
24:15
pre-processing phase and one online phase.
24:17
And the pre-processing phase is independent
24:20
of the actual... data being used
24:22
to the encrypted orders and parties,
24:24
whatever, and actually even the function,
24:27
it just is used to generate
24:29
so-called correlated randomness. And
24:31
that correlated randomness is
24:34
then being consumed in
24:36
the second phase, the online phase,
24:38
then operations can be performed way
24:41
more efficient. because we have had
24:43
this pre-processing. So one answer to
24:45
the question I think would be
24:47
okay. Since we are using the
24:49
setup, we can, we just need
24:52
to ensure pre-processing takes care, has
24:54
happened. Then we are able to
24:56
facilitate almost as fast as plain
24:58
text operation times for the online
25:01
phase. What we can also do
25:03
is we can make use of
25:05
oblivious data structure, oblivious sorting mechanisms,
25:07
where we can then asynchronously sort
25:09
large order books and crank up
25:12
the speed that way. And then
25:14
there's, I think, I would
25:16
say that performance-wise beer at a
25:18
point where it becomes feasible, but
25:20
yeah, basically I see my task
25:22
and the task of... my team
25:24
and our researchers at Arkham to
25:27
to just similarly to to what
25:29
a fired answer team is doing
25:31
but doing it for MPC to
25:33
just make that as fast as
25:35
possible come up with new cryptographic
25:37
protocols to to batch together communication
25:40
make use of hardware acceleration make
25:42
use of different yeah elliptic curve
25:44
operations we've been tinkering with so
25:46
yeah So I want to get
25:48
into a little bit of the
25:51
differences between fully homomorphic encryption and
25:53
MPC for sort of these use
25:55
cases. So there's sort of like
25:57
you can think about it as like.
26:00
What components do we care about
26:02
being private? Right? So there's like
26:04
the first version we talked about
26:06
that's like software encryption, where it's
26:08
like, I trust you and you want
26:10
to make sure I can verify that
26:12
what you've sent me actually arrived, right?
26:14
And no one like monkeyed with it
26:17
in the middle. There's the stuff you've
26:19
been talking about too, or it's like,
26:21
oh, we need to verify that like,
26:23
like, something exists without proving what
26:25
it is. Right, that's kind of one
26:27
piece we've talked about. And then like
26:29
FGE, the way folks talk about it
26:32
is like, oh, you can, a computer
26:34
can compute a result without knowing the
26:36
inputs or the output of that result,
26:39
which is like we've had black
26:41
box style computing for a while,
26:43
but usually the inputs are known
26:45
and the outputs are known and
26:47
that. process in the middle may
26:49
be like semi-blackboxed but like how
26:51
do these first off is that
26:53
a decent like explanation of how
26:55
this stuff works or like from
26:57
a from how you would interact
26:59
with it perspective and then where
27:01
does MPC versus FHE sort of
27:03
have different advantages and disadvantages?
27:06
So yeah I think this
27:08
idea of blackbox computing really
27:10
describes what both MPC and
27:12
FGE are trying to achieve
27:14
and but also trusted execution
27:16
environments. Yeah. So if you're
27:19
really talking about confidential computing
27:21
and black box operations those
27:23
are the free technologies to
27:25
think about because the K
27:27
allows for verifiability and two-party
27:29
private one-party privacy basically right
27:31
so those free FAGE MPC
27:33
and TES are what what
27:36
we should compare TES for me at
27:38
least from from. Yeah me
27:40
being able to look at Wikipedia
27:42
and news articles and and different
27:45
software exploits Just isn't the way
27:47
to go. Yeah, it seems like
27:49
SGX is hacked every other week
27:52
pretty much Yeah And also
27:54
if you don't have physical
27:56
security on the machines. Yeah,
27:58
it seems like it doesn't
28:00
particularly work. Yeah, I think it
28:03
works in a trusted cloud environment,
28:05
with a trusted software vendor, right?
28:07
Right, with physical security. Exactly, yeah.
28:10
Yeah. And it's also so striking
28:12
when Apple announced their T.E. hardware
28:14
for Apple Intelligence. in their blog
28:17
article they were describing them physically
28:19
guarding the machines exiting the manufacturing
28:21
plant and going to the data
28:24
center and you just have this
28:26
mental picture in mind with someone
28:28
with an assault rifle guarding a
28:31
computer right and you're like is
28:33
it actually this this amazing hardware
28:35
that can can do blackbox operations
28:37
doesn't seem that way right so
28:40
yeah it's sort of feels like
28:42
it being impossible. And to be
28:44
honest, we also have to look
28:46
at cryptography in a critical way
28:49
and have to think about quantum
28:51
resistance, right? So I think that's
28:53
also something that a lot of
28:55
folks, especially in our CK, MPC,
28:57
FGE space, try to ignore. But
28:59
I think that's something that we need
29:02
to work on with faster pace as
29:04
well. But I think these basically, and
29:06
for the folks. who are not as
29:08
familiar as Austin and IR with
29:10
that hardware-based security where basically the
29:13
hardware manufacturer has some privacy and
29:15
security promises that they give to
29:18
whoever purchases this hardware and we've
29:20
just seen this. huge amount of
29:22
exploits over the years a few
29:25
weeks ago. I think the most
29:27
recent one with Intel S3X, Intel's
29:30
trusted hardware for like the third
29:32
time. Yeah. And we even had
29:34
blockchain networks completely rely on that
29:37
and their privacy completely get Yeah,
29:39
researchers were able to completely decrypt the
29:41
secret network back to the Genesis block,
29:43
which is crazy. And the avalanche bridges
29:46
and a bunch of stuff like that.
29:48
Yeah. And so I think if you
29:50
want to build this Technology
29:53
that is future-proof, right, to build
29:55
this open network on which the
29:57
NASTEC can run, right? You don't
29:59
want... on some technology by every
30:01
five years, you need to
30:03
replace it with a new
30:06
trusted system by some other
30:08
manufacturer, you want something future
30:10
proof. So purely cryptographic systems,
30:12
I think is the answer.
30:14
And then the question is,
30:16
FGE versus MPC. FHE, basically,
30:18
as you described, allows you
30:20
to have an encrypted representation
30:23
of some value and perform
30:25
an operation. on that encrypted
30:27
representation of a value and
30:30
remain in that representation, right?
30:32
So it's, you're not exiting,
30:35
you're not decrypting, you remain
30:37
in this encrypted representation, which
30:39
is incredibly powerful. Yeah. It's
30:42
also very hard to think
30:44
through how that, like we don't have
30:46
a model for that. in the physical world.
30:49
Usually there's like a black box,
30:51
like something happens inside the box.
30:53
It's very hard to think of
30:55
like anything else where I can
30:57
perform an operation on something and
31:00
I have no ability to see
31:02
what the thing is. It's very
31:04
interesting. That's true actually, but
31:06
with FGE, powerful concept, and
31:08
the problem really is... It's
31:10
practical limitations and that
31:12
boils down to some
31:14
of the operations that
31:16
you're performing with FGE
31:19
fully homomorphic encryption Homomorphisms
31:21
mathematical homomorphisms basically means
31:23
that Yeah, two operations
31:25
are homomorphic which allows you
31:27
to remain in this representation
31:29
and multiplications as being one
31:32
of those operations are costly
31:34
because for when performing
31:36
those multiplications so-called noise
31:39
is being accumulated. And once
31:41
there's too much noise accumulated,
31:43
it's not possible to use
31:46
that ciphertext anymore. So this
31:48
noise needs to be reduced
31:50
using costly bootstrapping operations. And
31:53
that's mainly what makes FGE
31:55
so incredibly costly, meaning in
31:57
practice many orders of magnitude.
32:00
above your normal execution times
32:02
for whatever operations you
32:04
need to perform. And that's
32:06
of course not just
32:08
some latency penalty. That's also
32:10
computational power overhead, right? That
32:12
you have to work
32:14
with. So that's why we're
32:16
using MPC. Now MPC
32:18
is very closely related to
32:20
FHE. And so what
32:22
we're using is not fully
32:24
homomorphic encryption but somewhat
32:26
homomorphic encryption, which is a
32:28
subset and just makes
32:30
use of one operation being
32:32
homomorphic. And multiplication is
32:34
this very expensive operation we
32:36
are able to make use
32:38
of actually what we discussed
32:40
earlier, this preprocessing. That we
32:43
can make use of that
32:45
and that just makes it
32:47
very, very efficient. But of
32:49
course there's more trade -offs you
32:51
have to think about, right?
32:53
So with FHE, you're in this
32:55
encrypted representation, which is nice,
32:57
but what if you want to
32:59
exit that, right? What if
33:01
at some point under a given
33:03
condition you want to selectively
33:06
disclose some information about the encrypted
33:08
state, which especially becomes important on a
33:10
blockchain because I think you don't
33:13
want everything to be encrypted. I think
33:15
you want to be able to
33:17
move in and out of those representations.
33:19
Right, or even just like you
33:22
can see there being needs to
33:24
check point stuff and to provide
33:26
some amount of external observability at
33:28
like various pipeline points. Yes,
33:30
yeah, I think so too.
33:32
And so that's something that
33:34
MPC natively allows for this
33:36
form of selective disclosure. So
33:38
I feel like, and luckily
33:40
I'm not alone with that
33:42
opinion, but I mean, we
33:44
have one of the leading
33:46
FHE researchers in our team
33:48
and he's also fully convinced
33:50
of MPC being this technology
33:52
that allows for this generalized
33:54
confidential computing, both for on -chain
33:56
applications but also for off -chain
33:58
applications for AI. use cases. So
34:01
I think that's the correct path
34:03
to take and to make use
34:05
of any kind of hardware optimization
34:08
that we can make use of
34:10
both for MPC but also FGE
34:12
and CK. Yeah. So changing topic
34:14
a little bit, you're a co-founder.
34:17
Yes. You guys do a lot
34:19
of low-level research on this type
34:21
of work. Everything we've talked about
34:24
today. is about as far away
34:26
from a commercial product as you
34:28
can think of. How do you
34:31
think about balancing, when you're working
34:33
at this type of like fairly
34:35
groundbreaking technology, and a lot of
34:37
it being math, right? There's no
34:40
moat around math. You publish the
34:42
paper, it's out, everyone will use
34:44
it, right? How do you think
34:47
about building a product and a
34:49
business around this type of like
34:51
low level research? So I think
34:54
it's really important to... to have
34:56
a team that can satisfy both
34:58
ends of what's required to build
35:00
a product. It's this low-level research
35:03
being able to do the mathematics,
35:05
but at the same time, as
35:07
we discussed earlier, be able to
35:10
build programming interfaces, the developers can
35:12
just use. It doesn't make any
35:14
sense to build the best cryptography
35:17
and then, yeah, hand a research
35:19
paper to the developer and be
35:21
like, okay. read this manually, read
35:23
this research paper and then afterwards
35:26
you'll be able to to add
35:28
confidentiality. And so what we're working
35:30
on really is to have this
35:33
most accessible kind of confidential computing
35:35
because otherwise why would anybody use
35:37
it right? So I think that's
35:39
that's important striking this this balance
35:42
and to be honest. those low-level
35:44
researchers and PhDs often tend to
35:46
not find the most practical things.
35:49
And so it's important to have
35:51
also folks who, I mean, including
35:53
us when we were building a
35:56
lucid, but now in the meantime,
35:58
have more folks from the Solana
36:00
ecosystem be part. of our team
36:02
who then think like a developer,
36:05
right? They think like the end
36:07
user. And so I think in
36:09
general building a business around this
36:12
for us means building Arquium as
36:14
a global supercomputer and allowing anyone
36:16
to access that supercomputer for confidential
36:19
computations. And the supercomputer just needs
36:21
to be as accessible as possible
36:23
because then any kind of application
36:25
can have confidential computing. And what
36:28
do you see as sort of
36:30
technologies that are worth paying attention
36:32
to over the next few years
36:34
within this domain? It feels like
36:36
every few years there's a new
36:38
classification of something that gets stuck
36:40
on people's minds, raises a bunch
36:43
of money, maybe it works, maybe
36:45
it doesn't, but what are you
36:47
looking out as sort of some
36:49
of the advances coming to either
36:51
MPC or CK technology in general?
36:53
That's a very good question. Or
36:55
do we kind of have it?
36:58
We just need to make it
37:00
faster. So we have a lot of
37:02
what we need actually. I think
37:04
so too. And it's not just
37:06
about making it faster. It's also
37:09
about building those things. And as
37:11
I said, making them accessible. I
37:13
think that's one of the big
37:15
problems with CK, I think. Also,
37:18
what's not accessible is
37:20
requiring developers to learn a
37:22
new DSR. That's also not
37:24
accessible. And we've seen CK
37:26
teams that now introduce a
37:28
new domain-specific language that developers
37:30
need to learn. I think that's one
37:32
of the superior approaches that Solana has
37:34
taken as well. To by default, say,
37:37
hey developer, use rust, which is this
37:39
general programming language, which over the time
37:41
that Solana adopted it, has been adopted
37:43
by so many people in parallel, right?
37:46
So I think that was one of
37:48
the most logical decisions to make. to
37:50
be made and we are trying to
37:52
do it in the same way to
37:55
allow developers to add confidentiality
37:57
where they are at right now but that
37:59
takes time, tooling takes time and
38:02
optimization takes time as well.
38:04
So be it hardware acceleration,
38:06
be it communication optimization, be
38:09
it optimizing network communication protocols,
38:11
there's just a lot of
38:13
things that need to be
38:16
optimized. I think what I
38:18
personally would be or am
38:20
most excited about really is
38:23
protocols that optimize. network communication.
38:25
I think because that's one
38:27
of the big bottlenecks, optimizing
38:30
network communication, and the biggest
38:32
thing for me is oblivious
38:34
data structures. Because what my
38:37
vision really is, and that
38:39
might not be super relevant
38:41
for blockchain applications, I guess,
38:44
but my vision really is
38:46
to allow for fully encrypted
38:48
RAM programs. So it's not
38:51
just some small functionality, but
38:53
it's... anything your computer can
38:55
execute and it's not like
38:58
you pay gas for the
39:00
computation and then it's done
39:02
but it's run for the
39:05
next 10 years in a
39:07
confidential way right that's yeah
39:09
and and that requires complex
39:12
sort of encrypted data structures
39:14
that can have arbitrary size
39:16
and so yeah I think
39:19
that's what I'm most excited
39:21
about in that space. Excellent.
39:23
Well Henick thank you for
39:26
coming on validated. Thanks for
39:28
having me. Appreciate it. Valedated
39:30
is produced by Ray Belli
39:33
with help from Ross Cohen,
39:35
Brandon Hector, Amira Valiani, and
39:37
Ainsley Medford. Engineering by Tyler
39:40
Morissette.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More