Verifiable AI: Moving From a Black Box To a Glass House.


        ▓▓▓▓▓▓▓▓▓▓▓                              ░░░░░░░░░░░
        ▓           ▓                             ░    ░    ░
        ▓     ?     ▓                             ░  P ░ R  ░
        ▓           ▓                             ░    ░    ░
        ▓           ▓                             ░░░░░░░░░░░
        ▓           ▓                             ░  O ░ O  ░
        ▓           ▓                             ░    ░    ░
        ▓           ▓                             ░░░░░░░░░░░
        ▓           ▓                             ░  F ░ ✓  ░
        ▓▓▓▓▓▓▓▓▓▓▓                              ░░░░░░░░░░░
     ▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓                         ░░░░░░░░░░░░░░░

           ████████████████████████████████████████████
     
                ░░
              █ Ο █  ────►──────►──────►──────►  █▀█
              █ ▄▄▄█                               ███
              █   █                               █▀█
              ▀█ █▀                               

                            ____  
                        o8%8888,    
                      o88%8888888.  
                     8'-    -:8888b   
                    8'         8888  
                   d8.-=. ,==-.:888b  
                   >8 `~` :`~' d8888   
                   88         ,88888   
                   88b. `-~  ':88888  
                   888b ~==~ .:88888 
                   88888o--:':::8888      
                   `88888| :::' 8888b  
                     8888^^'       8888b  
                    d888           ,%888b.   
                   d88%            %%%8--'-.  
                  /88:.__ ,       _%-' ---  -  
                  '''::===..-'   =  --.
                             .:.    
                        ‘I see.’
                      ‘Verifiable.’
                    ‘No secrets inside.’

     ████████████████████████████████████████████████████████████████████

              ██████╗ ██╗  ██╗███╗   ███╗██╗     
              ╚══███╔╝██║ ██╔╝████╗ ████║██║     
                ███╔╝ █████╔╝ ██╔████╔██║██║     
                ███╔╝  ██╔═██╗ ██║╚██╔╝██║██║     
              ███████╗██║  ██╗██║ ╚═╝ ██║███████╗
              ╚══════╝╚═╝  ╚═╝╚═╝     ╚═╝╚══════╝


So there you are, minding your own business, when your Japanese wife mentions she needs a frying pan, the one you have is too curved to cook eggs.. Nothing unusual about that — couples talk about kitchen needs all the time. But then, like some digital stalker with perfect hearing, YouTube serves you an ad featuring an Asian woman lovingly seasoning her wok, speaking in that warm, familiar tone that says "I understand your household dynamics." The algorithm didn't just connect "frying pan" to your search history — it profiled your marriage, cross-referenced your location data, factored in your browsing patterns, and delivered what it calculated would be the most emotionally manipulative pitch possible.

It did this all without ever explaining how it knew, why it knew, or what else it thinks it knows about you. It's like having a private investigator living in your pocket, one who never shows you their case files but keeps rearranging your world based on conclusions you'll never see. The AI makes its move, pulls its strings, and expects you to just trust that this creepy precision is somehow in your best interest. But here's the thing: you never asked to be profiled, analyzed, and packaged into a marketing demographic. You just wanted to live your life without some black box playing puppet master with your personal data.

Now imagine handing that same black box your credit card and the keys to your digital kingdom. AI agents are the logical next step — sophisticated programs that don't just recommend frying pans, but actually buy them for you, book your flights, manage your investments, and negotiate your contracts. Silicon Valley evangelists paint a picture of frictionless living: your AI butler anticipating every need, executing every task, all while you sip coffee and live your best life. But here's what they don't mention in the glossy demos — you're essentially giving a sophisticated gambling machine access to your bank account and trusting it won't go rogue. We have all seen this movie before right?

When your AI agent decides to short NVIDIA stock because it "detected market sentiment," or books you a $3,000 flight to Tokyo because it misinterpreted your casual comment about wanting sushi, good luck getting an explanation that makes sense. Customer support - is likely to be that very same bot.

Does Web3 have an answer? Yes. Many propose Trusted Execution Environments— TEEs — those hardware secured enclaves that supposedly keep everything safe and honest. Sounds great in theory. But tell me, how many of your devices actually have TEE enabled? That laptop you're reading this on? Your phone? Your smart TV? Most people couldn't turn on a trusted enclave if their digital life depended on it, and frankly, most don't even know it exists. So we're building a world where everyone depends on trusted providers (Web3 protocols) that run trusted hardware (TEEs), that in turn run trusted and invisible algorithms (blackbox AI). Some propose a blend of TEE with ZKP, which is a great short term play. But what happens if AI is executing programs on my local device?

Cryptography meet AI.

This is where the cryptographic cavalry rides in—zero-knowledge proofs that don't need a supercomputer to run. When your AI agent is making moves on your behalf, you need more than blind faith; you need mathematical certainty wrapped in computational efficiency. Enter lookup arguments and memory-efficient ZKPs, the unsung heroes that could actually make this dystopian AI future — less dystopian. Think of it this way: every time your local AI executes a program, whether it's analyzing your spending patterns or deciding which stocks to trade — it generates a succinct cryptographic receipt proving it did exactly what it claimed to do, no more, no less.

These aren't the bloated, energy-guzzling proofs of old' that required server farms and special hardware to produce. We're talking about proofs small enough to fit in a tweet, fast enough to generate on your phone, yet mathematically bulletproof enough to stake your financial future on. The real magic happens when these efficient proofs become the lingua franca of AI communication — your local agent can trigger another AI halfway across the world, passing along a tiny cryptographic breadcrumb that says "I ran the calculation correctly, here's the proof, now you can trust my output enough to act on it." No need to re-run the computation, no need to trust the messenger or some special hardware regime, just pure mathematical verification.

Of course, mention zkML to most people and they'll look at you like you just suggested we colonize Mars by Tuesday. "Zero-knowledge machine learning? Come on, that's pure academic fantasy — there's no way you can prove neural network computations without turning your laptop into a space heater."

I get it.

The conventional wisdom says cryptographic proofs for AI are impossibly expensive, that you'd need Amazon's entire North East server fleet just to verify a simple image classification. But here's what the skeptics are missing: we're not trying to prove GPT-4 ran correctly on a billion-parameter model.

We're talking about the small, targeted programs that actually matter in your daily life — the local AI that decides whether to buy that stock, the agent that parses your email to book a meeting, the algorithm that analyzes your spending patterns before moving money around. These aren't moonshot problems anymore. The math has gotten good enough, the hardware has gotten fast enough, and frankly, the incentives have gotten desperate enough that zkML is happening whether people believe in it or not.

My team is using folding schemes with JOLT lookup arguments — think of it as a precompiled lookup-centric nuke for matmul and non-linearities when needed, but also a power plant turning out energy in a consistent rate with folding (space efficiency). Like a star fixed in the perfect balance of gravity and energy production. This setup also surprisingly and almost uniquely — supports true ZK (privacy). We're not talking about some distant future where quantum computers make everything possible. We're talking about this year, maybe sooner. While everyone else is debating whether it's theoretically feasible, we are already running the code.

And look, if we're being honest, not every AI decision needs to happen in real-time with zero-knowledge privacy. Sometimes the old-fashioned approach works just fine: delegate the heavy cryptographic lifting to a more powerful machine and wait for the receipt. Think about it — when your AI agent is about to send $10,000 in USDC to some DeFi protocol, or purchase that "great investment opportunity" it found, maybe a two-minute loading screen isn't the end of the world. Maybe that spinning wheel that says "generating cryptographic proof of correctness" is exactly what you want to see before your life savings gets moved around by an algorithm.

We've been conditioned by decades of instant gratification to think every digital interaction needs to complete in milliseconds, but some things are worth waiting for. Your coffee order? Sure, make it instant. Your mortgage refinancing decision? Maybe we can spare a few minutes to mathematically verify the AI isn't about to screw you over. The beauty of delegated proving is that you get the best of both worlds — the computational power of enterprise hardware generating bulletproof cryptographic evidence, with the convenience of running on whatever device you happen to have in your pocket. It's like having a team of forensic accountants double-checking every financial decision, except they work at light speed and speak only in mathematical certainties.

To rent or own?

So let's circle back to that frying pan moment—but imagine it playing out in a fundamentally different world. Instead of YouTube's servers profiling my marriage and serving targeted ads from some distant data center, picture this: my phone's local AI processes that conversation, generates insights about household needs, maybe even suggests products or timing for purchases. But here's the crucial difference — it all happens on my device, with my data, under my control.

Think Apple Intelligence, but instead of another black box you have to trust, every inference comes with a cryptographic proof you can verify. The AI shows its work: "I heard 'frying pan,' cross-referenced your cooking patterns from local data, determined optimal timing based on your calendar, here's the mathematical proof of my reasoning process." No raw data leaves your device, no corporate servers build shadow profiles of your life, but you still get intelligent assistance with full transparency.

This is what verifiable AI really promises — not just moving computation out of big tech's black boxes, but making personal AI that's actually personal. Your local models learn your preferences, anticipate your needs, make suggestions and decisions, all while generating portable proofs that you own and control. When your personal AI agent wants to interact with third-party services — making purchases, booking travel, managing investments — it can share these cryptographic proofs without revealing your underlying data. The service provider gets mathematical certainty that your AI ran correctly and you meet their requirements, while your personal information stays locked in your device.

We're not just moving from black box to glass house; we're moving from rented intelligence to owned intelligence, from surveillance capitalism to mathematical sovereignty with selective transparency.

Wyatt Benno

I build software and write about where AI meets cryptography.

My Twitter

Subscribe to ICME

Don’t miss out on the latest issues. Sign up now to get access to the library of members-only issues.
jamie@example.com
Subscribe