Blockchain and AI: Separating Fact from Fiction
Introduction
Industries everywhere are asking “What can AI do for us?” But the blockchain industry, known for challenging norms, is also asking the opposite question: “What can blockchain do for AI?”
Narrative #1: Blockchain can combat misinformation caused by generative AI
An expert panel at a recent Coinbase event concluded that “blockchain can counter misinformation with cryptographic digital signatures and timestamps, making it clear what’s authentic and what’s been manipulated.”
This is true only in a very narrow sense.
Blockchains can record digital-media creation in a tamper-proof way, i.e., so that modification of specific images is detectible. But this is a far cry from clarifying authenticity.
Consider a photo of a flying saucer hovering above the Washington Monument. Suppose that someone has registered its creation in, say, block 20,000,000 of the Ethereum blockchain. This fact tells you one thing: The flying saucer image was created before block 20,000,000. Additionally, whoever posted the image to the blockchain — let’s call her Alice — did so by digitally signing a transaction. Assuming that Alice’s signing key wasn’t stolen, it’s clear that Alice registered the photo on the blockchain.
None of this, however, tells you how the image was created. It might be a photo that Alice snapped with her own camera. Or Alice might have gotten the image from Bob, who Photoshopped it. Or maybe Carol created it with a generative AI tool. In short, the blockchain tells you nothing about whether aliens were touring Washington, D.C.—unless you already trust Alice to begin with.
Some cameras can digitally sign photos to authenticate them (assuming their sensors can’t be fooled, which is a big if), but this isn’t blockchain technology.
Narrative #2: Blockchain can bring privacy to AI
Model training is a data-hungry operation. The bigger and better the training data set, the better the resulting model. For many applications, training on private user data is essential. For instance, creating a good machine-learning model to diagnose medical conditions requires data from a population of real patients. Handling such highly sensitive data securely is a challenge. Some are trumpeting blockchain technologies as a solution.
Blockchains, however, are designed for transparency — a property at odds with confidentiality.
Proponents point to privacy-enhancing technologies advanced by the blockchain industry to address this tension — especially zero-knowledge proofs. Zero-knowledge proofs, however, don’t solve the problem of privacy in AI model training. That’s because a zero-knowledge proof doesn’t conceal secrets from whoever is constructing the proof. Zero-knowledge proofs are helpful if I want to conceal my transaction data from you. But they don’t enable me to compute privately over your data.
There are other, more relevant cryptographic and security tools with esoteric names, including fully homomorphic encryption (FHE), secure multiparty computation (MPC) and secure enclaves. These can in principle support privacy-preserving AI (specifically, “federated learning”). Each has important caveats, though. And claiming them as blockchain-specific technologies would be a stretch.
Narrative #3: Blockchains can empower AI bots with money — and that’s a good thing
Jeremy Allaire, CEO of Circle, has noted that bots are already performing transactions using cryptocurrency and tweeted that “AI and Blockchains are made for each other.” This is true in the sense that cryptocurrency is a good match for the capabilities of AI agents. But it’s also worrisome.
Many people fret about AI agents escaping human control. Classic nightmare scenarios involve autonomous vehicles killing people or AI-powered autonomous weapons going rogue. But there’s another vector of escape: The financial system. Money equals power. Give that power to an AI agent and it can do real damage.
This problem is the topic of a research paper that I co-authored in 2015/6. My colleagues and I examined the possibility of smart contracts, programs that autonomously intermediate transactions on Ethereum, being used to facilitate crime. Using the techniques in that paper and a blockchain oracle system with access to LLMs (Large Language Models) such as ChatGPT, bad actors could in principle launch “rogue” smart contracts that automatically pay bounties for committing serious crimes.
Conclusion
The integration of blockchains and AI does hold clear promise. AI may add unprecedented flexibility to blockchain systems by creating natural language interfaces to them. Blockchains may provide new financial and transparency frameworks for model training and data sourcing and put the power of AI in the hands of communities, not just enterprises.
It’s still early days, though, and as we wax lyrical about AI and blockchain as an enticing mix of buzzwords and technologies, we need to really think — and see — things through.
FAQs
Q: Can blockchain combat misinformation caused by generative AI?
A: No, blockchains can only record digital-media creation in a tamper-proof way, but this is a far cry from clarifying authenticity.
Q: Can blockchain bring privacy to AI?
A: No, blockchains are designed for transparency, which is at odds with confidentiality. Zero-knowledge proofs and other cryptographic tools can provide privacy, but these are not specific to blockchain technology.
Q: Can blockchains empower AI bots with money?
A: Yes, but this also raises concerns about AI agents escaping human control and the potential for financial damage.