ChatGPT - Solving AI’s Privacy Issue
There is no doubt that AI, and in particular Large Language Models (LLMs) such as ChatGPT will have a tremendous impact on society, perhaps even more than the Internet did before. From education to healthcare, movies to music, art to finance, not a single industry isn’t being disrupted by AI.
While foundational models that are trained on generic, publicly accessible data are powerful, they really become useful when contextualised for a given task or user, either through fine-tuning on private data, or pre-prompting with contextual information before sending in a query.
In both cases however there is a major privacy issue: all this private data goes to the company operating the service!
This is why several countries and companies, such as Italy and Samsung, are now limiting the use of ChatGPT and alike. Without strong privacy guarantees, the risk of data breach and manipulation is simply too high.
Can cryptography solve the privacy problem?
Fortunately, there is a way to both use AI and keep our data private: Fully Homomorphic Encryption (FHE) is a new encryption technique that enables computing on encrypted data, without actually decrypting it. And it may be a way to bridge the gap between the effective use of AI and keeping our data private.
When applied to AI, it works in the following way:
- The user encrypts their data and query using a secret key that they only know
- The encrypted data is then sent to the server running the AI model, which then processes it encrypted, producing a result which itself is encrypted. At no point does the server see the data, everything is done blindly!
- The user then decrypts the response from the AI, revealing its content.
What this means is that for users, nothing changes: they send queries and get an answer, but since the data is encrypted both in transit and during processing, nobody can see it: neither the company offering the service, nor governments or hackers. It’s end-to-end encryption for AI!
Of course, privacy is just a drop in a broader ocean of LLM-associated challenges that also involve discussions around copyright and unconscious bias, and FHE will therefore not offer a silver bullet to all the practical issues currently being discussed. However, it has the potential to evolve into a key piece of the current puzzle.
Why Aren’t We Using This Already?
The reason why FHE isn’t being used in widespread applications today, is because up until recently, it was too slow, too complicated and too limited to be useful. It took a PhD in cryptography to do a simple encrypted multiplication, and that would take minutes to complete. But thanks to recent development breakthroughs from a number of companies and academic institutions, as well as hardware acceleration efforts from companies such as Intel and Cornami, homomorphic encryption is quickly becoming a reality.
On the usability side, developers no longer need to know cryptography to use FHE. They can simply use homomorphic compilers to write Python code and have it automatically converted to an encrypted equivalent. On the feature side, we are also no longer limited to a handful of encrypted additions and multiplications. Anything is now doable in FHE, from deep neural networks to blockchain smart contracts to genomics. The only thing missing is performance.
Using traditional CPUs and GPUs to run ChatGPT encrypted end-to-end would cost tens of thousands of dollars per query, vs a few cents if the data isn’t encrypted. This means we need at least 100,000x better performance if we want FHE to be cost effective enough that it becomes the norm.
Thankfully, we have a solution: hardware acceleration. By creating dedicated chips for homomorphic encryption, we can make it anywhere from 1,000x to 10,000x faster, while simultaneously being 5-10x cheaper than conventional chips. Together, this means the 100,000x cost improvement we are looking for is within reach, and likely to happen in the next 5 years as these accelerators become available commercially.
While privacy isn’t the only issue with AI, it is a major hurdle for global adoption. Without it, we would need to trust a handful of companies with our most private information, or not use AI at all.
This is why homomorphic encryption is such a big deal: it solves the AI privacy dilemma, by allowing us to both use AI and keep our data private! Because in the end, if AIs don’t know anything about us, perhaps they won’t be able to harm us as much.
Dr Rand Hindi is CEO of Zama
Image: Shubham Dhage
You Might Also Read:
Guidelines For AI Systems Development:
___________________________________________________________________________________________
If you like this website and use the comprehensive 6,500-plus service supplier Directory, you can get unrestricted access, including the exclusive in-depth Directors Report series, by signing up for a Premium Subscription.
- Individual £5 per month or £50 per year. Sign Up
- Multi-User, Corporate & Library Accounts Available on Request
- Inquiries: Contact Cyber Security Intelligence
Cyber Security Intelligence: Captured Organised & Accessible