in AI for some time, you’re in all probability an LLM/Agent/Chat person, however have you ever ever requested your self how these instruments might be educated within the close to future, and what if we have now already used up the information we have to practice fashions? Many theories say that we’re working out of high-quality, human-generated information to coach our fashions.
New content material goes up daily, that’s a actuality, however an growing share of what will get added every day is itself AI-generated. So in the event you preserve coaching on public net information, you’re ultimately coaching on the outputs of your personal predecessors. The snake consuming its tail. Researchers name this phenomenon Mannequin Collapse, the place AI fashions begin studying from the errors of their predecessors till the entire system degrades into nonsense.
However what if I instructed you we aren’t really working out of knowledge? We’ve simply been trying within the incorrect place.
On this article, I’m going to interrupt down the important thing insights from this good paper.
The Net We Already use and the Net That Issues
Most of us think about the online as a singular supply of knowledge. In actuality, there are at the very least two.
There may be the Floor Net: the listed, public world like what we discover on Reddit, Wikipedia, and information websites. That is what we’ve already scraped and overused for years to coach the mainstream AI fashions of immediately. Then, there’s what we name the Deep Net, and right here I’m not speaking concerning the “Dark Web” or something unlawful.
The Deep Net is just every part behind a login or a firewall. It refers to something on-line that isn’t publicly listed. It may very well be your hospital’s affected person portal, your financial institution’s inside dashboard, enterprise doc archives, personal databases, and years of e mail sitting behind a login display screen. Regular, boring, however extremely worthwhile information.
Many research recommend the Deep Net is orders of magnitude bigger than the floor net. Extra importantly, it’s crucially higher high quality information. In comparison with floor net content material, which may be noisy, stuffed with misinformation, and strongly search engine optimisation optimized. Additionally, it more and more incorporates content material intentionally designed to mislead or poison AI fashions. Deep net information, like medical data or verified monetary paperwork or others inside databases, tends to be clear, authenticated, and arranged by individuals who care about its high quality.
The issue? I feel you may guess it, it’s personal. You’ll be able to’t simply extract one million medical data with out contemplating all of the authorized and moral catastrophes you’re going to trigger.
The PROPS Framework
That is the place a brand new framework referred to as PROPS (Protected Pipelines) is available in. Launched by Ari Juels (Cornell Tech), Farinaz Koushanfar (UCSD), and Laurence Moroney (former Google AI Lead), PROPS acts as a bridge between this delicate information and the AI fashions that want it.
The brilliance of PROPS is that it doesn’t ask you to “hand over” your information. As an alternative, it makes use of Privateness-Preserving Oracles. Consider an oracle as a “trusted middleman” that may take a look at your information, confirm it’s actual, after which inform the AI mannequin what it must know with out ever displaying the mannequin the uncooked data.
These ideas of props can sounds magical as it might probably remedy quite a lot of points associated to information availability that AI fashions face immediately. However how does this work precisely? Let’s take an instance of a medical firm that desires to coach a diagnostic software on actual well being data. Below the PROPS framework:
- Permission: As a person, you log into your personal well being portal and authorize a particular use on your information.
- The Oracle: Consider the Oracle as a digital notary. It goes to your personal portal (like your hospital database) to confirm that your information is actual. As an alternative of copying your information, it merely tells the AI system: “I have seen the original documents, and I testify they are authentic.” It supplies proof of the reality with out ever handing over the personal information itself. Instruments exist already for this, like DECO. It’s a protocol that lets customers show that they pulled a particular piece of knowledge from an online server over a safe TLS channel.
- The Safe Enclave: It is a “black box” inside the pc’s {hardware} the place the precise coaching occurs. We put the AI mannequin and your personal information inside and “lock the door.” No human or developer can see what is occurring inside. The AI “studies” the information and leaves with solely the mannequin weights. The uncooked information stays locked inside till the session is over.
- The Consequence: The mannequin trains on the information inside that field. Solely the up to date “weights” (the educational) come out. The uncooked information isn’t seen by human eyes.
The contributor is aware of precisely what they’re agreeing to, and they are often rewarded for taking part in a method that’s calibrated to how worthwhile their particular information really is. It’s a genuinely completely different relationship between information homeowners and AI techniques.
However why trouble with this as an alternative of Artificial Knowledge?
Some may ask: “Why bother with this complex setup when we can just generate synthetic data?”
The reply is that artificial information is a variety killer. By definition, artificial information era reinforces the center of the bell curve. When you’ve got a uncommon medical situation that impacts solely 0.01% of the inhabitants, an artificial information generator will doubtless easy you out as “noise.”
Fashions educated on artificial information grow to be progressively worse at serving outliers. PROPS solves this by making a safe method for actual individuals with uncommon circumstances or distinctive backgrounds to “opt-in.” It turns information sharing from a privateness danger right into a “data marketplace.” the place worthwhile information will get the compensation it deserves.
It’s not nearly coaching, inference issues too
Most discussions deal with coaching, however PROPS has an equally fascinating utility on the inference aspect.
For instance, getting a mortgage immediately entails quite a lot of doc submission: financial institution statements, pay stubs, and tax returns. In a PROPS-based system, they recommend the usage of a Mortgage Choice Mannequin (LDM):
- You authorize the LDM to speak on to your financial institution.
- The financial institution confirms your stability through a privacy-preserving oracle.
- The LDM decides.
- The outcome? The lender will get a verified “Yes” or “No” with out ever touching your personal paperwork. This eliminates the chance of knowledge leaks and makes it practically inconceivable for individuals to make use of fraudulent, photoshopped paperwork.
What’s really stopping this from occurring in 2026?
It merely comes right down to scale and infrastructure.
Essentially the most strong model of PROPS requires coaching to occur inside a hardware-backed safe enclave (like Intel SGX or NVIDIA’s H100 TEEs). These work properly at a small scale, however getting them to work for the large GPU clusters wanted for frontier LLMs remains to be an open engineering downside. It requires huge clusters to work in good, encrypted sync.
The researchers are clear: PROPS isn’t a completed product but. It’s a persuasive proof-of-concept. Nevertheless, a lighter-weight model is deployable immediately. Even with out full {hardware} ensures, you may construct techniques that give customers significant assurance, which is already an enchancment over asking somebody to e mail you a PDF.
My Personal Remaining Ideas
PROPS isn’t actually a “new” expertise; it’s a brand new utility of present instruments. Privateness-preserving oracles have been used within the blockchain and Web3 house (like Chainlink) for years. The perception right here is recognizing that the identical instruments can remedy the AI information disaster.
The “data crisis” isn’t a lack of understanding; it’s an absence of belief. We’ve got greater than sufficient information to construct the following era of AI, however it’s locked behind the doorways of the Deep Net. The snake doesn’t must eat its tail; it simply must discover a higher backyard.
👉 LinkedIn: Sabrine Bendimerad
👉 Medium:
👉 Instagram:



