Federated learning keeps gradients local. Clean rooms control access. But neither keeps raw data where it belongs and supports lossless, governed, multimodal sharing. Datasent does.

AI systems are data hungry. The larger and more capable the model, the more data it requires — and the more diverse that data needs to be. This creates a structural tension: the organizations with the most valuable data are often the ones least able to share it.
Healthcare providers hold patient records that would transform medical AI. Financial institutions hold transaction histories that would unlock credit models for the underserved. Manufacturers hold sensor streams from production lines that would make predictive maintenance dramatically more effective. But the compliance exposure, security risk, and operational complexity of sharing that data is prohibitive.
Existing solutions address parts of this problem. Federated learning keeps raw training data local, but it is lossy — gradient aggregation discards information — and application-specific. Secure data cleanrooms govern access, but they typically require raw data to enter a controlled environment managed by a third party. The raw data still moves.

"Datasent eliminates that requirement entirely. Raw data never leaves the Data Holder in any configuration. What AI systems receive is a structured residual representation — compact, governed, and lossless."
There is a second advantage beyond privacy: Datasent tokens are natively useful for machine learning, at zero additional computational cost.
Each token contains a coefficient matrix that captures local trend, curvature, and spectral components — precisely the features that preprocessing pipelines normally derive from raw data before feeding a model. The residual captures local deviations from the fitted structure. The model family identifier provides an unsupervised structural label for each segment. Together, these constitute a rich, structured feature representation that is available immediately from the encoding step.
For time-series data, this has dramatic implications for transformer architectures. A dataset with one million time steps, encoded with an average segment length of one thousand steps, produces a token sequence of one thousand elements rather than one million. Attention complexity falls from quadratic in one million to quadratic in one thousand — a reduction of six orders of magnitude, with no loss of information.
The trusted setup architecture makes federated machine learning across organisational boundaries structurally simpler than any existing approach. Organisations exchange tokenised residuals rather than raw datasets or gradient updates. No raw training data leaves any organisation's environment. The custodian model provides explicit authorisation and auditlogging for every reconstruction event.
This is not federated learning in the conventional sense —it is something more powerful. Federated learning aggregates gradient updates from distributed training runs, accepting information loss in exchange for privacy. Datasent allows the actual data to be shared, loss lessly, without the raw data ever moving. The receiving party can train on the full reconstructed dataset under authorised access, with a complete audit trail.

The longer-term direction is more radical: running MLinference directly on Datasent-encoded data, without a decode step. The token algebra established in Datasent's full mathematical framework provides a formal foundation for defining operations that compose naturally over the structured token representation — meaning that model weights trained on decoded data can, in principle, operate directly on tokens.
This direction is early. But it points toward a future in which the encoding layer is not a preprocessing step that happens before computation — it is the substrate on which computation runs.
AI is often framed as a problem of models, compute, or algorithms. But in practice, most limitations emerge much earlier — at the data layer.
Modern AI systems depend on continuous, large-scale access to data, yet the infrastructure supporting that data was never designed for this reality. As a result, organizations face bottlenecks in performance, rising costs, and increasing exposure risk — not because their models are insufficient, but because their data systems cannot keep up.

John Rhye
Position