Noise Without Regret: How Error Feedback Fixes Differentially Private Image Generation
Opening — Why this matters now Synthetic data has quietly become the backbone of privacy‑sensitive machine learning. Healthcare, surveillance, biometrics, and education all want the same thing: models that learn from sensitive images without ever touching them again. Differential privacy (DP) promises this bargain, but in practice it has been an expensive one. Every unit of privacy protection tends to shave off visual fidelity, diversity, or downstream usefulness. ...