IEEE FG 2026

Privacy-Compliant Human Data Synthesis in Images for GDPR

with Rendering Refined Stable Diffusion (RefSD)

Kartik Patwari*†1, David Schneider*†,2, Xiaoxiao Sun3, Chen-Nee Chuah1, Lingjuan Lyu4, Vivek Sharma*4
1University of California, Davis     2Karlsruhe Institute of Technology
3Stanford     4Sony AI, Sony Research

*Equal Contribution. Partial work done during internship at Sony AI.

RefSD teaser figure

RefSD pseudonymizes humans in commercially reusable public datasets by fully removing the original person, inserting a pose-aligned identity-free 3D avatar, and refining that avatar with text-guided diffusion. The result preserves scene layout and posture while pushing privacy closer to full masking and keeping the image useful for downstream learning.

Abstract

Privacy regulations (e.g., GDPR, CCPA) mandate that public datasets with permissive commercial licenses (e.g., CC BY 4.0) containing humans be pseudononymized before use. However, existing anonymization methods have notable limitations: blurring or masking degrade downstream utility, GAN-based synthesis offers limited control and photorealism, and diffusion editors may retain identity traces. To overcome these limitations, we propose Rendering Refined Stable Diffusion (RefSD), a three-stage pipeline that (1) removes real humans via segmentation and inpainting, (2) reconstructs pose-aligned, identity-free avatars through SMPL-based 3D rendering, and (3) refines appearance with text-guided diffusion for photorealism. By using rendering, RefSD provides explicit control over body shape, clothing and pose, enabling diverse yet structured avatar generation. To validate human alignment, we introduce HumanGenAI, a human-annotation suite for evaluating privacy preservation, perceptual satisfaction, and attribute-generation fidelity. Beyond HumanGenAI, we conduct re-identification and downstream task benchmarks, demonstrating that RefSD matches the re-ID performance of complete masking while achieving competitive utility relative to real images. Together, RefSD and HumanGenAI establish a scalable pipeline and benchmark for privacy-compliant human synthesis in image datasets.

Highlights

  • RefSD is a three-stage, training-free pseudonymization pipeline that sanitizes the scene, inserts a pose-aligned synthetic avatar, and refines it with diffusion for photorealism.
  • The rendering stage gives explicit control over pose, body shape, clothing, and attribute diversity while avoiding reuse of source-human textures or identity cues.
  • HumanGenAI provides a unified benchmark for privacy, pose preservation, perceptual satisfaction, prompt controllability, and fine-grained attribute fidelity.
  • RefSD reaches re-ID privacy nearly identical to full masking, outperforms prior methods in human evaluation, and improves downstream classification and detection when used as synthetic training data.

RefSD Pipeline

RefSD pipeline figure

RefSD is designed as a privacy-first full-body pseudonymization pipeline that replaces humans with pose-aligned synthetic counterparts rather than editing the original subject in place. The paper frames this as a privacy-utility bridge: the original human is completely removed, posture is preserved through 3D body estimation, and diffusion is used only after a synthetic avatar has already been inserted into the scene.

This separation between structure and appearance matters. Rendering gives RefSD explicit geometric control, while text-guided diffusion improves realism and attribute controllability without copying source identity. The resulting images stay closer to the original scene layout than pure masking, while avoiding the identity leakage risks of direct image editing.

1. Pose Estimation & Sanitization

Detect the person, recover 3D body parameters, then fully remove the original human with segmentation and inpainting so no identifiable traces remain in the background.

2. Rendering-Based Synthesis

Build a pose-aligned SMPL avatar, sample from a diverse bank of base bodies and textures, and composite that identity-free avatar back into the sanitized scene.

3. Diffusion-Based Refinement

Use Canny edges from the rendered avatar plus text prompts for demographics, clothing, and context to refine the synthetic person into a more photorealistic yet still privacy-compliant human.

HumanGenAI Evaluation Framework

HumanGenAI evaluation framework figure

HumanGenAI is introduced to benchmark whether pseudonymized humans are not only private, but also aligned with human expectations for realism, posture, and controllable attribute generation. The framework combines structured human evaluation with downstream task analysis so privacy, quality, and utility are judged together instead of in isolation.

Generated Attribute Fidelity (φ)

Tests how well RefSD follows prompts and preserves attribute intent.

  • φA: Prompt complexity across simple, medium, and complex prompts.
  • φB: Facial attribute fidelity over 50 single-attribute face prompts.
  • φC: Fine-grained attribute translation for close pairs such as ethnicity, age, emotion, and skin tone.
  • φD: Full-body attribute representation across 100 subcategories including clothing and occupation.

Generic Property Assurance (ψ)

Measures whether pseudonymized images remain private, aligned, and visually acceptable.

  • ψA: Privacy assessment of how distinguishable the synthesized human is from the original.
  • ψB: Pose preservation using human side-by-side comparisons.
  • ψC: Human satisfaction for realism, quality, and preference.

Quantitative Results

RefSD pairs strong privacy with usable synthetic data. In the paper’s human evaluation, RefSD scores highest across privacy, pose preservation, and perceptual satisfaction, and in the re-identification benchmark it nearly matches the privacy of completely masking the person.

Human Evaluation

RefSD achieves 4.5 / 4.6 privacy, 4.4 / 4.2 pose, and 4.3 / 4.4 satisfaction on face / full-body HumanGenAI ψ evaluations, outperforming SG-GAN, TriA-GAN, and Mask-SD.

Privacy via Re-ID

On Market1501, RefSD reaches R@1 = 27.5 and mAP = 23.9, essentially matching the privacy gold standard of MaskOut (27.4, 23.8) while remaining more visually useful than pure masking.

Utility for Training

RefSD synthetic data improves RAF-DB classification and OpenImages detection. The paper reports up to +11.1% gender classification gain over real-only training, plus +1.1 mAP@[.5:.95] and +1.0 mAP@0.5 for detection over real-only data.

ADD: Insert your quantitative figure, table screenshots, or HTML tables here.

Good candidates are Table II, Table III, Table IV, or a custom summary panel of the privacy and utility numbers.

Qualitative Results

Qualitatively, the paper positions RefSD as a better trade-off than both masked inpainting and prior full-body anonymization baselines: it keeps posture and scene alignment stronger than masking-only methods, and it looks more natural than raw rendering while avoiding direct reuse of source-human appearance.

  • Side-by-side comparisons show stronger pose retention and photorealism than Mask-SD and DP2 baselines.
  • HumanGenAI shows especially strong generation for clear attributes such as ethnicity, clothing, occupation, and distinct emotions like happy or angry.
  • The hardest cases are subtle fine-grained translations, such as closely related ethnicities, older age transitions, and niche facial details.
  • Failure cases include attribute conflicts, partial occlusions, limb/context mismatches, and occasional body-proportion errors from limited avatar coverage.

ADD: Insert qualitative comparison grids and failure cases here.

Good candidates are the main comparison figure, extra RefSD examples, HumanGenAI examples, or the failure-case panel.

BibTeX

@inproceedings{patwari2026refsd,
  title     = {Privacy-Compliant Human Data Synthesis in Images for GDPR},
  author    = {Patwari, Kartik and Schneider, David and Sun, Xiaoxiao and Chuah, Chen-Nee and Lyu, Lingjuan and Sharma, Vivek},
  booktitle = {IEEE International Conference on Automatic Face and Gesture Recognition (FG)},
  year      = {2026}
}