Disentangled Contrastive Learning From Synthetic Matching Pairs for Targeted Chest X-Ray Generation

Disentangled generation enables the synthesis of images with explicit control over disentangled attributes. However, traditional generative models often struggle to independently disentangle these attributes while maintaining the ability to generate entirely new, fully randomized, and diverse synthe...

Full description

Saved in:
Bibliographic Details
Main Authors: Euyoung Kim, Soochahn Lee, Kyoung Mu Lee
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10844299/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Disentangled generation enables the synthesis of images with explicit control over disentangled attributes. However, traditional generative models often struggle to independently disentangle these attributes while maintaining the ability to generate entirely new, fully randomized, and diverse synthetic data. In this study, we propose a novel framework for disentangled Chest X-ray (CXR) generation that enables explicit control over person-specific and disease-specific attributes. This framework synthesizes CXR images that preserve the same patient identity—either real or randomly generated—while selectively varying the presence or absence of specific diseases. These synthesized matching-paired CXRs not only augment training datasets but also aid in identifying lesions more effectively by comparing attribute-specific differences between paired images. The proposed method leverages contrastive learning to disentangle latent spaces for patient and disease attributes, modeling these spaces with multivariate Gaussians for precise and exclusive attribute sampling. This disentangled representation enables the training of a controllable generative model capable of manipulating disease attributes in CXR images. Experimental results demonstrate the fidelity and diversity of the generated images through qualitative assessments and quantitative comparisons, outperforming state-of-the-art class-conditional generative adversarial networks on two public CXR datasets. Further experiments on clinical efficacy demonstrate that our method improves disease classification and detection tasks by leveraging data augmentation and employing the difference maps generated from paired images as effective attention maps for lesion localization. These findings underscore the potential of our framework to improve medical imaging analysis and facilitate novel clinical applications.
ISSN:2169-3536