TLDR: A new research paper introduces Partial Feature Membership Inference (PFMI), a privacy attack where adversaries can determine if a data sample was used in a model’s training set, even with only partial feature information. The paper proposes MRAD, a two-stage framework that reconstructs missing features and uses anomaly detection to identify members. Empirical results show MRAD’s effectiveness across various datasets, highlighting a significant new privacy vulnerability in machine learning models.
Machine learning models, while powerful, are increasingly recognized for their vulnerability to privacy breaches. One significant threat is the Membership Inference Attack (MIA), which allows an adversary to determine if a specific piece of data was used to train a machine learning model. This can expose sensitive personal information, such as medical records or financial data.
Traditionally, these attacks assume that the attacker has complete access to all features of the target data sample. However, in many real-world scenarios, this assumption doesn’t hold. For instance, an attacker might only know a person’s age, gender, and race, but not their full medical history. This gap in information has limited the applicability of existing MIA methods.
Introducing Partial Feature Membership Inference (PFMI)
A new research paper, titled Membership Inference Attack with Partial Features, addresses this critical limitation by studying a new inference scenario called Partial Feature Membership Inference (PFMI). In PFMI, the attacker observes only a subset of a sample’s features and aims to infer whether this partial information was part of the model’s training set. This reflects a more realistic threat model, as complete data is often difficult to acquire due to privacy concerns or incomplete collection.
MRAD: A Two-Stage Attack Framework
To tackle the PFMI problem, researchers Xurun Wang, Guangrui Liu, Xinjie Li, Haoyu He, Lin Yao, and Weizhe Zhang propose a novel two-stage attack framework called MRAD (Memory-guided Reconstruction and Anomaly Detection).
The first stage of MRAD focuses on ‘Memory-guided Reconstruction’. Since incomplete data cannot be directly fed into a neural network, MRAD optimizes the unknown feature values by minimizing the sample’s loss to the target model. This process leverages the model’s inherent ‘memorization’ of its training data. If the known features come from a training member, the reconstruction tends to align better with the true data distribution. If not, the reconstruction deviates more, leading to ‘anomalous’ data.
The second stage, ‘Anomaly Detection’, then measures the deviation between the reconstructed sample and the training data distribution. By using anomaly detection techniques, MRAD can distinguish between reconstructed samples that originated from member features and those from non-member features. The framework is designed to be compatible with various existing anomaly detection methods.
Empirical Results and Real-World Implications
The effectiveness of MRAD was demonstrated through extensive experiments on various datasets, including image datasets like CIFAR-10, Fashion-MNIST, and STL-10, as well as the tabular Epsilon dataset. The results show that MRAD is effective even with a significant portion of features missing. For example, on the STL-10 dataset, the attack achieved an AUC (Area Under the ROC Curve) of approximately 0.6 even when 40% of the features were unknown.
The study also found that the importance of the known features significantly impacts the attack’s performance. When the known features are those that contribute most to the model’s predictions, the attack performs better. This highlights that even a small amount of highly relevant partial information can be enough to compromise privacy.
A compelling case study using a simulated patient risk prediction system based on a diabetes dataset illustrated the practical relevance of PFMI. By knowing only publicly accessible demographic information (race, age, gender), an attacker could infer that an individual was part of the hospital’s training dataset, thereby revealing their likely diabetic status. This demonstrates how MRAD can lead to privacy leakage in real-world scenarios.
Also Read:
- Advancing Private AI: A New Framework for Neural Fields on Edge Devices
- The Hidden Truth: LLMs Deceive Even Without Prompts
Defenses and Future Work
The researchers note that defenses against overfitting, such as early stopping and regularization, can help mitigate these attacks, as overfitting plays a key role in their success. Differential privacy is also highlighted as a provable privacy-preserving framework that can reduce the risk of such inferences.
While MRAD shows promising results, the paper acknowledges limitations, such as its current reliance on a white-box setting (full access to the model). Future work could explore gradient approximation or model stealing techniques to enable black-box attacks. Additionally, improving performance when a very high percentage of features are known, and exploring label-free attacks, are areas for future research.
This research underscores a new dimension in privacy risk assessment, demonstrating that even with incomplete or partial data, membership inference attacks can be successfully mounted, posing significant threats to user confidentiality and data privacy.


