TLDR: This comprehensive survey reviews graph-based techniques for detecting fake accounts in online social networks. It categorizes existing methods based on their approach (classical, traditional ML, deep learning), features used, and detection time. The paper discusses the strengths and limitations of various techniques, including random walks, probabilistic models, GNNs, contrastive learning, and federated learning. It also investigates available real-world and synthesized datasets and proposes several avenues for future research, such as improving dataset quality, handling heterophily, few-shot learning, and enhancing adversarial robustness and explainability.
Online social networks like Facebook, X (formerly Twitter), and Instagram have transformed how we connect and share information. However, these platforms have also become fertile ground for harmful activities, particularly through the creation of fake accounts. These fake personas, often called bots, sock puppets, or sybils, pose significant threats to the security, trust, and integrity of online interactions. They are used to manipulate public opinion, spread misinformation, commit financial crimes, and inflate engagement metrics, leading to substantial financial losses and undermining democratic processes.
Detecting these fake accounts, especially those empowered by advanced generative AI and deepfake technologies, is a major challenge. While fake accounts can easily mimic real profiles and content, they have less control over their position within the social network’s graph structure. This is because forming connections often requires acceptance from other accounts. Therefore, methods that leverage the underlying graph structure of social networks have gained significant attention, as they have shown to greatly improve detection accuracy.
Categorizing Detection Methods
Fake Account Detection (FAD) methods can be classified in several ways. Based on their methodological approach, they fall into three main categories: Classical methods, which use algorithms like random walks without requiring training; Traditional Machine Learning (ML) methods, which apply algorithms to extracted features; and Deep Learning (DL) methods, particularly those using Graph Neural Networks (GNNs), which learn directly from the graph structure and features to model complex relationships.
Methods can also be categorized by the features they use: graph-based features (like connections and centrality), profile-based features (like usernames and account age), content-based features (like shared text or images), and personal features (like IP addresses). The timing of detection is another key distinction: some methods aim for detection at registration, others for early detection based on initial activity, and more sophisticated ones for detection after extended activity, targeting accounts that have evaded earlier checks.
Classical Approaches to FAD
Early FAD techniques, often called classical methods, rely on graph algorithms and probabilistic mechanisms. Random walk-based methods, like SybilRank and SybilWalk, use the idea that real accounts tend to connect with other real accounts, while fake accounts might be isolated or form distinct clusters. By simulating ‘walks’ through the network, these methods estimate the likelihood of an account being fake based on where these walks end up.
Probabilistic inference methods, such as SybilBelief, use models like Markov Random Fields to iteratively refine the probability of a node being fake or real by passing ‘messages’ between connected nodes. More advanced techniques, like SybilSCAR, combine the strengths of random walks and probabilistic methods to improve efficiency and robustness. Additionally, structural preprocessing methods enhance detection by refining the graph structure, for example, by removing ‘weak’ connections that might link fake accounts to real ones, or by integrating additional interaction data beyond simple friendships.
Traditional Machine Learning and Graph-Based Insights
Traditional ML methods for FAD focus on extracting discriminative features from profiles and graph structures. Simple classifiers like Support Vector Machines (SVM) or Random Forests (RF) can be effective when fed the right features, such as account activity patterns over time or various graph-based metrics like node degree and centrality. Some approaches analyze ‘ego networks’ (subgraphs centered around individual accounts) to identify structural differences between real and fake accounts.
More advanced traditional ML techniques incorporate higher-order graph representations and probabilistic models. Methods like UFA and GUFAD use unsupervised learning, community detection, and graph embeddings to identify clusters of fake accounts based on shared registration attributes or dynamic behavioral patterns. SimilCatch and UnDBot leverage structural entropy to uncover irregular network structures associated with fake accounts, often working with multi-relational graphs that encode different types of interactions.
The Rise of Deep Learning in FAD
Deep Learning (DL) has become a powerful tool for FAD, especially with the advent of Graph Neural Networks (GNNs). GNNs can automatically extract complex patterns from raw data and capture the relational and structural dependencies in social networks. Variants like Graph Convolutional Networks (GCN) and Graph Attention Networks (GAT) have shown improved performance in identifying fake accounts.
DL extends classical methods by integrating GNNs with techniques like random walks (e.g., TrustGCN) or ensemble learning (e.g., RF-GNN, SStackGNN). These hybrid approaches leverage the strengths of both worlds. A significant challenge in FAD is ‘heterophily,’ where fake accounts intentionally connect with real accounts to avoid detection. Heterophily-aware methods, such as BothH and CGNN, explicitly handle these mixed-label connections by adjusting message passing or pruning unreliable edges.
Further advancements in DL for FAD include subgraph-level and community-aware modeling, which extract richer structural patterns by focusing on local communities or specific network motifs. Multi-relational graphs, which represent different types of interactions (e.g., follows, retweets, likes), are also increasingly used with models like Relational Graph Convolutional Networks (RGCNs) to capture the complexity of online social networks.
The integration of content features, such as textual and visual data, has also become crucial. Models like BIC and ETS-MM combine textual semantics (topics, emotions) with graph structures, while multi-modal frameworks integrate CNNs for images and LSTMs for text alongside GNNs for network structure. Contrastive Learning (CL) is another powerful paradigm, enabling robust representation learning by pulling similar accounts closer and pushing dissimilar ones apart in the embedding space, even with limited labels.
Reinforcement Learning (RL) allows FAD models to dynamically adapt their graph sampling strategies and optimize network depth in response to evolving fake account tactics. Temporal methods, like BotDGT, track behavioral changes over time by modeling social networks as sequences of graph snapshots, capturing the evolution of connections and content patterns. Mixture of Experts (MoE) frameworks, such as DSBD and BotMoE, use multiple specialized neural networks to handle diverse account behaviors across different interest categories or communities.
Federated Learning is gaining traction in FAD, offering data privacy by training models across multiple participants without sharing raw data. This is particularly important for sensitive account information. Finally, research into adversarial attacks, where small perturbations to features or graph structure can mislead detection models, highlights the ongoing ‘arms race’ between fake account creators and detection systems. For more in-depth technical details, you can refer to the full research paper available at arXiv:2507.06541.
Datasets for FAD Research
The effectiveness of FAD models heavily relies on the quality and diversity of datasets. Both real-world and synthesized datasets are used. Real-world datasets, often from platforms like Twitter (e.g., Cresci-15, TwiBot-20, TwiBot-22, MGTAB), provide realistic social behaviors and diverse features, including textual content, profile metadata, and graph structures. These are typically labeled manually or semi-automatically.
Synthesized datasets, on the other hand, allow for controlled experimentation and provide explicit ground-truth labels. Researchers can simulate various fake account behaviors and attack strategies, such as random or targeted attacks, to evaluate detection methods under specific conditions. Some synthesis algorithms even build upon unlabeled real-world data to retain realistic characteristics.
Also Read:
- A Comprehensive Overview of Graph Learning: Methods, Challenges, and Future Directions
- Unveiling Key Players: A Comprehensive Look at Critical Node Identification in Complex Networks
Future Directions in FAD
Despite significant progress, several challenges and opportunities remain in FAD research. There’s a need for more high-quality datasets from diverse platforms beyond Twitter, incorporating multiple media modalities (images, audio, video), and providing more granular labels (e.g., distinguishing spammers from bots or data collectors). Addressing ‘heterophilic edges’ more effectively, perhaps by incorporating causality between nodes, is another promising area.
Furthermore, few-shot and zero-shot learning approaches, which can classify new accounts with minimal or no labeled data, are crucial for low-supervision settings. The ‘cold-start’ problem, where new accounts have limited connections, aligns with early detection challenges and requires further investigation. Enhancing the explainability of FAD models is also vital to build trust and ensure responsible deployment, allowing humans to understand why an account is flagged as fake. Finally, continued research into designing realistic and sophisticated adversarial attacks is essential to benchmark and improve the robustness of FAD models against evolving evasion tactics.


