TLDR: Researchers from Georgia Institute of Technology explored advanced techniques for detecting subjective and objective sentences in news text. Their study for CheckThat! Lab at CLEF 2025 demonstrated that using AI models pre-trained on sentiment and emotion data (transfer-learning) significantly improves subjectivity detection. Furthermore, they introduced a novel data augmentation pipeline where GPT-4o generates stylistic paraphrases and then self-corrects any inconsistencies in the generated data, leading to enhanced model robustness and classification performance, particularly for subjective content.
In an era grappling with the rapid spread of misinformation, the development of automated fact-checking systems has become critically important. A key component of such systems is the ability to accurately identify subjective and objective sentences in text, as objective statements can be directly fact-checked, while subjective ones often require further processing to remove opinions or emotions before verification. This challenge was the focus of Task 1, Subjectivity Detection, at the CheckThat! Lab during CLEF 2025.
Researchers from Georgia Institute of Technology, Maximilian Heil and Dionne Bang, presented their work on detecting subjectivity in English news text. Their approach explored the effectiveness of combining transfer-learning techniques with innovative data augmentation strategies, including a unique self-correction mechanism powered by GPT-4o.
Enhancing Detection with Specialized AI Models
The team investigated how different types of pre-trained AI models, known as encoders, perform in distinguishing subjective from objective language. They compared general-purpose encoders like RoBERTa-base and MiniLM-L12-v2 with specialized encoders such as Sentiment-Analysis-BERT and Emotion-English-DistilRoBERTa-base. These specialized models had already been fine-tuned on datasets related to sentiment analysis and emotion recognition, making them inherently better equipped to understand emotional tones and subjective language.
Their findings confirmed that models pre-trained on related tasks significantly outperformed general-purpose models in identifying subjective linguistic cues. This suggests that leveraging AI models with prior exposure to sentiment and emotion data can greatly improve their sensitivity to subjectivity in news articles.
Smart Data Augmentation with Self-Correction
Given the often limited size of labeled datasets for training, the researchers also explored the benefits of synthetically expanding their training data. They used GPT-4o, a large language model, to generate paraphrases of existing sentences, creating both subjective and objective versions to enrich the dataset. For instance, an objective sentence like “The trend is expected to reverse as soon as next month” could be transformed into a subjective one like “A promising turnaround is on the horizon, with expectations for change as early as next month.”
However, simply generating more data isn’t always enough. The team discovered that uncorrected augmented data could sometimes introduce inconsistencies, where generated sentences didn’t perfectly align with their intended subjective or objective labels. To address this, they introduced a crucial second-stage validation and correction pipeline. Using GPT-4o again, this pipeline automatically reviewed each generated sentence. If a sentence didn’t match its assigned label or stylistic intent (e.g., propaganda, emotional), GPT-4o would rewrite it to ensure consistency, while preserving the original subject matter.
This self-correction mechanism proved vital. While naive data augmentation did not consistently improve performance, the corrected augmented datasets led to significant gains in classification accuracy, particularly for detecting subjective content. This highlights that the quality and consistency of synthetic data are as important as the quantity.
Also Read:
- Assessing Large Language Models for Text Summarization: A Deep Dive into Prompt Engineering and Performance
- Unmasking AI: New Research Advances in Detecting and Attributing Machine-Generated Text
Impact and Future Directions
Although the improved models trained on the self-corrected data could not be submitted for the official competition deadline, the post-submission results clearly demonstrated their superior performance. The team’s official submission, using a model fine-tuned on the original dataset, placed 16th out of 24 participants, outperforming the organizers’ baseline. The research underscores the value of combining specialized AI models with carefully curated, high-quality synthetic data for improving performance in tasks like subjectivity detection, especially in scenarios with limited initial data.
Future work includes applying this approach to multilingual contexts, incorporating more labeled data into model fine-tuning, and exploring ensemble methods that combine different models to leverage their individual strengths. This research contributes significantly to the ongoing efforts to build more robust and accurate fact-checking systems. You can read the full research paper at https://arxiv.org/pdf/2507.06189.


