Detecting Misinformation, Bias & Hostile Rhetoric in Political Communication Using NLP and Deep Learning
This project analyzes and flags manipulative language in political discourseβspanning speeches, tweets, and media articles. Using advanced Natural Language Processing (NLP) and Transformer-based Deep Learning models, it classifies political statements for truthfulness, detects manipulative or hostile rhetoric, analyzes media bias, and extracts key rhetorical themes.
β¨ Real-time, multi-angle detection pipeline revealing how and why political speech may be misleading, emotionally charged, or biased.
- π Fact-Check Claims: Classify political statements as true, false, or misleading.
- π¬ Rhetoric Detection: Identify manipulation tactics (e.g., emotional appeal, nationalism, repetition).
- π§ Hostile Speech Analysis: Detect personal attacks and hate speech in tweets.
- π° Media Bias Framing: Analyze partisan leanings and framing in political news.
- π Theme Extraction: Uncover core themes and ideological messaging in speeches.
π SPEECH 1 - MANIPULATION ANALYSIS REPORT
π Overall Sentiment: π‘ Negative (-0.46)
β οΈ Fact-Check: Likely False Statement Detected
π Detected Tactics:
β
Repetition | β
Emotional Appeal | β
Personal Attacks | β
Nationalism | β
Misinformation
π§ Verdict: The speech contains manipulative or misleading content with emotionally charged, divisive language.
| Dataset | Description | Use Case |
|---|---|---|
| LIAR | 12.8K fact-checked political statements | Fake news classification |
| Trump Speeches | 45+ official transcripts | Rhetorical + sentiment analysis |
| Trump Tweets | 5.6K tweets labeled for hostility | Hate speech / insult detection |
| All The News | 140K+ news articles (2016β2022) | Media bias & framing analysis |
- Named Entity Recognition (NER)
- Tokenization, Lemmatization
- Sentiment & Emotion Analysis
- TF-IDF & Hate Speech Lexicons
- Entity Co-occurrence Mapping
- Logistic Regression, SVM, XGBoost, Random Forest
- Topic Modeling: LDA for ideological themes and topics
- BERT / BERT-Large: Context-aware classification
- DistilBERT, RoBERTa: Light & interpretable models
- BiLSTM: Sequential pattern recognition
- Hybrid Architectures (e.g., BERT + BiLSTM)
| Task | Best Model | Score / Insight |
|---|---|---|
| β Fake News Detection | Logistic Regression | 88.8% Accuracy |
| π¨ Insult Detection | BERT + BiLSTM | 93.6% F1 Score |
| π° Media Bias Detection | Co-occurrence Analysis | Based on framing patterns (heuristic) |
| π Theme Extraction | LDA + Sentiment | Topics: Witch Hunt, Fake News, Elections |
| π¬ Deep Rhetoric Class. | DistilBERT (Trial) | 88.8% Accuracy |
Natural Language Understanding, Transformer Fine-Tuning, Text Classification, Sentiment Analysis, Rhetoric Detection, Topic Modeling, Explainable AI, Data Visualization, and Model Interpretation (SHAP).
- Multi-language political discourse analysis π
- Real-time Twitter/News feed monitoring π
- Dashboard with dynamic visualizations π