News
Paper List Available
Written on 15.04.2026 09:16 by Ziqing Yang
Dear all,
The paper list is online. Please select three papers (ranked by preference) and send them to Ziqing Yang (ziqing.yang@cispa.de) by 17.04.2026.
Note that the assignment will be based on the first-come, first-served principle.
The assignment will be informed by 20.04.2026.
Best,
Ziqing
Paper List
- HATEDAY: Insights from a Global Hate Speech Dataset Representative of a Day on Twitter
- Content Moderation and Hate Speech on Alternative Platforms: A Case Study of BitChute
- Investigating Moderation Challenges to Combating Hate and Harassment: The Case of Mod-Admin Power Dynamics and Feature Misuse on Reddit
- Understanding the Security and Privacy Implications of Online Toxic Content on Refugees
- Supporting Human Raters with the Detection of Harmful Content using Large Language Models
- HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns
- Moderating Illicit Online Image Promotion for Unsafe User-Generated Content Games Using Large Vision-Language Models
- You Know What I Meme: Enhancing People’s Understanding and Awareness of Hateful Memes Using Crowdsourced Explanations
- Exploring fake news awareness and trust in the age of social media among university student TikTok users
- A Representative Study on Human Detection of Artificially Generated Media Across Countries
- "Better Be Computer or I'm Dumb": A Large-Scale Evaluation of Humans as Audio Deepfake Detectors
- Combating misinformation in the age of LLMs: Opportunities and challenges
- Machine-Made Media: Monitoring the Mobilization of Machine-Generated Articles on Misinformation and Mainstream News Websites
- You Only Prompt Once: On the Capabilities of Prompt Learning on Large Language Models to Tackle Toxic Content.
- Breaking Bad Tokens: Detoxification of LLMs Using Sparse Autoencoder
- from Benign import Toxic: Jailbreaking the Language Model via Adversarial Metaphors.
- Toxicity Detection for Free.
- Robust Adaptation of Large Multimodal Models for Augmented Hateful Meme Detection
- MIND: A Multi-agent Framework for Zero-shot Harmful Meme Detection
- From Meme to Threat: On the Hateful Meme Understanding and Induced Hateful Content Generation in Open-Source Vision Language Models
- On the Proactive Generation of Unsafe Images From Text-To-Image Models Using Benign Prompts
- SneakyPrompt: Jailbreaking Text-to-image Generative Models
- DRES: Fake news detection by dynamic representation and ensemble selection
- Can Large Language Models Identify Authorship?
- Adapting fake news detection to the era of large language models
- Fake News in Sheep's Clothing: Robust Fake News Detection Against LLM-Empowered Style Attacks
- On Fake News Detection with LLM Enhanced Semantics Mining
- Is LLMs Hallucination Usable? LLM-based Negative Reasoning for Fake News Detection
- Generate First, Then Sample: Enhancing Fake News Detection with LLM-Augmented Reinforced Sampling
- Explainable Fake News Detection with Large Language Model via Defense Among Competing Wisdom
- Bad Actor, Good Advisor: Exploring the Role of Large Language Models in Fake News Detection
- Can LLM-Generated Misinformation Be Detected?
