|
Written on 15.04.26 by Ziqing Yang
Dear all,
The paper list is online. Please select three papers (ranked by preference) and send them to Ziqing Yang (ziqing.yang@cispa.de) by 17.04.2026.
Note that the assignment will be based on the first-come, first-served principle.
The assignment will be informed by… Read more
Dear all,
The paper list is online. Please select three papers (ranked by preference) and send them to Ziqing Yang (ziqing.yang@cispa.de) by 17.04.2026.
Note that the assignment will be based on the first-come, first-served principle.
The assignment will be informed by 20.04.2026.
Best,
Ziqing
Paper List
- HATEDAY: Insights from a Global Hate Speech Dataset Representative of a Day on Twitter
- Content Moderation and Hate Speech on Alternative Platforms: A Case Study of BitChute
- Investigating Moderation Challenges to Combating Hate and Harassment: The Case of Mod-Admin Power Dynamics and Feature Misuse on Reddit
- Understanding the Security and Privacy Implications of Online Toxic Content on Refugees
- Supporting Human Raters with the Detection of Harmful Content using Large Language Models
- HateBench: Benchmarking Hate Speech Detectors on LLM-Generated Content and Hate Campaigns
- Moderating Illicit Online Image Promotion for Unsafe User-Generated Content Games Using Large Vision-Language Models
- You Know What I Meme: Enhancing People’s Understanding and Awareness of Hateful Memes Using Crowdsourced Explanations
- Exploring fake news awareness and trust in the age of social media among university student TikTok users
- A Representative Study on Human Detection of Artificially Generated Media Across Countries
- "Better Be Computer or I'm Dumb": A Large-Scale Evaluation of Humans as Audio Deepfake Detectors
- Combating misinformation in the age of LLMs: Opportunities and challenges
- Machine-Made Media: Monitoring the Mobilization of Machine-Generated Articles on Misinformation and Mainstream News Websites
- You Only Prompt Once: On the Capabilities of Prompt Learning on Large Language Models to Tackle Toxic Content.
- Breaking Bad Tokens: Detoxification of LLMs Using Sparse Autoencoder
- from Benign import Toxic: Jailbreaking the Language Model via Adversarial Metaphors.
- Toxicity Detection for Free.
- Robust Adaptation of Large Multimodal Models for Augmented Hateful Meme Detection
- MIND: A Multi-agent Framework for Zero-shot Harmful Meme Detection
- From Meme to Threat: On the Hateful Meme Understanding and Induced Hateful Content Generation in Open-Source Vision Language Models
- On the Proactive Generation of Unsafe Images From Text-To-Image Models Using Benign Prompts
- SneakyPrompt: Jailbreaking Text-to-image Generative Models
- DRES: Fake news detection by dynamic representation and ensemble selection
- Can Large Language Models Identify Authorship?
- Adapting fake news detection to the era of large language models
- Fake News in Sheep's Clothing: Robust Fake News Detection Against LLM-Empowered Style Attacks
- On Fake News Detection with LLM Enhanced Semantics Mining
- Is LLMs Hallucination Usable? LLM-based Negative Reasoning for Fake News Detection
- Generate First, Then Sample: Enhancing Fake News Detection with LLM-Augmented Reinforced Sampling
- Explainable Fake News Detection with Large Language Model via Defense Among Competing Wisdom
- Bad Actor, Good Advisor: Exploring the Role of Large Language Models in Fake News Detection
- Can LLM-Generated Misinformation Be Detected?
|