Rumor Stance Detection (PHEME)
Classify stance toward rumors in social media threads. Based on PHEME (Zubiaga et al.). Label replies as supporting, denying, querying, or commenting on rumorous claims.
配置文件config.yaml
# Rumor Stance Detection (PHEME)
# Based on Zubiaga et al., EMNLP 2018
# Paper: https://aclanthology.org/D18-1401/
# Dataset: https://figshare.com/articles/dataset/PHEME_dataset
#
# This task classifies how users respond to rumors on social media.
# Understanding stance helps assess rumor veracity and spread patterns.
#
# Stance Labels (SDQC):
# - Support: The reply endorses or agrees with the rumor
# - Deny: The reply refutes or contradicts the rumor
# - Query: The reply questions the rumor's veracity
# - Comment: The reply discusses the rumor without taking a stance
#
# Annotation Guidelines:
# 1. Read the original rumor (source claim) first
# 2. Read the reply in context of the conversation thread
# 3. Focus on the reply's stance toward the SOURCE rumor
# 4. Support: Agrees, shares, adds confirming information
# 5. Deny: Disagrees, corrects, provides counter-evidence
# 6. Query: Asks for evidence, expresses doubt, questions truth
# 7. Comment: Neutral discussion, jokes, tangential remarks
#
# Note: A reply might support the rumor while denying someone else's
# denial. Focus on stance toward the ORIGINAL claim.
annotation_task_name: "Rumor Stance Detection"
task_dir: "."
data_files:
- sample-data.json
item_properties:
id_key: "id"
text_key: "reply"
output_annotation_dir: "annotation_output/"
output_annotation_format: "json"
annotation_schemes:
# Step 1: Stance classification
- annotation_type: radio
name: stance
description: "What is the stance of this REPLY toward the RUMOR?"
labels:
- "Support"
- "Deny"
- "Query"
- "Comment"
tooltips:
"Support": "The reply endorses, agrees with, or spreads the rumor as true"
"Deny": "The reply refutes, contradicts, or provides evidence against the rumor"
"Query": "The reply questions the rumor, asks for evidence, or expresses doubt"
"Comment": "The reply discusses the topic without taking a clear stance on truth/falsity"
# Step 2: Certainty
- annotation_type: likert
name: certainty
description: "How certain does the reply author seem about their stance?"
min_value: 1
max_value: 5
labels:
1: "Very uncertain"
2: "Somewhat uncertain"
3: "Neutral"
4: "Somewhat certain"
5: "Very certain"
tooltips:
1: "Expresses strong doubt or hedging"
2: "Shows some hesitation"
3: "Neutral tone"
4: "Fairly confident expression"
5: "Expresses strong conviction"
# Step 3: Confidence in annotation
- annotation_type: likert
name: confidence
description: "How confident are you in your stance classification?"
min_value: 1
max_value: 5
labels:
1: "Very uncertain"
2: "Somewhat uncertain"
3: "Moderately confident"
4: "Confident"
5: "Very confident"
allow_all_users: true
instances_per_annotator: 100
annotation_per_instance: 3
allow_skip: true
skip_reason_required: false
示例数据sample-data.json
[
{
"id": "rum_001",
"rumor": "Breaking: Major earthquake reported in Los Angeles, buildings collapsing across the city.",
"reply": "My cousin lives there and says everything is fine. This is fake news."
},
{
"id": "rum_002",
"rumor": "Breaking: Major earthquake reported in Los Angeles, buildings collapsing across the city.",
"reply": "OMG I hope everyone is okay! Sharing this so people can stay safe."
}
]
// ... and 8 more items获取此设计
Clone or download from the repository
快速开始:
git clone https://github.com/davidjurgens/potato-showcase.git cd potato-showcase/text/argumentation-stance/rumor-stance potato start config.yaml
详情
标注类型
领域
应用场景
标签
发现问题或想改进此设计?
提交 Issue相关设计
Clickbait Detection (Webis Clickbait Corpus)
Classify headlines and social media posts as clickbait or non-clickbait based on the Webis Clickbait Corpus. Identify manipulative content designed to attract clicks through sensationalism, curiosity gaps, or misleading framing.
Deceptive Review Detection
Distinguish between truthful and deceptive (fake) reviews. Based on Ott et al., ACL 2011. Identify fake reviews written to deceive vs genuine customer experiences.
Dynamic Hate Speech Detection
Hate speech classification with fine-grained type labels based on the Dynamically Generated Hate Speech Dataset (Vidgen et al., ACL 2021). Classify content as hateful or not, then identify hate type (animosity, derogation, dehumanization, threatening, support for hateful entities) and target group.