Financial PhraseBank - Sentiment Classification
Sentiment classification of financial news sentences based on the Financial PhraseBank dataset (Malo et al., JASIST 2014). Annotators classify sentences from financial news articles into fine-grained and coarse sentiment categories.
配置文件config.yaml
# Financial PhraseBank - Sentiment Classification
# Based on Malo et al., JASIST 2014
# Paper: https://arxiv.org/abs/1307.5336
# Dataset: https://huggingface.co/datasets/financial_phrasebank
#
# This task classifies the sentiment of financial news sentences.
# Annotators assign both a fine-grained sentiment level (5-point scale)
# and a coarse-grained sentiment label (positive, negative, neutral)
# to each sentence from financial news articles.
#
# Fine-Grained Sentiment:
# - Strong Positive: Clear positive impact on company/market
# - Moderate Positive: Mild positive implications
# - Neutral: No sentiment or balanced positive/negative
# - Moderate Negative: Mild negative implications
# - Strong Negative: Clear negative impact on company/market
#
# Annotation Guidelines:
# 1. Read the sentence carefully in its financial context
# 2. Select the fine-grained sentiment level
# 3. Select the coarse-grained overall sentiment
# 4. Focus on investor/market perspective, not general sentiment
annotation_task_name: "Financial PhraseBank - Sentiment Classification"
task_dir: "."
data_files:
- sample-data.json
item_properties:
id_key: "id"
text_key: "text"
output_annotation_dir: "annotation_output/"
output_annotation_format: "json"
port: 8000
server_name: localhost
annotation_schemes:
- annotation_type: select
name: fine_grained_sentiment
description: "Select the fine-grained sentiment level of this financial sentence"
labels:
- "Strong Positive"
- "Moderate Positive"
- "Neutral"
- "Moderate Negative"
- "Strong Negative"
- annotation_type: radio
name: overall_sentiment
description: "What is the overall sentiment of this financial sentence?"
labels:
- "Positive"
- "Negative"
- "Neutral"
keyboard_shortcuts:
"Positive": "1"
"Negative": "2"
"Neutral": "3"
tooltips:
"Positive": "The sentence conveys positive financial news or outlook"
"Negative": "The sentence conveys negative financial news or outlook"
"Neutral": "The sentence is neutral or purely factual without sentiment"
annotation_instructions: |
You will be shown a sentence from a financial news article. Your task is to:
1. Determine the fine-grained sentiment (Strong Positive to Strong Negative).
2. Assign the overall sentiment label (Positive, Negative, or Neutral).
Important: Judge sentiment from an investor or market perspective, not general sentiment.
For example, "layoffs" might be negative generally but could be positive for investors
if it signals cost-cutting measures.
html_layout: |
<div style="padding: 15px; max-width: 800px; margin: auto;">
<div style="background: #f0f9ff; border: 1px solid #bae6fd; border-radius: 8px; padding: 16px; margin-bottom: 16px;">
<strong style="color: #0369a1;">Financial News Sentence:</strong>
<p style="font-size: 16px; line-height: 1.7; margin: 8px 0 0 0;">{{text}}</p>
</div>
</div>
allow_all_users: true
instances_per_annotator: 50
annotation_per_instance: 2
allow_skip: true
skip_reason_required: false
示例数据sample-data.json
[
{
"id": "finphrase_001",
"text": "The company reported a 25% increase in quarterly revenue, exceeding analyst expectations by a significant margin."
},
{
"id": "finphrase_002",
"text": "Operating profit fell to EUR 35.4 million from EUR 68.1 million in the corresponding period last year."
}
]
// ... and 8 more items获取此设计
Clone or download from the repository
快速开始:
git clone https://github.com/davidjurgens/potato-showcase.git cd potato-showcase/text/financial/financial-phrasebank-sentiment potato start config.yaml
详情
标注类型
领域
应用场景
标签
发现问题或想改进此设计?
提交 Issue相关设计
KG-BERT Knowledge Graph Triple Validation
Validate knowledge graph triples for correctness and annotate relation types based on the KG-BERT framework. Annotators assess whether entity-relation-entity triples are valid, classify the relation type, and provide entity descriptions.
MS MARCO - Passage Relevance Ranking
Passage relevance ranking based on the MS MARCO dataset (Nguyen et al., NeurIPS 2016 Workshop). Annotators assess the relevance of a candidate passage to a given search query using a graded relevance scale.
SemEval-2007 - Word Sense Disambiguation
Word sense disambiguation task based on the SemEval-2007 English lexical sample (Pradhan et al.). Annotators identify the correct sense of a target word in context from a provided list of sense definitions.