File size: 8,266 Bytes
d56a8cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37e6b02
 
 
 
 
 
 
 
 
 
 
 
 
2c4d061
37e6b02
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9478435
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
---
license: mit
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: test
    path: data/test-*
dataset_info:
  features:
  - name: text
    dtype: string
  - name: context
    dtype: string
  - name: answer_1
    dtype: string
  - name: answer_2
    dtype: string
  - name: subreddit
    dtype: string
  - name: num_comments
    dtype: int64
  - name: score
    dtype: int64
  - name: upvote_ratio
    dtype: float64
  - name: ups
    dtype: float64
  - name: downs
    dtype: float64
  - name: author
    dtype: string
  - name: created_utc
    dtype: int64
  - name: retrieved_on
    dtype: float64
  - name: retrieved_utc
    dtype: float64
  - name: id
    dtype: string
  - name: comment_is_submission
    dtype: bool
  - name: score_answer1
    dtype: int64
  - name: upvote_ratio_answer1
    dtype: 'null'
  - name: ups_answer1
    dtype: float64
  - name: downs_answer1
    dtype: float64
  - name: author_answer1
    dtype: string
  - name: name_answer1
    dtype: string
  - name: id_answer1
    dtype: string
  - name: created_utc_answer1
    dtype: int64
  - name: retrieved_on_answer1
    dtype: float64
  - name: retrieved_utc_answer1
    dtype: float64
  - name: score_answer2
    dtype: int64
  - name: upvote_ratio_answer2
    dtype: 'null'
  - name: ups_answer2
    dtype: float64
  - name: downs_answer2
    dtype: float64
  - name: author_answer2
    dtype: string
  - name: name_answer2
    dtype: string
  - name: id_answer2
    dtype: string
  - name: created_utc_answer2
    dtype: int64
  - name: retrieved_on_answer2
    dtype: float64
  - name: retrieved_utc_answer2
    dtype: float64
  splits:
  - name: train
    num_bytes: 116393440
    num_examples: 53061
  - name: test
    num_bytes: 1076885
    num_examples: 500
  download_size: 73381121
  dataset_size: 117470325
task_categories:
- question-answering
language:
- en
tags:
- finance
size_categories:
- 10K<n<100K
---
# Dataset Card for Dataset Name

<!-- Provide a quick summary of the dataset. -->

SocialFinanceQA is the first social-media preference dataset for fine-tuning and aligning LLMs for the finance domain.

## Dataset Details

### Dataset Description

<!-- Provide a longer summary of what this dataset is. -->

- **Curated by:** Kris-Fillip Kahl, Tolga Buz, Russa Biswas, Gerard de Melo
- **Language(s) (NLP):** English
- **License:** MIT

### Dataset Sources

- **Repository:** https://github.com/kris-fillip/SocialFinanceQA
<!-- - **Paper [optional]:** [More Information Needed] -->

## Uses

<!-- Address questions around how the dataset is intended to be used. -->
This dataset is intended to be used for fine-tuning and aligning LLMs for the task of Long Form Question Answering (LFQA).

### Direct Use

<!-- This section describes suitable use cases for the dataset. -->
This dataset is suitable for supervised fine-tuning as well as preference alignment of LLMs.

## Dataset Structure

<!-- This section provides a description of the dataset fields, and additional information about the dataset structure such as criteria used to create the splits, relationships between data points, etc. -->

The dataset consists of questions from submissions and corresponding answers from comments within the selected subreddits.
It is a preference dataset constructed based on the scores (upvotes and downvotes) of the individual submissions and comments within their subreddits.
All fields entailing answer_1 correspond to the preferred answer for a specific question, while the answer_2 fields correspond to the non-preferred answer.

## Dataset Creation

### Curation Rationale

<!-- Motivation for the creation of this dataset. -->

Retail investing is on the rise, and a growing number of users are relying on online finance communities to educate themselves.
However, recent years have positioned Large Language Models (LLMs) as powerful question-answering (QA) tools, shifting users away from interacting in communities toward discourse with AI-driven conversational interfaces.
These AI tools are currently limited by the availability of labeled data containing domain-specific financial knowledge.
Therefore, in this work, we curate a QA preference dataset SOCIALFINANCEQA for fine-tuning and aligning LLMs, extracted from over 7.4 million submissions and 82 million comments from 2008 to 2022 in Reddit’s 15 largest finance communities.

### Source Data

<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->

The data originally stems from Reddit submissions and comments made to 15 of the largest financial subreddits (personalfinance, financialindependence, FinancialPlanning, investing, wallstreetbets, Wallstreetbetsnew, stocks, StockMarket, pennystocks, options, RealEstate, Economics, realestateinvesting, AskEconomics, explainlikeimfive) between 2008 and 2022.

#### Data Collection and Processing

<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->


#### Who are the source data producers?

<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->

Reddit users who contributed to the selected financial subreddits between 2008 and 2022.

#### Personal and Sensitive Information

<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->

The dataset contains the usernames of the authors of the texts.
These user names are pseudonyms that do not contain information about the author's identity unless the author actively decides to share personal information in their Reddit profile.
We have not obtained dedicated permission from the authors of the texts to include them in our dataset.
However, this is mitigated by the users agreeing to Reddit’s terms, which allow data extraction services such as Pushshift, which is the source for our dataset and was accessed legitimately.
Additionally, the texts may contain the names of individuals, usually public figures relevant to the stock market, whom the authors included in their questions or responses.
We have decided to leave these names in the dataset, as they are usually critical for understanding the text (e.g. when a company’s CEO is mentioned in a text about how the same company performs).

## Bias, Risks, and Limitations

<!-- This section is meant to convey both technical and sociotechnical limitations. -->

Social media should never be the only source for investors planning to make financial decisions.
We have chosen 15 communities covering various topics to provide a broad foundation.
Nonetheless, we recommend retail investors consult other sources, such as relevant books, scientific literature, and trained professionals, before making significant investment decisions.
Investing often bears the risk of losing money, so investors should be careful with money they cannot afford to lose.
Inappropriate, toxic, or offensive language is always a risk when working with social media data.
We mitigate this by filtering our dataset carefully, leveraging Reddit’s internal mechanisms and additional techniques, such as toxicity detection.
Nonetheless, there is a possibility that our dataset still harbors remaining instances of text that may be considered offensive or may cause fine-tuned LLMs to generate such.

<!-- ## Citation [optional] -->

<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->

<!-- **BibTeX:** -->

<!-- **APA:** -->

<!-- ## Glossary [optional] -->

<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->

## Dataset Card Authors

Kris-Fillip Kahl, Tolga Buz