mirageco commited on
Commit
00a2d22
1 Parent(s): c6e846c

Add Decision-Making task. FinTrade.

Browse files
Files changed (1) hide show
  1. src/about.py +42 -39
src/about.py CHANGED
@@ -52,6 +52,7 @@ class Tasks(Enum):
52
  task53 = Task("EFPA", "F1", "EFPA", category="Spanish")
53
  task54 = Task("FinanceES", "F1", "FinanceES", category="Spanish")
54
  task55 = Task("TSA-Spanish", "F1", "TSA-Spanish", category="Spanish")
 
55
 
56
  NUM_FEWSHOT = 0 # Change with your few shot
57
  # ---------------------------------------------------
@@ -100,45 +101,47 @@ If the icon is "?", it indicates that there is insufficient information about th
100
 
101
  Our evaluation metrics include, but are not limited to, Accuracy, F1 Score, ROUGE score, BERTScore, and Matthews correlation coefficient (MCC), providing a multidimensional assessment of model performance. Metrics for specific tasks are as follows:
102
 
103
- - **FPB**: F1, Accuracy. Financial PhraseBank classification task.
104
- - **FiQA-SA**: F1. Sentiment analysis on FiQA financial domain.
105
- - **TSA**: F1, Accuracy. Sentiment analysis.
106
- - **Headlines**: AvgF1. News headline classification.
107
- - **FOMC**: F1, Accuracy. Hawkish-dovish classification.
108
- - **FinArg-ACC**: F1, Accuracy. Financial argument unit classification.
109
- - **FinArg-ARC**: F1, Accuracy. Financial argument relation classification.
110
- - **MultiFin**: F1, Accuracy. Multi-class financial sentiment analysis.
111
- - **MA**: F1, Accuracy. Deal completeness classification.
112
- - **MLESG**: F1, Accuracy. ESG issue identification.
113
- - **NER**: EntityF1. Named entity recognition in financial texts.
114
- - **FINER-ORD**: EntityF1. Ordinal classification in financial NER.
115
- - **FinRED**: F1, EntityF1. Financial relation extraction from text.
116
- - **SC**: F1, EntityF1. Causal classification task in the financial domain.
117
- - **CD**: F1, EntityF1. Causal detection.
118
- - **FinQA**: EmAcc. Numerical question answering in finance.
119
- - **TATQA**: F1, EmAcc. Table-based question answering in financial documents.
120
- - **ConvFinQA**: EmAcc. Multi-turn question answering in finance.
121
- - **FNXL**: F1, EmAcc. Numeric labeling in financial texts.
122
- - **FSRL**: F1, EmAcc. Financial statement relation linking.
123
- - **EDTSUM**: ROUGE, BERTScore, BARTScore. Extractive document summarization in finance.
124
- - **ECTSUM**: ROUGE, BERTScore, BARTScore. Extractive content summarization.
125
- - **BigData22**: Accuracy, MCC. Stock movement prediction.
126
- - **ACL18**: Accuracy, MCC. Financial news-based stock prediction.
127
- - **CIKM18**: Accuracy, MCC. Financial market prediction using news.
128
- - **German**: F1, MCC. Credit scoring in the German market.
129
- - **Australian**: F1, MCC. Credit scoring in the Australian market.
130
- - **LendingClub**: F1, MCC. Peer-to-peer lending risk prediction.
131
- - **ccf**: F1, MCC. Credit card fraud detection.
132
- - **ccfraud**: F1, MCC. Credit card transaction fraud detection.
133
- - **polish**: F1, MCC. Credit risk prediction in the Polish market.
134
- - **taiwan**: F1, MCC. Credit risk prediction in the Taiwanese market.
135
- - **portoseguro**: F1, MCC. Claim analysis in the Brazilian market.
136
- - **travelinsurance**: F1, MCC. Travel insurance claim prediction.
137
- - **MultiFin-ES**: F1. Multi-class financial sentiment analysis in Spanish.
138
- - **EFP**: F1. Financial phrase classification in Spanish.
139
- - **EFPA**: F1. Financial argument classification in Spanish.
140
- - **FinanceES**: F1. Financial sentiment classification in Spanish.
141
- - **TSA-Spanish**: F1. Sentiment analysis in Spanish.
 
 
142
 
143
 
144
  To ensure a fair and unbiased assessment of the models' true capabilities, all evaluations are conducted in zero-shot settings (0-shots). This approach eliminates any potential advantage from task-specific fine-tuning, providing a clear indication of how well the models can generalize to new tasks.
 
52
  task53 = Task("EFPA", "F1", "EFPA", category="Spanish")
53
  task54 = Task("FinanceES", "F1", "FinanceES", category="Spanish")
54
  task55 = Task("TSA-Spanish", "F1", "TSA-Spanish", category="Spanish")
55
+ task56 = Task("FinTrade", "CR", "FinTrade", category="Decision-Making (DM)")
56
 
57
  NUM_FEWSHOT = 0 # Change with your few shot
58
  # ---------------------------------------------------
 
101
 
102
  Our evaluation metrics include, but are not limited to, Accuracy, F1 Score, ROUGE score, BERTScore, and Matthews correlation coefficient (MCC), providing a multidimensional assessment of model performance. Metrics for specific tasks are as follows:
103
 
104
+ - **FPB**: F1, Accuracy. Financial PhraseBank classification task. This dataset is from the Financial PhraseBank, containing annotated phrases used in financial contexts. The classification task involves determining sentiment (positive, negative, neutral) for each phrase, essential for understanding financial news and reports.
105
+ - **FiQA-SA**: F1. Sentiment analysis on FiQA financial domain. Derived from the FiQA dataset, this task focuses on sentiment analysis in the financial domain, particularly within news and social media. The dataset is crucial for gauging market sentiment based on financial communications.
106
+ - **TSA**: F1, RMSE. Sentiment analysis on social media. The TSA dataset is utilized to analyze sentiment from tweets related to financial markets. The dataset is essential for real-time sentiment analysis, providing insights into market trends influenced by public opinion.
107
+ - **Headlines**: AvgF1. News headline classification. This dataset consists of financial news headlines, with each headline categorized into various financial events or sentiment classes. The task challenges models to understand and classify brief, context-rich text segments that drive market movements.
108
+ - **FOMC**: F1, Accuracy. Hawkish-dovish classification. Derived from transcripts of the Federal Open Market Committee (FOMC) meetings, this dataset involves classifying statements as hawkish or dovish, which indicates the stance of monetary policy. Accurate classification helps predict market reactions to central bank communications.
109
+ - **FinArg-ACC**: F1, Accuracy. Financial argument unit classification. This dataset involves the classification of argument units in financial documents, such as identifying the main claim, supporting evidence, or counterarguments. The task is crucial for automated financial document analysis, enabling the extraction of structured information from unstructured text.
110
+ - **FinArg-ARC**: F1, Accuracy. Financial argument relation classification. This task focuses on classifying relationships between different argument units within financial texts, such as support, opposition, or neutrality. Understanding these relations is critical for constructing coherent financial narratives from fragmented data.
111
+ - **MultiFin**: F1, Accuracy. Multi-class financial sentiment analysis. The MultiFin dataset includes diverse financial texts requiring sentiment classification across multiple categories, such as bullish, bearish, or neutral. The task is pivotal for analyzing sentiment in financial markets from varied sources like reports, news articles, and social media.
112
+ - **MA**: F1, Accuracy. Deal completeness classification. The dataset revolves around classifying mergers and acquisitions (M&A) reports to determine whether a deal has been completed. The task helps in tracking and analyzing the outcomes of corporate transactions, which is key for investment decisions.
113
+ - **MLESG**: F1, Accuracy. ESG issue identification. This dataset focuses on identifying Environmental, Social, and Governance (ESG) issues within financial texts. Models are evaluated on their ability to correctly classify and categorize ESG-related content, which is increasingly important for responsible investing.
114
+ - **NER**: EntityF1. Named entity recognition in financial texts. This task involves identifying and classifying named entities (e.g., companies, financial instruments, persons) within financial documents. Accurate NER is crucial for information extraction and financial analysis automation.
115
+ - **FINER-ORD**: EntityF1. Ordinal classification in financial NER. This dataset extends standard NER by requiring models to classify entities not just by type but also by their ordinal relevance (e.g., primary, secondary importance) within the text. This is useful for prioritizing information in financial summaries.
116
+ - **FinRED**: F1, EntityF1. Financial relation extraction from text. The task involves extracting relationships between financial entities, such as ownership, acquisition, or partnership relations. This is important for building knowledge graphs and conducting in-depth financial analysis.
117
+ - **SC**: F1, EntityF1. Causal classification task in the financial domain. The dataset requires models to classify causal relationships in financial texts, such as determining whether one event causes another. Understanding causality is critical for risk assessment and decision-making in finance.
118
+ - **CD**: F1, EntityF1. Causal detection. Similar to SC, but focused on detecting causality in a broader range of financial texts, including reports, news, and social media. The task evaluates the model's ability to identify causal links, which are key drivers in financial analysis.
119
+ - **FinQA**: EmAcc. Numerical question answering in finance. FinQA involves answering numerical questions based on financial documents, such as balance sheets or income statements. The task tests a model's ability to perform calculations or identify numerical data in a text.
120
+ - **TATQA**: F1, EmAcc. Table-based question answering in financial documents. This task is centered around answering questions that require interpreting and extracting information from tables in financial documents. It's crucial for automating the analysis of structured financial data.
121
+ - **ConvFinQA**: EmAcc. Multi-turn question answering in finance. ConvFinQA extends standard QA tasks by requiring models to handle multi-turn dialogues, where each question builds on the previous one. This simulates real-world scenarios where financial analysts ask a series of related questions.
122
+ - **FNXL**: F1, EmAcc. Numeric labeling in financial texts. This dataset requires models to label numeric values within financial documents, categorizing them by type (e.g., revenue, profit) and relevance. It tests the model's ability to understand the role of numbers in financial contexts.
123
+ - **FSRL**: F1, EmAcc. Financial statement relation linking. The task involves linking related information across different financial statements, such as matching revenue figures from income statements with corresponding cash flow data. This is key for comprehensive financial analysis.
124
+ - **EDTSUM**: ROUGE, BERTScore, BARTScore. Extractive document summarization in finance. The dataset involves summarizing lengthy financial documents by extracting the most relevant sentences. This task evaluates a model's ability to generate concise summaries that retain critical information.
125
+ - **ECTSUM**: ROUGE, BERTScore, BARTScore. Extractive content summarization. Similar to EDTSUM, but with a broader focus on summarizing content from various financial document types, including reports, articles, and regulatory filings.
126
+ - **BigData22**: Accuracy, MCC. Stock movement prediction. This dataset is used for predicting stock price movements based on financial news and reports. The task evaluates a model's ability to forecast market trends, which is essential for investment strategies.
127
+ - **ACL18**: Accuracy, MCC. Financial news-based stock prediction. The ACL18 dataset focuses on predicting stock movements specifically using news headlines and articles. It's a benchmark for evaluating the impact of news on stock prices.
128
+ - **CIKM18**: Accuracy, MCC. Financial market prediction using news. This task involves predicting broader market movements, such as indices, based on financial news. It tests the model's ability to aggregate and interpret multiple sources of financial information.
129
+ - **German**: F1, MCC. Credit scoring in the German market. The dataset includes data on loan applicants in Germany, with the task being to predict creditworthiness. This is important for financial institutions in assessing loan risks.
130
+ - **Australian**: F1, MCC. Credit scoring in the Australian market. Similar to the German dataset, but tailored for the Australian financial context, this task evaluates the model's ability to predict credit risk in this specific market.
131
+ - **LendingClub**: F1, MCC. Peer-to-peer lending risk prediction. This dataset involves predicting the risk of default for loans issued through the LendingClub platform, which is a major peer-to-peer lending service. The task is crucial for risk management in alternative finance.
132
+ - **ccf**: F1, MCC. Credit card fraud detection. The dataset is used to identify fraudulent transactions within a large dataset of credit card operations. Accurate detection is critical for financial security and fraud prevention.
133
+ - **ccfraud**: F1, MCC. Credit card transaction fraud detection. Similar to the ccf dataset but focusing on transaction-level analysis, this task evaluates the model's ability to detect anomalies that indicate fraud.
134
+ - **polish**: F1, MCC. Credit risk prediction in the Polish market. This task involves predicting the likelihood of default for loan applicants in Poland, with the dataset tailored to local economic and financial conditions.
135
+ - **taiwan**: F1, MCC. Credit risk prediction in the Taiwanese market. Similar to the Polish dataset but focused on Taiwan, this task evaluates the model's ability to assess credit risk in this market.
136
+ - **portoseguro**: F1, MCC. Claim analysis in the Brazilian market. The dataset involves predicting insurance claim risks in Brazil, specifically for auto insurance. The task tests the model's ability to assess and manage insurance risks.
137
+ - **travelinsurance**: F1, MCC. Travel insurance claim prediction. This dataset is used for predicting the likelihood of a travel insurance claim being made, which is important for risk pricing and policy management in the travel insurance industry.
138
+ - **MultiFin-ES**: F1. Multi-class financial sentiment analysis in Spanish. This dataset is used to analyze sentiment in Spanish-language financial texts. It evaluates the model's ability to handle sentiment classification across multiple categories in a non-English context.
139
+ - **EFP**: F1. Financial phrase classification in Spanish. Similar to the FPB dataset but in Spanish, this task involves classifying financial phrases according to sentiment or intent, specifically for Spanish-language content.
140
+ - **EFPA**: F1. Financial argument classification in Spanish. This dataset requires the classification of arguments in Spanish financial documents, focusing on identifying claims, evidence, and other argumentative structures.
141
+ - **FinanceES**: F1. Financial sentiment classification in Spanish. The task involves classifying sentiment in a broad range of Spanish financial documents, including news articles and reports. It tests the model's ability to adapt sentiment analysis techniques to a non-English language.
142
+ - **TSA-Spanish**: F1. Sentiment analysis in Spanish. This dataset involves sentiment analysis on Spanish-language tweets and short texts, similar to the English TSA dataset but tailored for Spanish speakers. It evaluates the model's ability to process and analyze sentiment in social media content.
143
+ - **FinTrade**: CR, SR, DV, AV, MD. Stock trading dataset. FinTrade is a novel dataset developed specifically for evaluating stock trading tasks using LLMs. It incorporates historical stock prices, financial news, and sentiment data from 10 different stocks over a year. This dataset is designed to simulate real-world trading scenarios, allowing models to perform agent-based financial trading. The task evaluates the models on multiple financial metrics such as Cumulative Return (CR), Sharpe Ratio (SR), Daily Volatility (DV), Annualized Volatility (AV), and Maximum Drawdown (MD). These metrics provide a comprehensive assessment of the model's profitability, risk management, and decision-making capabilities.
144
+
145
 
146
 
147
  To ensure a fair and unbiased assessment of the models' true capabilities, all evaluations are conducted in zero-shot settings (0-shots). This approach eliminates any potential advantage from task-specific fine-tuning, providing a clear indication of how well the models can generalize to new tasks.