Model Card for Meta LLAMA2-7B Custom Task-Oriented Chatbot

This model is a fine-tuned version of Meta's LLAMA2-7B model, adapted to function as a task-oriented chatbot that processes and answers questions related to financial 10K reports.

Model Details

Model Description

Developed by Yatharth Mahesh Sant, this model is a causal language model fine-tuned from Meta's LLAMA2-7B to specifically handle queries and tasks associated with 10K financial reports. It is designed to assist financial analysts and stakeholders by providing detailed, accurate answers to inquiries about company performances, financial standings, and other key metrics contained within 10K reports.

  • Developed by: Yatharth Mahesh Sant
  • Model type: Causal LM
  • Language(s) (NLP): English
  • Finetuned from: meta/llama2-7b
  • Repository: Meta LLAMA2-7B

Uses

Intended Use

This model is meant to be used as a task-oriented bot to interact with users querying about details in 10K financial reports, enhancing the efficiency of financial analysis and decision-making processes.

Direct Use

The model can directly answer questions from financial reports, serving as an automated assistant to financial analysts, investors, and regulatory authorities who require quick, reliable interpretations of financial data.

Downstream Use

The model can be integrated into financial analysis software, used to power internal data review tools in corporations, or serve as a support system in investor relations departments to automate responses to common shareholder inquiries.

Out-of-Scope Use

This model is not designed for non-financial texts or languages other than English. It may not perform well in informal conversational settings or handle off-topic inquiries effectively.

Bias, Risks, and Limitations

The model's performance and responses are based on the data it was trained on, which primarily includes structured financial texts. As such, it may inherit biases from this data or fail to comprehend nuanced questions not directly related to financial reporting.

Recommendations

It is recommended that responses generated by this model be reviewed by a qualified financial analyst to confirm their accuracy before being used in critical decision-making processes.

How to Get Started with the Model

To get started with this model, please refer to the specific deployment guides and API documentation provided in the repository linked above.

Training Details

Training Data

The model was fine-tuned on a comprehensive dataset comprising several years' worth of 10K reports from companies across various industries, annotated for key financial metrics and queries.

Training Procedure

Preprocessing

Training data was preprocessed to normalize financial terminology and remove any non-relevant sections of the reports, focusing on the sections most pertinent to common queries.

Training Hyperparameters

The model was trained using a learning rate of 5e-5 with a batch size of 32 over 4 epochs, employing a transformer-based architecture optimized for natural language understanding tasks.

Evaluation

Testing Data

The model was evaluated on a separate validation set consisting of annotated 10K reports not seen during training to ensure it can generalize across different texts and query types.

Metrics

Evaluation metrics included accuracy, F1 score, and a custom metric for response relevance to financial queries.

Technical Specifications

Model Architecture

The model employs a transformer-based architecture, leveraging attention mechanisms to focus on relevant parts of the text when generating responses.

Compute Infrastructure

Training was conducted on cloud-based GPUs with support for high-throughput training sessions.

Downloads last month
30
Safetensors
Model size
6.74B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.