Binarybardakshat commited on
Commit
ee6dfc1
1 Parent(s): 71a2ac0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -88
README.md CHANGED
@@ -1,100 +1,44 @@
1
- # Model Card for SVLM
2
-
3
- This model is a Seq2Seq Language Model (SVLM) fine-tuned to answer questions from the ACL research paper dataset. It generates responses related to academic research questions, making it useful for research and academic inquiry.
 
 
 
 
 
 
 
 
 
 
4
 
5
  ## Model Details
6
 
7
- ### Model Description
8
-
9
- - **Developed by:** @binarybardakshat
10
- - **Model type:** Seq2Seq Language Model (BART-based)
11
- - **Language(s) (NLP):** English
12
- - **License:** [More Information Needed]
13
- - **Finetuned from model:** facebook/bart-base
14
-
15
- ### Model Sources
16
-
17
- - **Repository:** [More Information Needed]
18
-
19
- ## Uses
20
-
21
- ### Direct Use
22
-
23
- This model can be directly used to answer questions based on research data from ACL papers. It is suitable for academic and research purposes.
24
-
25
- ### Out-of-Scope Use
26
 
27
- The model may not work well for general conversation or non-research-related queries.
28
 
29
- ## Bias, Risks, and Limitations
30
-
31
- The model may carry biases present in the training data, which consists of ACL research papers. It might not generalize well outside this domain.
32
-
33
- ### Recommendations
34
-
35
- Users should be cautious of biases and ensure that outputs align with their academic requirements.
36
-
37
- ## How to Get Started with the Model
38
-
39
- Use the code below to get started with the model:
40
 
41
  ```python
42
- from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
43
-
44
- tokenizer = AutoTokenizer.from_pretrained("path_to_your_tokenizer")
45
- model = AutoModelForSeq2SeqLM.from_pretrained("path_to_your_model")
46
- ````
47
- ## Training Details
48
-
49
- ### Training Data
50
-
51
- The model was trained using the ACL dataset, which consists of research papers focused on computational linguistics.
52
-
53
- ### Training Procedure
54
-
55
- #### Training Hyperparameters
56
-
57
- - **Training regime:** fp32
58
- - **Learning rate:** 2e-5
59
- - **Epochs:** 3
60
- - **Batch size:** 8
61
-
62
- ## Evaluation
63
-
64
- ### Testing Data
65
-
66
- The model was evaluated on a subset of the ACL dataset, focusing on research-related questions.
67
-
68
- ### Metrics
69
-
70
- - **Accuracy**
71
- - **Loss**
72
-
73
- ### Results
74
-
75
- The model performs best in research-related question-answering tasks. Further evaluation metrics will be added as the model is used more widely.
76
-
77
- ## Environmental Impact
78
-
79
- - **Hardware Type:** GPU (NVIDIA V100)
80
- - **Hours used:** [More Information Needed]
81
- - **Compute Region:** [More Information Needed]
82
- - **Carbon Emitted:** [More Information Needed]
83
-
84
- ## Technical Specifications
85
-
86
- ### Model Architecture and Objective
87
-
88
- The model is based on BART architecture, designed to perform sequence-to-sequence tasks like text summarization and translation.
89
 
90
- ### Compute Infrastructure
 
 
91
 
92
- #### Hardware
 
93
 
94
- - **NVIDIA V100 GPU**
 
95
 
96
- #### Software
 
 
97
 
98
- - **TensorFlow**
99
- - **Transformers**
100
- - **Safetensors**
 
1
+ ---
2
+ license: openrail
3
+ datasets:
4
+ - Binarybardakshat/SVLM-ACL-DATASET
5
+ language:
6
+ - en
7
+ library_name: transformers
8
+ tags:
9
+ - code
10
+ ---
11
+ # SVLM: A Question-Answering Model for ACL Research Papers
12
+
13
+ This model, `SVLM`, is designed to answer questions based on research papers from the ACL dataset. It leverages the BART architecture to generate precise answers from scientific abstracts.
14
 
15
  ## Model Details
16
 
17
+ - **Model Architecture:** BART (Bidirectional and Auto-Regressive Transformers)
18
+ - **Framework:** TensorFlow
19
+ - **Dataset:** [Binarybardakshat/SVLM-ACL-DATASET](https://huggingface.co/datasets/Binarybardakshat/SVLM-ACL-DATASET)
20
+ - **Author:** @binarybard (Akshat Shukla)
21
+ - **Purpose:** The model is trained to provide answers to questions from the ACL research paper dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
+ ## Usage
24
 
25
+ To use this model with the Hugging Face Interface API:
 
 
 
 
 
 
 
 
 
 
26
 
27
  ```python
28
+ from transformers import AutoTokenizer, TFAutoModelForSeq2SeqLM
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
+ # Load the model and tokenizer
31
+ tokenizer = AutoTokenizer.from_pretrained("Binarybardakshat/SVLM")
32
+ model = TFAutoModelForSeq2SeqLM.from_pretrained("Binarybardakshat/SVLM")
33
 
34
+ # Example input
35
+ input_text = "What is the main contribution of the paper titled 'Your Paper Title'?"
36
 
37
+ # Tokenize input
38
+ inputs = tokenizer(input_text, return_tensors="tf", padding=True, truncation=True)
39
 
40
+ # Generate answer
41
+ outputs = model.generate(inputs.input_ids, max_length=50, num_beams=5, early_stopping=True)
42
+ answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
43
 
44
+ print("Answer:", answer)