Binarybardakshat commited on
Commit
0a08052
1 Parent(s): 50bc80b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +81 -38
README.md CHANGED
@@ -1,56 +1,99 @@
1
- # SVLM (Scientific Virtual Language Model)
2
 
3
- This model is designed to answer questions based on research papers from the ACL Anthology. The model uses a Seq2Seq approach and was fine-tuned on the `SVLM-ACL-DATASET` dataset, focusing on generating high-quality responses related to research topics.
4
 
5
  ## Model Details
6
 
7
- - **Model Architecture:** BART (Base)
8
- - **Libraries Used:**
9
- - `transformers`
10
- - `tensorflow`
11
- - `datasets`
12
- - `safetensors`
13
- - `pandas`
14
 
15
- - **Training Framework:** TensorFlow
16
- - **Data Source:** [SVLM-ACL-DATASET](https://huggingface.co/datasets/Binarybardakshat/SVLM-ACL-DATASET)
 
 
 
17
 
18
- ## Usage
19
 
20
- ```python
21
- from transformers import pipeline
22
 
23
- # Load the model and tokenizer
24
- qa_pipeline = pipeline("question-answering", model="Binarybardakshat/svlm", tokenizer="Binarybardakshat/svlm")
25
 
26
- # Ask a question
27
- result = qa_pipeline(question="What is the purpose of this research?", context="Context of the research paper...")
28
- print(result)
29
- # SVLM (Scientific Virtual Language Model)
30
 
31
- This model is designed to answer questions based on research papers from the ACL Anthology. The model uses a Seq2Seq approach and was fine-tuned on the `SVLM-ACL-DATASET` dataset, focusing on generating high-quality responses related to research topics.
32
 
33
- ## Model Details
 
 
 
 
34
 
35
- - **Model Architecture:** BART (Base)
36
- - **Libraries Used:**
37
- - `transformers`
38
- - `tensorflow`
39
- - `datasets`
40
- - `safetensors`
41
- - `pandas`
42
 
43
- - **Training Framework:** TensorFlow
44
- - **Data Source:** [SVLM-ACL-DATASET](https://huggingface.co/datasets/Binarybardakshat/SVLM-ACL-DATASET)
45
 
46
- ## Usage
 
 
 
 
47
 
48
  ```python
49
- from transformers import pipeline
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
 
51
- # Load the model and tokenizer
52
- qa_pipeline = pipeline("question-answering", model="Binarybardakshat/svlm", tokenizer="Binarybardakshat/svlm")
53
 
54
- # Ask a question
55
- result = qa_pipeline(question="What is the purpose of this research?", context="Context of the research paper...")
56
- print(result)
 
1
+ # Model Card for SVLM
2
 
3
+ This model is a Seq2Seq Language Model (SVLM) fine-tuned to answer questions from the ACL research paper dataset. It generates responses related to academic research questions, making it useful for research and academic inquiry.
4
 
5
  ## Model Details
6
 
7
+ ### Model Description
 
 
 
 
 
 
8
 
9
+ - **Developed by:** @binarybardakshat
10
+ - **Model type:** Seq2Seq Language Model (BART-based)
11
+ - **Language(s) (NLP):** English
12
+ - **License:** [More Information Needed]
13
+ - **Finetuned from model:** facebook/bart-base
14
 
15
+ ### Model Sources
16
 
17
+ - **Repository:** [More Information Needed]
 
18
 
19
+ ## Uses
 
20
 
21
+ ### Direct Use
 
 
 
22
 
23
+ This model can be directly used to answer questions based on research data from ACL papers. It is suitable for academic and research purposes.
24
 
25
+ ### Out-of-Scope Use
26
+
27
+ The model may not work well for general conversation or non-research-related queries.
28
+
29
+ ## Bias, Risks, and Limitations
30
 
31
+ The model may carry biases present in the training data, which consists of ACL research papers. It might not generalize well outside this domain.
 
 
 
 
 
 
32
 
33
+ ### Recommendations
 
34
 
35
+ Users should be cautious of biases and ensure that outputs align with their academic requirements.
36
+
37
+ ## How to Get Started with the Model
38
+
39
+ Use the code below to get started with the model:
40
 
41
  ```python
42
+ from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
43
+
44
+ tokenizer = AutoTokenizer.from_pretrained("path_to_your_tokenizer")
45
+ model = AutoModelForSeq2SeqLM.from_pretrained("path_to_your_model")
46
+ ## Training Details
47
+
48
+ ### Training Data
49
+
50
+ The model was trained using the ACL dataset, which consists of research papers focused on computational linguistics.
51
+
52
+ ### Training Procedure
53
+
54
+ #### Training Hyperparameters
55
+
56
+ - **Training regime:** fp32
57
+ - **Learning rate:** 2e-5
58
+ - **Epochs:** 3
59
+ - **Batch size:** 8
60
+
61
+ ## Evaluation
62
+
63
+ ### Testing Data
64
+
65
+ The model was evaluated on a subset of the ACL dataset, focusing on research-related questions.
66
+
67
+ ### Metrics
68
+
69
+ - **Accuracy**
70
+ - **Loss**
71
+
72
+ ### Results
73
+
74
+ The model performs best in research-related question-answering tasks. Further evaluation metrics will be added as the model is used more widely.
75
+
76
+ ## Environmental Impact
77
+
78
+ - **Hardware Type:** GPU (NVIDIA V100)
79
+ - **Hours used:** [More Information Needed]
80
+ - **Compute Region:** [More Information Needed]
81
+ - **Carbon Emitted:** [More Information Needed]
82
+
83
+ ## Technical Specifications
84
+
85
+ ### Model Architecture and Objective
86
+
87
+ The model is based on BART architecture, designed to perform sequence-to-sequence tasks like text summarization and translation.
88
+
89
+ ### Compute Infrastructure
90
+
91
+ #### Hardware
92
+
93
+ - **NVIDIA V100 GPU**
94
 
95
+ #### Software
 
96
 
97
+ - **TensorFlow**
98
+ - **Transformers**
99
+ - **Safetensors**