Update README.md
Browse files
README.md
CHANGED
@@ -102,7 +102,7 @@ print(respuesta)
|
|
102 |
|
103 |
# Referencias
|
104 |
1- **MS MARCO Dataset:**
|
105 |
-
|
106 |
@misc{bajaj2018msmarcohumangenerated,
|
107 |
title={MS MARCO: A Human Generated MAchine Reading COmprehension Dataset},
|
108 |
author={Payal Bajaj and Daniel Campos and Nick Craswell and Li Deng and Jianfeng Gao and Xiaodong Liu and Rangan Majumder and Andrew McNamara and Bhaskar Mitra and Tri Nguyen and Mir Rosenberg and Xia Song and Alina Stoica and Saurabh Tiwary and Tong Wang},
|
@@ -112,9 +112,9 @@ print(respuesta)
|
|
112 |
primaryClass={cs.CL},
|
113 |
url={https://arxiv.org/abs/1611.09268},
|
114 |
}
|
115 |
-
|
116 |
2- **QLoRA Paper:**
|
117 |
-
|
118 |
@misc{dettmers2023qloraefficientfinetuningquantized,
|
119 |
title={QLoRA: Efficient Finetuning of Quantized LLMs},
|
120 |
author={Tim Dettmers and Artidoro Pagnoni and Ari Holtzman and Luke Zettlemoyer},
|
@@ -124,11 +124,11 @@ print(respuesta)
|
|
124 |
primaryClass={cs.LG},
|
125 |
url={https://arxiv.org/abs/2305.14314},
|
126 |
}
|
|
|
127 |
|
128 |
3- [**PEFT Library**](https://huggingface.co/docs/peft/index)
|
129 |
|
130 |
4- **LoRA Paper:**
|
131 |
-
|
132 |
```bibtex
|
133 |
@misc{hu2021loralowrankadaptationlarge,
|
134 |
title={LoRA: Low-Rank Adaptation of Large Language Models},
|
|
|
102 |
|
103 |
# Referencias
|
104 |
1- **MS MARCO Dataset:**
|
105 |
+
```bibtex
|
106 |
@misc{bajaj2018msmarcohumangenerated,
|
107 |
title={MS MARCO: A Human Generated MAchine Reading COmprehension Dataset},
|
108 |
author={Payal Bajaj and Daniel Campos and Nick Craswell and Li Deng and Jianfeng Gao and Xiaodong Liu and Rangan Majumder and Andrew McNamara and Bhaskar Mitra and Tri Nguyen and Mir Rosenberg and Xia Song and Alina Stoica and Saurabh Tiwary and Tong Wang},
|
|
|
112 |
primaryClass={cs.CL},
|
113 |
url={https://arxiv.org/abs/1611.09268},
|
114 |
}
|
115 |
+
```
|
116 |
2- **QLoRA Paper:**
|
117 |
+
```bibtex
|
118 |
@misc{dettmers2023qloraefficientfinetuningquantized,
|
119 |
title={QLoRA: Efficient Finetuning of Quantized LLMs},
|
120 |
author={Tim Dettmers and Artidoro Pagnoni and Ari Holtzman and Luke Zettlemoyer},
|
|
|
124 |
primaryClass={cs.LG},
|
125 |
url={https://arxiv.org/abs/2305.14314},
|
126 |
}
|
127 |
+
```
|
128 |
|
129 |
3- [**PEFT Library**](https://huggingface.co/docs/peft/index)
|
130 |
|
131 |
4- **LoRA Paper:**
|
|
|
132 |
```bibtex
|
133 |
@misc{hu2021loralowrankadaptationlarge,
|
134 |
title={LoRA: Low-Rank Adaptation of Large Language Models},
|