AdaptLLM commited on
Commit
4020fcc
·
verified ·
1 Parent(s): 3bf3a0a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -33
README.md CHANGED
@@ -41,29 +41,6 @@ We investigate domain adaptation of MLLMs through post-training, focusing on dat
41
 
42
  **Code**: [https://github.com/bigai-ai/QA-Synthesizer](https://github.com/bigai-ai/QA-Synthesizer)
43
 
44
- ## About
45
-
46
- AdaMLLM is our latest effort to enhance task generalization of (M)LLMs by scaling synthetic supervised tasks based on unsupervised contexts.
47
-
48
- <p align='left'>
49
- <img src="https://cdn-uploads.huggingface.co/production/uploads/650801ced5578ef7e20b33d4/HUN3Cr66w_xpj5_c7QQaI.png" width="1000">
50
- </p>
51
-
52
- - [AdaptLLM](https://huggingface.co/papers/2309.09530)
53
- We employ rule-based methods to extract tasks from domain-specific corpora, reformatting them into reading comprehension tasks for continued pre-training. Our 7B finance model outperforms domain-specific models of much larger scales, such as BloombergGPT-50B.
54
-
55
- - [Instruction Pre-Training](https://huggingface.co/papers/2406.14491)
56
- We develop a general-purpose instruction synthesizer which significantly increases task diversity for LM pre-training, outperforming vanilla pre-training in both general pre-training from scratch and domain-adaptive continual pre-training.
57
-
58
- - AdaMLLM
59
- We extend supervised task synthesis to multimodality, introducing a unified visual instruction synthesizer to extract instruction-response pairs from image-caption data. Our synthetic tasks outperform those generated by manual rules, GPT-4, and GPT-4V in improving domain-specific performance for MLLMs.
60
-
61
- Looking ahead, we envision further broadening the scope of supervised task synthesis, efficiently enhancing the general capabilities of trained models.
62
-
63
- ## Contact
64
- Daixuan Cheng: `[email protected]`
65
-
66
-
67
  ## Citation
68
  If you find our work helpful, please cite us.
69
 
@@ -77,16 +54,6 @@ If you find our work helpful, please cite us.
77
  }
78
  ```
79
 
80
- [Instruction Pre-Training](https://huggingface.co/papers/2406.14491) (EMNLP 2024)
81
- ```bibtex
82
- @article{instructPT,
83
- title={Instruction Pre-Training: Language Models are Supervised Multitask Learners},
84
- author={Cheng, Daixuan and Gu, Yuxian and Huang, Shaohan and Bi, Junyu and Huang, Minlie and Wei, Furu},
85
- journal={arXiv preprint arXiv:2406.14491},
86
- year={2024}
87
- }
88
- ```
89
-
90
  [Adapt LLM to Domains](https://huggingface.co/papers/2309.09530) (ICLR 2024)
91
  ```bibtex
92
  @inproceedings{
 
41
 
42
  **Code**: [https://github.com/bigai-ai/QA-Synthesizer](https://github.com/bigai-ai/QA-Synthesizer)
43
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
44
  ## Citation
45
  If you find our work helpful, please cite us.
46
 
 
54
  }
55
  ```
56
 
 
 
 
 
 
 
 
 
 
 
57
  [Adapt LLM to Domains](https://huggingface.co/papers/2309.09530) (ICLR 2024)
58
  ```bibtex
59
  @inproceedings{