probablybots
commited on
Commit
•
0b20da3
1
Parent(s):
27f7f52
Update README.md
Browse files
README.md
CHANGED
@@ -2,14 +2,14 @@
|
|
2 |
tags:
|
3 |
- biology
|
4 |
---
|
5 |
-
# DNA
|
6 |
-
DNA
|
7 |
|
8 |
-
By scaling model depth while maintaining a short context length of 4000 nucleotides, DNA
|
9 |
-
<center><img src="DNA_RNA FM model architecture.png" alt="An Overview of DNA
|
10 |
|
11 |
## Model Architectural Details
|
12 |
-
DNA
|
13 |
|
14 |
To learn semantically meaningful representations, we employed an BERT-style encoder-only dense transformer architecture. We make minor updates to this architecture to align with current best practices, including using SwiGLU and LayerNorms. Additionally, we use Rotary Positional Embeddings (RoPE), given that DNA syntax does not function based on absolute nucleotide positions but nucleotides interact in highly local and context-specific ways. Below are more detailes about the model architecture:
|
15 |
| Model Arch Component | Value |
|
@@ -21,14 +21,14 @@ tags:
|
|
21 |
| Vocab Size | 16 |
|
22 |
| Context Length | 4000 |
|
23 |
|
24 |
-
## Pre-training of DNA
|
25 |
-
Here we briefly introduce the details of pre-training of DNA
|
26 |
|
27 |
### Data
|
28 |
-
To test whether representation capacity has limited the development of DNA language models in previous studies, we utilize the data set and splits from the Nucleotide Transformer. Starting from a total of `812` genomes with `712` for training, `50` for validation, and `50` for testing, we removed `17` entries which had been deleted from NCBI since the original dataset’s publication on Hugging Face. One of these was the important model organism Rattus norvegigus, which we replaced with the current reference genome. This resulted in `696` genomes for training, `50` for validation, and `50` for testing. We pre-trained DNA
|
29 |
|
30 |
### Training Details
|
31 |
-
The weights of our seven billion parameter model occupy over 200GB of memory in 32 bit precision. To train a model of this size, we use model parallelism to split training across 256 H100 GPUs using the Megatron-LM framework. We also employed bfloat16 mixed precision training and FlashAttention-2 to allow for training with large context length at scale. With this configuration, DNA
|
32 |
| Hyper-params | Value |
|
33 |
| ------------- |:-------------:|
|
34 |
| Global Batch Size | 1024 |
|
@@ -37,13 +37,13 @@ To test whether representation capacity has limited the development of DNA langu
|
|
37 |
| Total Iters | 100000 |
|
38 |
|
39 |
### Tokenization
|
40 |
-
To minimize bias and learn high-resolution single-nucleotide dependencies, we opted to align closely with the real data and use character-level tokenization with a 5-letter vocabulary: `A, T, C, G, N`, where `N` is commonly used in gene sequencing to denote uncertain elements. Sequences were also prefixed with a `[CLS]` token and suffixed with a `[EOS]` token as hooks for downstream tasks. We chose a context length of 4,000 nucleotides as the longest context which would fit within DNA
|
41 |
|
42 |
-
## Evaluation of DNA
|
43 |
-
We evaluate the benefits of pretraining DNA
|
44 |
|
45 |
## Results
|
46 |
-
<center><img src="circle_benchmarks.png" alt="Downstream results of DNA
|
47 |
|
48 |
## How to Use
|
49 |
### Build any downstream models from this backbone with ModelGenerator
|
@@ -94,7 +94,7 @@ print(logits)
|
|
94 |
|
95 |
|
96 |
## Citation
|
97 |
-
Please cite DNA
|
98 |
```
|
99 |
@inproceedings{ellington2024accurate,
|
100 |
title={Accurate and General {DNA} Representations Emerge from Genome Foundation Models at Scale},
|
|
|
2 |
tags:
|
3 |
- biology
|
4 |
---
|
5 |
+
# AIDO.DNA-7B
|
6 |
+
AIDO.DNA-7B is DNA foundation model trained on 10.6 billion nucleotides from 796 species, enabling genome mining, in silico mutagenesis studies, gene expression prediction, and directed sequence generation.
|
7 |
|
8 |
+
By scaling model depth while maintaining a short context length of 4000 nucleotides, AIDO.DNA shows substantial improvements across a breadth of tasks in functional genomics using transfer learning, sequence generation, and unsupervised annotation of functional elements. Notably, AIDO.DNA outperforms prior encoder-only architectures without new data, suggesting that new scaling laws are needed to achieve compute-optimal DNA language models.
|
9 |
+
<center><img src="DNA_RNA FM model architecture.png" alt="An Overview of AIDO.DNA-7B" style="width:70%; height:auto;" /></center>
|
10 |
|
11 |
## Model Architectural Details
|
12 |
+
AIDO.DNA-7B is based on the bidirectional transformer encoder (BERT) architecture with single-nucleotide tokenization, and is optimized using a masked language modeling (MLM) training objective.
|
13 |
|
14 |
To learn semantically meaningful representations, we employed an BERT-style encoder-only dense transformer architecture. We make minor updates to this architecture to align with current best practices, including using SwiGLU and LayerNorms. Additionally, we use Rotary Positional Embeddings (RoPE), given that DNA syntax does not function based on absolute nucleotide positions but nucleotides interact in highly local and context-specific ways. Below are more detailes about the model architecture:
|
15 |
| Model Arch Component | Value |
|
|
|
21 |
| Vocab Size | 16 |
|
22 |
| Context Length | 4000 |
|
23 |
|
24 |
+
## Pre-training of AIDO.DNA-7B
|
25 |
+
Here we briefly introduce the details of pre-training of AIDO.DNA-7B. For more detailed information, please refer to [our paper](https://openreview.net/forum?id=Kis8tVUeNi&referrer=%5BAuthor%20Console%5D(%2Fgroup%3Fid%3DNeurIPS.cc%2F2024%2FWorkshop%2FAIDrugX%2FAuthors%23your-submissions)).
|
26 |
|
27 |
### Data
|
28 |
+
To test whether representation capacity has limited the development of DNA language models in previous studies, we utilize the data set and splits from the Nucleotide Transformer. Starting from a total of `812` genomes with `712` for training, `50` for validation, and `50` for testing, we removed `17` entries which had been deleted from NCBI since the original dataset’s publication on Hugging Face. One of these was the important model organism Rattus norvegigus, which we replaced with the current reference genome. This resulted in `696` genomes for training, `50` for validation, and `50` for testing. We pre-trained AIDO.DNA-7B With a total of 10.6 billion training tokens.
|
29 |
|
30 |
### Training Details
|
31 |
+
The weights of our seven billion parameter model occupy over 200GB of memory in 32 bit precision. To train a model of this size, we use model parallelism to split training across 256 H100 GPUs using the Megatron-LM framework. We also employed bfloat16 mixed precision training and FlashAttention-2 to allow for training with large context length at scale. With this configuration, AIDO.DNA-7B took 8 days to train.
|
32 |
| Hyper-params | Value |
|
33 |
| ------------- |:-------------:|
|
34 |
| Global Batch Size | 1024 |
|
|
|
37 |
| Total Iters | 100000 |
|
38 |
|
39 |
### Tokenization
|
40 |
+
To minimize bias and learn high-resolution single-nucleotide dependencies, we opted to align closely with the real data and use character-level tokenization with a 5-letter vocabulary: `A, T, C, G, N`, where `N` is commonly used in gene sequencing to denote uncertain elements. Sequences were also prefixed with a `[CLS]` token and suffixed with a `[EOS]` token as hooks for downstream tasks. We chose a context length of 4,000 nucleotides as the longest context which would fit within AIDO.DNA-7B during pretraining, and chunked our dataset of 796 genomes into non-overlapping segments.
|
41 |
|
42 |
+
## Evaluation of AIDO.DNA-7B
|
43 |
+
We evaluate the benefits of pretraining AIDO.DNA-7B by conducting a comprehensive series of experiments related to functional genomics, genome mining, metabolic engineering, synthetic biology, and therapeutics design, covering supervised, unsupervised, and generative objectives. Unless otherwise stated, hyperparameters were determined by optimizing model performance on a 10% validation split of the training data, and models were tested using the checkpoint with the lowest validation loss. For more detailed information, please refer to [our paper](https://doi.org/10.1101/2024.12.01.625444)).
|
44 |
|
45 |
## Results
|
46 |
+
<center><img src="circle_benchmarks.png" alt="Downstream results of AIDO.DNA-7B" style="width:70%; height:auto;" /></center>
|
47 |
|
48 |
## How to Use
|
49 |
### Build any downstream models from this backbone with ModelGenerator
|
|
|
94 |
|
95 |
|
96 |
## Citation
|
97 |
+
Please cite AIDO.DNA using the following BibTeX code:
|
98 |
```
|
99 |
@inproceedings{ellington2024accurate,
|
100 |
title={Accurate and General {DNA} Representations Emerge from Genome Foundation Models at Scale},
|