lalaland1125 commited on
Commit
9f06fd3
·
1 Parent(s): 2d320ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -4,13 +4,13 @@ library_name: femr
4
  tags:
5
  - healthcare
6
  ---
7
- # CLMBR-T-Base-Synthetic
8
 
9
- This is a CLMBR model that was "pretrained" on **synthetic data**. The purpose of this model is to test code pipelines and demonstrate how to use CLMBR before applying for access to the official CLMBR release that was trained on real Stanford Hospital data.
10
 
11
  The model architecture is CLMBR-T-Base (144M params), as originally described in [the EHRSHOT paper (Wornow et al. 2023)](https://arxiv.org/abs/2307.02028), and based on the architecture originally developed in [the Clinical Language Modeling Based Representations paper (Steinberg et al. 2021)](https://www.sciencedirect.com/science/article/pii/S1532046420302653)
12
 
13
- The synthetic data used for training is uniformly sampled random data with no intrinsic signal, so **this model has no clinical or research use.**
14
 
15
  ## Model Details
16
 
@@ -35,7 +35,7 @@ This model generates dense representations for patients based on the structured
35
 
36
  These representations can then be used for downstream tasks such as predicting diagnoses, detecting anomalies, or doing propensity score matching for causal inference.
37
 
38
- However, please note that **this version of the model was trained on synthetic data.** Thus, the outputs will be random/meaningless.
39
 
40
 
41
  ## How to Get Started with the Model
@@ -93,11 +93,11 @@ with torch.no_grad():
93
 
94
  ## Training Details
95
 
96
- The synthetic data used for training is uniformly sampled random data with no intrinsic signal, so this model itself has no clinical or research use.
97
 
98
  ## Evaluation
99
 
100
- As this model was trained on random synthetic data, we expect it to perform no better than chance.
101
 
102
  ## Technical Specifications
103
 
@@ -105,7 +105,7 @@ Please see [the EHRSHOT paper (Wornow et al. 2023)](https://arxiv.org/abs/2307.0
105
 
106
  ### Compute Infrastructure
107
 
108
- This model was trained on a single V100 GPU.
109
 
110
  #### Software
111
 
 
4
  tags:
5
  - healthcare
6
  ---
7
+ # CLMBR-T-Base-Random
8
 
9
+ This is a CLMBR model that was randomly initialized with a dummy vocabulary. The purpose of this model is to test code pipelines and demonstrate how to use CLMBR before applying for access to the official CLMBR release that was trained on real Stanford Hospital data.
10
 
11
  The model architecture is CLMBR-T-Base (144M params), as originally described in [the EHRSHOT paper (Wornow et al. 2023)](https://arxiv.org/abs/2307.02028), and based on the architecture originally developed in [the Clinical Language Modeling Based Representations paper (Steinberg et al. 2021)](https://www.sciencedirect.com/science/article/pii/S1532046420302653)
12
 
13
+ The weights are random, so **this model has no clinical or research use.**
14
 
15
  ## Model Details
16
 
 
35
 
36
  These representations can then be used for downstream tasks such as predicting diagnoses, detecting anomalies, or doing propensity score matching for causal inference.
37
 
38
+ However, please note that **this version of the model has random weights.** Thus, the outputs will be random/meaningless.
39
 
40
 
41
  ## How to Get Started with the Model
 
93
 
94
  ## Training Details
95
 
96
+ This model is not trained.
97
 
98
  ## Evaluation
99
 
100
+ None, as the weights are random.
101
 
102
  ## Technical Specifications
103
 
 
105
 
106
  ### Compute Infrastructure
107
 
108
+ This model was not trained.
109
 
110
  #### Software
111