ihk commited on
Commit
6e4dbe9
·
1 Parent(s): 6327a59

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -10
README.md CHANGED
@@ -1,7 +1,5 @@
1
  ---
2
  base_model: jjzha/jobbert-base-cased
3
- tags:
4
- - generated_from_trainer
5
  metrics:
6
  - accuracy
7
  - precision
@@ -10,6 +8,9 @@ metrics:
10
  model-index:
11
  - name: results
12
  results: []
 
 
 
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -27,17 +28,11 @@ It achieves the following results on the evaluation set:
27
 
28
  ## Model description
29
 
30
- More information needed
31
-
32
- ## Intended uses & limitations
33
-
34
- More information needed
35
 
36
  ## Training and evaluation data
37
 
38
- More information needed
39
-
40
- ## Training procedure
41
 
42
  ### Training hyperparameters
43
 
 
1
  ---
2
  base_model: jjzha/jobbert-base-cased
 
 
3
  metrics:
4
  - accuracy
5
  - precision
 
8
  model-index:
9
  - name: results
10
  results: []
11
+ widget:
12
+ - text: You should be a skilled communicator.
13
+ - text: You can programme in Python and CSS.
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
28
 
29
  ## Model description
30
 
31
+ The base model (`jjzha/jobbert-base-cased`) is a BERT transformer model, pretrained on a corpus of ~3.2 million sentences from job adverts for the objective of Masked Language Modelling (MLM). A token classification head is added to the top of the model to predict a label for every token in a given spequence. In this instance, it is predicting a label for every token in a job description, where the label is either a 'B-SKILL', 'I-SKILL' or 'O' (not a skill).
 
 
 
 
32
 
33
  ## Training and evaluation data
34
 
35
+ The model was trained on 4112 job advert sentences.
 
 
36
 
37
  ### Training hyperparameters
38