End of training
Browse files- README.md +7 -4
 - config.json +1 -43
 - logs/events.out.tfevents.1705982868.70e47a1f5afe.42.9 +3 -0
 - logs/events.out.tfevents.1705982894.70e47a1f5afe.42.10 +3 -0
 - model.safetensors +2 -2
 - trainer_state.json +18 -13
 - training_args.bin +1 -1
 
    	
        README.md
    CHANGED
    
    | 
         @@ -14,6 +14,13 @@ should probably proofread and complete it, then remove this comment. --> 
     | 
|
| 14 | 
         
             
            # ner-bert-ingredients
         
     | 
| 15 | 
         | 
| 16 | 
         
             
            This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 17 | 
         | 
| 18 | 
         
             
            ## Model description
         
     | 
| 19 | 
         | 
| 
         @@ -43,10 +50,6 @@ The following hyperparameters were used during training: 
     | 
|
| 43 | 
         
             
            - num_epochs: 50
         
     | 
| 44 | 
         
             
            - mixed_precision_training: Native AMP
         
     | 
| 45 | 
         | 
| 46 | 
         
            -
            ### Training results
         
     | 
| 47 | 
         
            -
             
     | 
| 48 | 
         
            -
             
     | 
| 49 | 
         
            -
             
     | 
| 50 | 
         
             
            ### Framework versions
         
     | 
| 51 | 
         | 
| 52 | 
         
             
            - Transformers 4.36.0
         
     | 
| 
         | 
|
| 14 | 
         
             
            # ner-bert-ingredients
         
     | 
| 15 | 
         | 
| 16 | 
         
             
            This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
         
     | 
| 17 | 
         
            +
            It achieves the following results on the evaluation set:
         
     | 
| 18 | 
         
            +
            - eval_loss: 4.4464
         
     | 
| 19 | 
         
            +
            - eval_runtime: 0.0176
         
     | 
| 20 | 
         
            +
            - eval_samples_per_second: 56.724
         
     | 
| 21 | 
         
            +
            - eval_steps_per_second: 56.724
         
     | 
| 22 | 
         
            +
            - epoch: 10.0
         
     | 
| 23 | 
         
            +
            - step: 10
         
     | 
| 24 | 
         | 
| 25 | 
         
             
            ## Model description
         
     | 
| 26 | 
         | 
| 
         | 
|
| 50 | 
         
             
            - num_epochs: 50
         
     | 
| 51 | 
         
             
            - mixed_precision_training: Native AMP
         
     | 
| 52 | 
         | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 53 | 
         
             
            ### Framework versions
         
     | 
| 54 | 
         | 
| 55 | 
         
             
            - Transformers 4.36.0
         
     | 
    	
        config.json
    CHANGED
    
    | 
         @@ -63,28 +63,7 @@ 
     | 
|
| 63 | 
         
             
                "50": "LABEL_50",
         
     | 
| 64 | 
         
             
                "51": "LABEL_51",
         
     | 
| 65 | 
         
             
                "52": "LABEL_52",
         
     | 
| 66 | 
         
            -
                "53": "LABEL_53" 
     | 
| 67 | 
         
            -
                "54": "LABEL_54",
         
     | 
| 68 | 
         
            -
                "55": "LABEL_55",
         
     | 
| 69 | 
         
            -
                "56": "LABEL_56",
         
     | 
| 70 | 
         
            -
                "57": "LABEL_57",
         
     | 
| 71 | 
         
            -
                "58": "LABEL_58",
         
     | 
| 72 | 
         
            -
                "59": "LABEL_59",
         
     | 
| 73 | 
         
            -
                "60": "LABEL_60",
         
     | 
| 74 | 
         
            -
                "61": "LABEL_61",
         
     | 
| 75 | 
         
            -
                "62": "LABEL_62",
         
     | 
| 76 | 
         
            -
                "63": "LABEL_63",
         
     | 
| 77 | 
         
            -
                "64": "LABEL_64",
         
     | 
| 78 | 
         
            -
                "65": "LABEL_65",
         
     | 
| 79 | 
         
            -
                "66": "LABEL_66",
         
     | 
| 80 | 
         
            -
                "67": "LABEL_67",
         
     | 
| 81 | 
         
            -
                "68": "LABEL_68",
         
     | 
| 82 | 
         
            -
                "69": "LABEL_69",
         
     | 
| 83 | 
         
            -
                "70": "LABEL_70",
         
     | 
| 84 | 
         
            -
                "71": "LABEL_71",
         
     | 
| 85 | 
         
            -
                "72": "LABEL_72",
         
     | 
| 86 | 
         
            -
                "73": "LABEL_73",
         
     | 
| 87 | 
         
            -
                "74": "LABEL_74"
         
     | 
| 88 | 
         
             
              },
         
     | 
| 89 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 90 | 
         
             
              "intermediate_size": 3072,
         
     | 
| 
         @@ -139,29 +118,8 @@ 
     | 
|
| 139 | 
         
             
                "LABEL_51": 51,
         
     | 
| 140 | 
         
             
                "LABEL_52": 52,
         
     | 
| 141 | 
         
             
                "LABEL_53": 53,
         
     | 
| 142 | 
         
            -
                "LABEL_54": 54,
         
     | 
| 143 | 
         
            -
                "LABEL_55": 55,
         
     | 
| 144 | 
         
            -
                "LABEL_56": 56,
         
     | 
| 145 | 
         
            -
                "LABEL_57": 57,
         
     | 
| 146 | 
         
            -
                "LABEL_58": 58,
         
     | 
| 147 | 
         
            -
                "LABEL_59": 59,
         
     | 
| 148 | 
         
             
                "LABEL_6": 6,
         
     | 
| 149 | 
         
            -
                "LABEL_60": 60,
         
     | 
| 150 | 
         
            -
                "LABEL_61": 61,
         
     | 
| 151 | 
         
            -
                "LABEL_62": 62,
         
     | 
| 152 | 
         
            -
                "LABEL_63": 63,
         
     | 
| 153 | 
         
            -
                "LABEL_64": 64,
         
     | 
| 154 | 
         
            -
                "LABEL_65": 65,
         
     | 
| 155 | 
         
            -
                "LABEL_66": 66,
         
     | 
| 156 | 
         
            -
                "LABEL_67": 67,
         
     | 
| 157 | 
         
            -
                "LABEL_68": 68,
         
     | 
| 158 | 
         
            -
                "LABEL_69": 69,
         
     | 
| 159 | 
         
             
                "LABEL_7": 7,
         
     | 
| 160 | 
         
            -
                "LABEL_70": 70,
         
     | 
| 161 | 
         
            -
                "LABEL_71": 71,
         
     | 
| 162 | 
         
            -
                "LABEL_72": 72,
         
     | 
| 163 | 
         
            -
                "LABEL_73": 73,
         
     | 
| 164 | 
         
            -
                "LABEL_74": 74,
         
     | 
| 165 | 
         
             
                "LABEL_8": 8,
         
     | 
| 166 | 
         
             
                "LABEL_9": 9
         
     | 
| 167 | 
         
             
              },
         
     | 
| 
         | 
|
| 63 | 
         
             
                "50": "LABEL_50",
         
     | 
| 64 | 
         
             
                "51": "LABEL_51",
         
     | 
| 65 | 
         
             
                "52": "LABEL_52",
         
     | 
| 66 | 
         
            +
                "53": "LABEL_53"
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 67 | 
         
             
              },
         
     | 
| 68 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 69 | 
         
             
              "intermediate_size": 3072,
         
     | 
| 
         | 
|
| 118 | 
         
             
                "LABEL_51": 51,
         
     | 
| 119 | 
         
             
                "LABEL_52": 52,
         
     | 
| 120 | 
         
             
                "LABEL_53": 53,
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 121 | 
         
             
                "LABEL_6": 6,
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 122 | 
         
             
                "LABEL_7": 7,
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 123 | 
         
             
                "LABEL_8": 8,
         
     | 
| 124 | 
         
             
                "LABEL_9": 9
         
     | 
| 125 | 
         
             
              },
         
     | 
    	
        logs/events.out.tfevents.1705982868.70e47a1f5afe.42.9
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:21b3f3ff7b8fb5dbecad8aef3836d435dac2c5d5fb697882eadaaf21055007d0
         
     | 
| 3 | 
         
            +
            size 7028
         
     | 
    	
        logs/events.out.tfevents.1705982894.70e47a1f5afe.42.10
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:2330695fbad9f24ae1b10f12ef6d1ba3a29aa60bcef4ee3f64dae7b35fa5d04e
         
     | 
| 3 | 
         
            +
            size 7028
         
     | 
    	
        model.safetensors
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
            -
            size  
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b1f2ed33d49476ecdf341fac292ec64b65edf6c4f0c485b9fcbbab100d62d596
         
     | 
| 3 | 
         
            +
            size 435756040
         
     | 
    	
        trainer_state.json
    CHANGED
    
    | 
         @@ -1,29 +1,34 @@ 
     | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "best_metric": null,
         
     | 
| 3 | 
         
             
              "best_model_checkpoint": null,
         
     | 
| 4 | 
         
            -
              "epoch":  
     | 
| 5 | 
         
            -
              "eval_steps":  
     | 
| 6 | 
         
            -
              "global_step":  
     | 
| 7 | 
         
             
              "is_hyper_param_search": false,
         
     | 
| 8 | 
         
             
              "is_local_process_zero": true,
         
     | 
| 9 | 
         
             
              "is_world_process_zero": true,
         
     | 
| 10 | 
         
             
              "log_history": [
         
     | 
| 11 | 
         
             
                {
         
     | 
| 12 | 
         
            -
                  "epoch":  
     | 
| 13 | 
         
            -
                  " 
     | 
| 14 | 
         
            -
                  " 
     | 
| 15 | 
         
            -
                  " 
     | 
| 16 | 
         
            -
             
     | 
| 17 | 
         
            -
             
     | 
| 18 | 
         
            -
                  " 
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 19 | 
         
             
                }
         
     | 
| 20 | 
         
             
              ],
         
     | 
| 21 | 
         
            -
              "logging_steps":  
     | 
| 22 | 
         
             
              "max_steps": 50,
         
     | 
| 23 | 
         
             
              "num_input_tokens_seen": 0,
         
     | 
| 24 | 
         
             
              "num_train_epochs": 50,
         
     | 
| 25 | 
         
            -
              "save_steps":  
     | 
| 26 | 
         
            -
              "total_flos":  
     | 
| 27 | 
         
             
              "train_batch_size": 16,
         
     | 
| 28 | 
         
             
              "trial_name": null,
         
     | 
| 29 | 
         
             
              "trial_params": null
         
     | 
| 
         | 
|
| 1 | 
         
             
            {
         
     | 
| 2 | 
         
             
              "best_metric": null,
         
     | 
| 3 | 
         
             
              "best_model_checkpoint": null,
         
     | 
| 4 | 
         
            +
              "epoch": 10.0,
         
     | 
| 5 | 
         
            +
              "eval_steps": 10,
         
     | 
| 6 | 
         
            +
              "global_step": 10,
         
     | 
| 7 | 
         
             
              "is_hyper_param_search": false,
         
     | 
| 8 | 
         
             
              "is_local_process_zero": true,
         
     | 
| 9 | 
         
             
              "is_world_process_zero": true,
         
     | 
| 10 | 
         
             
              "log_history": [
         
     | 
| 11 | 
         
             
                {
         
     | 
| 12 | 
         
            +
                  "epoch": 10.0,
         
     | 
| 13 | 
         
            +
                  "learning_rate": 4e-05,
         
     | 
| 14 | 
         
            +
                  "loss": 0.3826,
         
     | 
| 15 | 
         
            +
                  "step": 10
         
     | 
| 16 | 
         
            +
                },
         
     | 
| 17 | 
         
            +
                {
         
     | 
| 18 | 
         
            +
                  "epoch": 10.0,
         
     | 
| 19 | 
         
            +
                  "eval_loss": 4.4464287757873535,
         
     | 
| 20 | 
         
            +
                  "eval_runtime": 0.0176,
         
     | 
| 21 | 
         
            +
                  "eval_samples_per_second": 56.724,
         
     | 
| 22 | 
         
            +
                  "eval_steps_per_second": 56.724,
         
     | 
| 23 | 
         
            +
                  "step": 10
         
     | 
| 24 | 
         
             
                }
         
     | 
| 25 | 
         
             
              ],
         
     | 
| 26 | 
         
            +
              "logging_steps": 10,
         
     | 
| 27 | 
         
             
              "max_steps": 50,
         
     | 
| 28 | 
         
             
              "num_input_tokens_seen": 0,
         
     | 
| 29 | 
         
             
              "num_train_epochs": 50,
         
     | 
| 30 | 
         
            +
              "save_steps": 10,
         
     | 
| 31 | 
         
            +
              "total_flos": 6571230957720.0,
         
     | 
| 32 | 
         
             
              "train_batch_size": 16,
         
     | 
| 33 | 
         
             
              "trial_name": null,
         
     | 
| 34 | 
         
             
              "trial_params": null
         
     | 
    	
        training_args.bin
    CHANGED
    
    | 
         @@ -1,3 +1,3 @@ 
     | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            -
            oid sha256: 
     | 
| 3 | 
         
             
            size 4283
         
     | 
| 
         | 
|
| 1 | 
         
             
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:665509b7779b6a7e7ab19bb9a60700f3660066d0bf783678aea312e07320debf
         
     | 
| 3 | 
         
             
            size 4283
         
     |