End of training
Browse files
    	
        README.md
    CHANGED
    
    | @@ -7,41 +7,7 @@ metrics: | |
| 7 | 
             
            - rouge
         | 
| 8 | 
             
            model-index:
         | 
| 9 | 
             
            - name: PTS-Bart-Large-CNN
         | 
| 10 | 
            -
              results:
         | 
| 11 | 
            -
              - task:
         | 
| 12 | 
            -
                  type: summarization
         | 
| 13 | 
            -
                  name: Summarization
         | 
| 14 | 
            -
                dataset:
         | 
| 15 | 
            -
                  name: PTS Dataset
         | 
| 16 | 
            -
                  type: PTS-Dataset
         | 
| 17 | 
            -
                metrics:
         | 
| 18 | 
            -
                - name: Rouge1
         | 
| 19 | 
            -
                  type: rouge
         | 
| 20 | 
            -
                  value: 0.6591
         | 
| 21 | 
            -
                - name: Rouge2
         | 
| 22 | 
            -
                  type: rouge
         | 
| 23 | 
            -
                  value: 0.449
         | 
| 24 | 
            -
                - name: Rougel
         | 
| 25 | 
            -
                  type: rouge
         | 
| 26 | 
            -
                  value: 0.5635
         | 
| 27 | 
            -
                - name: Rougelsum
         | 
| 28 | 
            -
                  type: rouge
         | 
| 29 | 
            -
                  value: 0.5633
         | 
| 30 | 
            -
            datasets:
         | 
| 31 | 
            -
            - ahmedmbutt/PTS-Dataset
         | 
| 32 | 
            -
            language:
         | 
| 33 | 
            -
            - en
         | 
| 34 | 
            -
            library_name: transformers
         | 
| 35 | 
            -
            widget:
         | 
| 36 | 
            -
            - text: >-
         | 
| 37 | 
            -
                I have to say that I do miss talking to a good psychiatrist- however. I
         | 
| 38 | 
            -
                could sit and argue for ages with a psychiatrist who is intelligent and kind
         | 
| 39 | 
            -
                (quite hard to find- but they do exist). Especially now that I have a PhD in
         | 
| 40 | 
            -
                philosophy and have read everything that can be found on madness- including
         | 
| 41 | 
            -
                the notes they wrote about me when I was in the hospital. Nowadays-
         | 
| 42 | 
            -
                psychiatrists have a tendency to sign me off pretty quickly when I come onto
         | 
| 43 | 
            -
                their radar. They don’t wish to deal with me- I tire them out.
         | 
| 44 | 
            -
            pipeline_tag: summarization
         | 
| 45 | 
             
            ---
         | 
| 46 |  | 
| 47 | 
             
            <!-- This model card has been generated automatically according to the information the Trainer had access to. You
         | 
| @@ -49,14 +15,14 @@ should probably proofread and complete it, then remove this comment. --> | |
| 49 |  | 
| 50 | 
             
            # PTS-Bart-Large-CNN
         | 
| 51 |  | 
| 52 | 
            -
            This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the  | 
| 53 | 
             
            It achieves the following results on the evaluation set:
         | 
| 54 | 
            -
            - Loss: 1. | 
| 55 | 
            -
            - Rouge1: 0. | 
| 56 | 
            -
            - Rouge2: 0. | 
| 57 | 
            -
            - Rougel: 0. | 
| 58 | 
            -
            - Rougelsum: 0. | 
| 59 | 
            -
            - Gen Len:  | 
| 60 |  | 
| 61 | 
             
            ## Model description
         | 
| 62 |  | 
| @@ -88,14 +54,14 @@ The following hyperparameters were used during training: | |
| 88 |  | 
| 89 | 
             
            | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
         | 
| 90 | 
             
            |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
         | 
| 91 | 
            -
            | No log        | 1.0   | 220  | 0. | 
| 92 | 
            -
            | No log        | 2.0   | 440  | 0. | 
| 93 | 
            -
            | 0. | 
| 94 | 
            -
            | 0. | 
| 95 | 
            -
            | 0. | 
| 96 | 
            -
            | 0. | 
| 97 | 
            -
            | 0. | 
| 98 | 
            -
            | 0. | 
| 99 |  | 
| 100 |  | 
| 101 | 
             
            ### Framework versions
         | 
| @@ -103,4 +69,4 @@ The following hyperparameters were used during training: | |
| 103 | 
             
            - Transformers 4.41.2
         | 
| 104 | 
             
            - Pytorch 2.3.0+cu121
         | 
| 105 | 
             
            - Datasets 2.20.0
         | 
| 106 | 
            -
            - Tokenizers 0.19.1
         | 
|  | |
| 7 | 
             
            - rouge
         | 
| 8 | 
             
            model-index:
         | 
| 9 | 
             
            - name: PTS-Bart-Large-CNN
         | 
| 10 | 
            +
              results: []
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 11 | 
             
            ---
         | 
| 12 |  | 
| 13 | 
             
            <!-- This model card has been generated automatically according to the information the Trainer had access to. You
         | 
|  | |
| 15 |  | 
| 16 | 
             
            # PTS-Bart-Large-CNN
         | 
| 17 |  | 
| 18 | 
            +
            This model is a fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the None dataset.
         | 
| 19 | 
             
            It achieves the following results on the evaluation set:
         | 
| 20 | 
            +
            - Loss: 1.1760
         | 
| 21 | 
            +
            - Rouge1: 0.6551
         | 
| 22 | 
            +
            - Rouge2: 0.4332
         | 
| 23 | 
            +
            - Rougel: 0.5543
         | 
| 24 | 
            +
            - Rougelsum: 0.5541
         | 
| 25 | 
            +
            - Gen Len: 80.0886
         | 
| 26 |  | 
| 27 | 
             
            ## Model description
         | 
| 28 |  | 
|  | |
| 54 |  | 
| 55 | 
             
            | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
         | 
| 56 | 
             
            |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
         | 
| 57 | 
            +
            | No log        | 1.0   | 220  | 0.8239          | 0.6263 | 0.3973 | 0.5238 | 0.5237    | 84.2023 |
         | 
| 58 | 
            +
            | No log        | 2.0   | 440  | 0.8201          | 0.6461 | 0.4184 | 0.5417 | 0.5416    | 81.1659 |
         | 
| 59 | 
            +
            | 0.7121        | 3.0   | 660  | 0.8661          | 0.6479 | 0.4226 | 0.5448 | 0.5454    | 80.5409 |
         | 
| 60 | 
            +
            | 0.7121        | 4.0   | 880  | 0.9784          | 0.6474 | 0.4242 | 0.5424 | 0.5425    | 82.2932 |
         | 
| 61 | 
            +
            | 0.2619        | 5.0   | 1100 | 1.0645          | 0.655  | 0.4327 | 0.5517 | 0.5517    | 80.8386 |
         | 
| 62 | 
            +
            | 0.2619        | 6.0   | 1320 | 1.1098          | 0.6548 | 0.4339 | 0.5542 | 0.5543    | 81.3545 |
         | 
| 63 | 
            +
            | 0.1124        | 7.0   | 1540 | 1.1528          | 0.6528 | 0.4298 | 0.5511 | 0.551     | 80.5705 |
         | 
| 64 | 
            +
            | 0.1124        | 8.0   | 1760 | 1.1760          | 0.6551 | 0.4332 | 0.5543 | 0.5541    | 80.0886 |
         | 
| 65 |  | 
| 66 |  | 
| 67 | 
             
            ### Framework versions
         | 
|  | |
| 69 | 
             
            - Transformers 4.41.2
         | 
| 70 | 
             
            - Pytorch 2.3.0+cu121
         | 
| 71 | 
             
            - Datasets 2.20.0
         | 
| 72 | 
            +
            - Tokenizers 0.19.1
         | 
    	
        model.safetensors
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
             
            size 1625422896
         | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:decb05eb85026d23066f4d7c898a7942df3fc8ebe27585ae528b438f830947ee
         | 
| 3 | 
             
            size 1625422896
         | 
    	
        runs/Jun24_16-23-47_a407000c0675/events.out.tfevents.1719246228.a407000c0675.257.0
    CHANGED
    
    | @@ -1,3 +1,3 @@ | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            -
            oid sha256: | 
| 3 | 
            -
            size  | 
|  | |
| 1 | 
             
            version https://git-lfs.github.com/spec/v1
         | 
| 2 | 
            +
            oid sha256:8d2b5624a59c6350b968ea01ae584711ab985e9bd917b67f626fce0cd3da4af7
         | 
| 3 | 
            +
            size 11131
         | 
