root
		
	commited on
		
		
					Commit 
							
							·
						
						102e286
	
1
								Parent(s):
							
							3308376
								
update config.json
Browse files- config.json +1 -2
 
    	
        config.json
    CHANGED
    
    | 
         @@ -7,9 +7,8 @@ 
     | 
|
| 7 | 
         
             
                "AutoModelForCausalLM": "modeling_polylm.PolyLMHeadModel"
         
     | 
| 8 | 
         
             
              },
         
     | 
| 9 | 
         
             
              "attn_pdrop": 0.0,
         
     | 
| 10 | 
         
            -
              "bos_token_id": 255999,
         
     | 
| 11 | 
         
             
              "embd_pdrop": 0.0,
         
     | 
| 12 | 
         
            -
              "eos_token_id":  
     | 
| 13 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 14 | 
         
             
              "layer_norm_epsilon": 1e-05,
         
     | 
| 15 | 
         
             
              "model_type": "gpt2",
         
     | 
| 
         | 
|
| 7 | 
         
             
                "AutoModelForCausalLM": "modeling_polylm.PolyLMHeadModel"
         
     | 
| 8 | 
         
             
              },
         
     | 
| 9 | 
         
             
              "attn_pdrop": 0.0,
         
     | 
| 
         | 
|
| 10 | 
         
             
              "embd_pdrop": 0.0,
         
     | 
| 11 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 12 | 
         
             
              "initializer_range": 0.02,
         
     | 
| 13 | 
         
             
              "layer_norm_epsilon": 1e-05,
         
     | 
| 14 | 
         
             
              "model_type": "gpt2",
         
     |