Spaces:
				
			
			
	
			
			
		Running
		
			on 
			
			Zero
	
	
	
			
			
	
	
	
	
		
		
		Running
		
			on 
			
			Zero
	Commit 
							
							·
						
						e470146
	
1
								Parent(s):
							
							c795693
								
revert
Browse files
    	
        app.py
    CHANGED
    
    | 
         @@ -254,6 +254,9 @@ def start_training( 
     | 
|
| 254 | 
         
             
                    f"num_train_epochs={int(num_train_epochs)}",
         
     | 
| 255 | 
         
             
                    f"adam_weight_decay={adam_weight_decay}",
         
     | 
| 256 | 
         
             
                    f"adam_epsilon={adam_epsilon}",
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 257 | 
         
             
                    f"max_grad_norm={max_grad_norm}",
         
     | 
| 258 | 
         
             
                    f"lr_num_cycles={int(lr_num_cycles)}",
         
     | 
| 259 | 
         
             
                    f"lr_power={lr_power}",
         
     | 
| 
         @@ -279,12 +282,6 @@ def start_training( 
     | 
|
| 279 | 
         
             
                    commands.append(f"snr_gamma={snr_gamma}")
         
     | 
| 280 | 
         
             
                if scale_lr:
         
     | 
| 281 | 
         
             
                    commands.append("scale_lr")
         
     | 
| 282 | 
         
            -
                if prodigy_decouple:
         
     | 
| 283 | 
         
            -
                    commands.append("prodigy_decouple")
         
     | 
| 284 | 
         
            -
                if prodigy_use_bias_correction:
         
     | 
| 285 | 
         
            -
                    commands.append("prodigy_use_bias_correction")
         
     | 
| 286 | 
         
            -
                if prodigy_safeguard_warmup:
         
     | 
| 287 | 
         
            -
                    commands.append("prodigy_safeguard_warmup")
         
     | 
| 288 | 
         
             
                if with_prior_preservation:
         
     | 
| 289 | 
         
             
                    commands.append("with_prior_preservation")
         
     | 
| 290 | 
         
             
                    commands.append(f"class_prompt={class_prompt}")
         
     | 
| 
         | 
|
| 254 | 
         
             
                    f"num_train_epochs={int(num_train_epochs)}",
         
     | 
| 255 | 
         
             
                    f"adam_weight_decay={adam_weight_decay}",
         
     | 
| 256 | 
         
             
                    f"adam_epsilon={adam_epsilon}",
         
     | 
| 257 | 
         
            +
                    f"prodigy_decouple={prodigy_decouple}",
         
     | 
| 258 | 
         
            +
                    f"prodigy_use_bias_correction={prodigy_use_bias_correction}",
         
     | 
| 259 | 
         
            +
                    f"prodigy_safeguard_warmup={prodigy_safeguard_warmup}",
         
     | 
| 260 | 
         
             
                    f"max_grad_norm={max_grad_norm}",
         
     | 
| 261 | 
         
             
                    f"lr_num_cycles={int(lr_num_cycles)}",
         
     | 
| 262 | 
         
             
                    f"lr_power={lr_power}",
         
     | 
| 
         | 
|
| 282 | 
         
             
                    commands.append(f"snr_gamma={snr_gamma}")
         
     | 
| 283 | 
         
             
                if scale_lr:
         
     | 
| 284 | 
         
             
                    commands.append("scale_lr")
         
     | 
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 285 | 
         
             
                if with_prior_preservation:
         
     | 
| 286 | 
         
             
                    commands.append("with_prior_preservation")
         
     | 
| 287 | 
         
             
                    commands.append(f"class_prompt={class_prompt}")
         
     |