Spaces:
				
			
			
	
			
			
					
		Running
		
	
	
	
			
			
	
	
	
	
		
		
					
		Running
		
	
		Clémentine
		
	commited on
		
		
					Commit 
							
							·
						
						32b707a
	
1
								Parent(s):
							
							01ea22b
								
removed quantization to simplify
Browse files- src/about.py +4 -3
- src/backend/manage_requests.py +8 -9
- src/display/utils.py +9 -9
    	
        src/about.py
    CHANGED
    
    | @@ -8,15 +8,16 @@ class Task: | |
| 8 | 
             
                col_name: str
         | 
| 9 |  | 
| 10 |  | 
| 11 | 
            -
            #  | 
|  | |
| 12 | 
             
            class Tasks(Enum):
         | 
| 13 | 
             
                # task_key in the json file, metric_key in the json file, name to display in the leaderboard 
         | 
| 14 | 
             
                task0 = Task("anli_r1", "acc", "ANLI")
         | 
| 15 | 
             
                task1 = Task("logiqa", "acc_norm", "LogiQA")
         | 
| 16 |  | 
| 17 | 
            -
            TASKS_HARNESS = [task.value.benchmark for task in Tasks]
         | 
| 18 | 
            -
             | 
| 19 | 
             
            NUM_FEWSHOT = 0 # Change with your few shot
         | 
|  | |
|  | |
| 20 |  | 
| 21 |  | 
| 22 | 
             
            # Your leaderboard name
         | 
|  | |
| 8 | 
             
                col_name: str
         | 
| 9 |  | 
| 10 |  | 
| 11 | 
            +
            # Select your tasks here
         | 
| 12 | 
            +
            # ---------------------------------------------------
         | 
| 13 | 
             
            class Tasks(Enum):
         | 
| 14 | 
             
                # task_key in the json file, metric_key in the json file, name to display in the leaderboard 
         | 
| 15 | 
             
                task0 = Task("anli_r1", "acc", "ANLI")
         | 
| 16 | 
             
                task1 = Task("logiqa", "acc_norm", "LogiQA")
         | 
| 17 |  | 
|  | |
|  | |
| 18 | 
             
            NUM_FEWSHOT = 0 # Change with your few shot
         | 
| 19 | 
            +
            # ---------------------------------------------------
         | 
| 20 | 
            +
             | 
| 21 |  | 
| 22 |  | 
| 23 | 
             
            # Your leaderboard name
         | 
    	
        src/backend/manage_requests.py
    CHANGED
    
    | @@ -14,7 +14,7 @@ class EvalRequest: | |
| 14 | 
             
                json_filepath: str
         | 
| 15 | 
             
                weight_type: str = "Original"
         | 
| 16 | 
             
                model_type: str = ""  # pretrained, finetuned, with RL
         | 
| 17 | 
            -
                precision: str = ""  # float16, bfloat16 | 
| 18 | 
             
                base_model: Optional[str] = None # for adapter models
         | 
| 19 | 
             
                revision: str = "main" # commit
         | 
| 20 | 
             
                submitted_time: Optional[str] = "2022-05-18T11:40:22.519222"  # random date just so that we can still order requests by date
         | 
| @@ -28,11 +28,12 @@ class EvalRequest: | |
| 28 |  | 
| 29 | 
             
                    if self.precision in ["float16", "bfloat16"]:
         | 
| 30 | 
             
                        model_args += f",dtype={self.precision}"
         | 
| 31 | 
            -
                     | 
| 32 | 
            -
             | 
| 33 | 
            -
                     | 
| 34 | 
            -
             | 
| 35 | 
            -
                     | 
|  | |
| 36 | 
             
                        # A GPTQ model does not need dtype to be specified,
         | 
| 37 | 
             
                        # it will be inferred from the config
         | 
| 38 | 
             
                        pass
         | 
| @@ -42,9 +43,7 @@ class EvalRequest: | |
| 42 | 
             
                    return model_args
         | 
| 43 |  | 
| 44 |  | 
| 45 | 
            -
            def set_eval_request(
         | 
| 46 | 
            -
                api: HfApi, eval_request: EvalRequest, set_to_status: str, hf_repo: str, local_dir: str
         | 
| 47 | 
            -
            ):
         | 
| 48 | 
             
                """Updates a given eval request with its new status on the hub (running, completed, failed, ...)"""
         | 
| 49 | 
             
                json_filepath = eval_request.json_filepath
         | 
| 50 |  | 
|  | |
| 14 | 
             
                json_filepath: str
         | 
| 15 | 
             
                weight_type: str = "Original"
         | 
| 16 | 
             
                model_type: str = ""  # pretrained, finetuned, with RL
         | 
| 17 | 
            +
                precision: str = ""  # float16, bfloat16
         | 
| 18 | 
             
                base_model: Optional[str] = None # for adapter models
         | 
| 19 | 
             
                revision: str = "main" # commit
         | 
| 20 | 
             
                submitted_time: Optional[str] = "2022-05-18T11:40:22.519222"  # random date just so that we can still order requests by date
         | 
|  | |
| 28 |  | 
| 29 | 
             
                    if self.precision in ["float16", "bfloat16"]:
         | 
| 30 | 
             
                        model_args += f",dtype={self.precision}"
         | 
| 31 | 
            +
                    # Quantized models need some added config, the install of bits and bytes, etc
         | 
| 32 | 
            +
                    #elif self.precision == "8bit":
         | 
| 33 | 
            +
                    #    model_args += ",load_in_8bit=True"
         | 
| 34 | 
            +
                    #elif self.precision == "4bit":
         | 
| 35 | 
            +
                    #    model_args += ",load_in_4bit=True"
         | 
| 36 | 
            +
                    #elif self.precision == "GPTQ":
         | 
| 37 | 
             
                        # A GPTQ model does not need dtype to be specified,
         | 
| 38 | 
             
                        # it will be inferred from the config
         | 
| 39 | 
             
                        pass
         | 
|  | |
| 43 | 
             
                    return model_args
         | 
| 44 |  | 
| 45 |  | 
| 46 | 
            +
            def set_eval_request(api: HfApi, eval_request: EvalRequest, set_to_status: str, hf_repo: str, local_dir: str):
         | 
|  | |
|  | |
| 47 | 
             
                """Updates a given eval request with its new status on the hub (running, completed, failed, ...)"""
         | 
| 48 | 
             
                json_filepath = eval_request.json_filepath
         | 
| 49 |  | 
    	
        src/display/utils.py
    CHANGED
    
    | @@ -94,9 +94,9 @@ class WeightType(Enum): | |
| 94 | 
             
            class Precision(Enum):
         | 
| 95 | 
             
                float16 = ModelDetails("float16")
         | 
| 96 | 
             
                bfloat16 = ModelDetails("bfloat16")
         | 
| 97 | 
            -
                qt_8bit = ModelDetails("8bit")
         | 
| 98 | 
            -
                qt_4bit = ModelDetails("4bit")
         | 
| 99 | 
            -
                qt_GPTQ = ModelDetails("GPTQ")
         | 
| 100 | 
             
                Unknown = ModelDetails("?")
         | 
| 101 |  | 
| 102 | 
             
                def from_str(precision):
         | 
| @@ -104,12 +104,12 @@ class Precision(Enum): | |
| 104 | 
             
                        return Precision.float16
         | 
| 105 | 
             
                    if precision in ["torch.bfloat16", "bfloat16"]:
         | 
| 106 | 
             
                        return Precision.bfloat16
         | 
| 107 | 
            -
                    if precision in ["8bit"]:
         | 
| 108 | 
            -
             | 
| 109 | 
            -
                    if precision in ["4bit"]:
         | 
| 110 | 
            -
             | 
| 111 | 
            -
                    if precision in ["GPTQ", "None"]:
         | 
| 112 | 
            -
             | 
| 113 | 
             
                    return Precision.Unknown
         | 
| 114 |  | 
| 115 | 
             
            # Column selection
         | 
|  | |
| 94 | 
             
            class Precision(Enum):
         | 
| 95 | 
             
                float16 = ModelDetails("float16")
         | 
| 96 | 
             
                bfloat16 = ModelDetails("bfloat16")
         | 
| 97 | 
            +
                #qt_8bit = ModelDetails("8bit")
         | 
| 98 | 
            +
                #qt_4bit = ModelDetails("4bit")
         | 
| 99 | 
            +
                #qt_GPTQ = ModelDetails("GPTQ")
         | 
| 100 | 
             
                Unknown = ModelDetails("?")
         | 
| 101 |  | 
| 102 | 
             
                def from_str(precision):
         | 
|  | |
| 104 | 
             
                        return Precision.float16
         | 
| 105 | 
             
                    if precision in ["torch.bfloat16", "bfloat16"]:
         | 
| 106 | 
             
                        return Precision.bfloat16
         | 
| 107 | 
            +
                    #if precision in ["8bit"]:
         | 
| 108 | 
            +
                    #    return Precision.qt_8bit
         | 
| 109 | 
            +
                    #if precision in ["4bit"]:
         | 
| 110 | 
            +
                    #    return Precision.qt_4bit
         | 
| 111 | 
            +
                    #if precision in ["GPTQ", "None"]:
         | 
| 112 | 
            +
                    #    return Precision.qt_GPTQ
         | 
| 113 | 
             
                    return Precision.Unknown
         | 
| 114 |  | 
| 115 | 
             
            # Column selection
         | 
