Spaces:
Running
Running
XufengDuan
commited on
Commit
·
9ba63d6
1
Parent(s):
9dc9335
update scripts
Browse files
src/backend/evaluate_model.py
CHANGED
@@ -85,8 +85,18 @@ class Evaluator:
|
|
85 |
# exit()
|
86 |
# avg_summary_len = self.summary_generator.avg_length
|
87 |
# answer_rate = self.summary_generator.answer_rate
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
88 |
'''开始评估模型的结果'''
|
89 |
self.humanlike = self.eval_model.evaluate_humanlike(self.generated_summaries_df, envs.HUMAN_DATA, f"./generation_results/{self.model}.csv")
|
|
|
|
|
|
|
90 |
'''åŽŸå§‹æŒ‡æ ‡'''
|
91 |
# self.hallucination_scores, self.eval_results = self.eval_model.evaluate_hallucination(
|
92 |
# self.generated_summaries_df)
|
|
|
85 |
# exit()
|
86 |
# avg_summary_len = self.summary_generator.avg_length
|
87 |
# answer_rate = self.summary_generator.answer_rate
|
88 |
+
envs.API.upload_file(
|
89 |
+
path_or_fileobj=f"./generation_results/{self.model}.csv",
|
90 |
+
path_in_repo=f"{self.model}.csv",
|
91 |
+
repo_id=envs.RESULTS_REPO,
|
92 |
+
repo_type="dataset",
|
93 |
+
)
|
94 |
+
|
95 |
'''开始评估模型的结果'''
|
96 |
self.humanlike = self.eval_model.evaluate_humanlike(self.generated_summaries_df, envs.HUMAN_DATA, f"./generation_results/{self.model}.csv")
|
97 |
+
|
98 |
+
|
99 |
+
|
100 |
'''åŽŸå§‹æŒ‡æ ‡'''
|
101 |
# self.hallucination_scores, self.eval_results = self.eval_model.evaluate_hallucination(
|
102 |
# self.generated_summaries_df)
|
src/backend/model_operations.py
CHANGED
@@ -187,7 +187,7 @@ class SummaryGenerator:
|
|
187 |
# print(ID, q_ID, prompt_value)
|
188 |
system_prompt = envs.SYSTEM_PROMPT
|
189 |
_user_prompt = prompt_value
|
190 |
-
for ii in range(
|
191 |
# user_prompt = f"{envs.USER_PROMPT}\nPassage:\n{_source}"
|
192 |
while True:
|
193 |
try:
|
@@ -1242,6 +1242,14 @@ class EvaluationModel:
|
|
1242 |
'''coding llm data'''
|
1243 |
save_path = result_save_path.replace('.csv','_coding.csv')
|
1244 |
self.llm_df = self.code_results_llm(summaries_df)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1245 |
if save_path is not None:
|
1246 |
print(f'Save LLM coding results to {save_path}')
|
1247 |
fpath = Path(save_path)
|
|
|
187 |
# print(ID, q_ID, prompt_value)
|
188 |
system_prompt = envs.SYSTEM_PROMPT
|
189 |
_user_prompt = prompt_value
|
190 |
+
for ii in range(10):
|
191 |
# user_prompt = f"{envs.USER_PROMPT}\nPassage:\n{_source}"
|
192 |
while True:
|
193 |
try:
|
|
|
1242 |
'''coding llm data'''
|
1243 |
save_path = result_save_path.replace('.csv','_coding.csv')
|
1244 |
self.llm_df = self.code_results_llm(summaries_df)
|
1245 |
+
|
1246 |
+
envs.API.upload_file(
|
1247 |
+
path_or_fileobj=f"./generation_results/{self.model}_coding.csv",
|
1248 |
+
path_in_repo=f"{self.model}_coding.csv",
|
1249 |
+
repo_id=envs.RESULTS_REPO,
|
1250 |
+
repo_type="dataset",
|
1251 |
+
)
|
1252 |
+
|
1253 |
if save_path is not None:
|
1254 |
print(f'Save LLM coding results to {save_path}')
|
1255 |
fpath = Path(save_path)
|