test
Browse files- code_bleu.py +2 -2
code_bleu.py
CHANGED
|
@@ -69,7 +69,7 @@ class CodeBleu(evaluate.Metric):
|
|
| 69 |
],
|
| 70 |
)
|
| 71 |
|
| 72 |
-
def compute_codebleu_score(ground_truth, generated_answer, lang="python"):
|
| 73 |
"""
|
| 74 |
Function to compute CodeBLEU score between ground truth code and generated code
|
| 75 |
Has keywords for C, C#, C++, Go, Java, JavaScript, PHP, Python, Ruby, and Rust.
|
|
@@ -79,5 +79,5 @@ class CodeBleu(evaluate.Metric):
|
|
| 79 |
return result["codebleu"]
|
| 80 |
|
| 81 |
def _compute(self, references, predictions):
|
| 82 |
-
average_codebleu_score = sum([compute_codebleu_score(r, p) for r, p in zip(references, predictions)])/len(references)
|
| 83 |
return {"codebleu_score": average_codebleu_score}
|
|
|
|
| 69 |
],
|
| 70 |
)
|
| 71 |
|
| 72 |
+
def compute_codebleu_score(self, ground_truth, generated_answer, lang="python"):
|
| 73 |
"""
|
| 74 |
Function to compute CodeBLEU score between ground truth code and generated code
|
| 75 |
Has keywords for C, C#, C++, Go, Java, JavaScript, PHP, Python, Ruby, and Rust.
|
|
|
|
| 79 |
return result["codebleu"]
|
| 80 |
|
| 81 |
def _compute(self, references, predictions):
|
| 82 |
+
average_codebleu_score = sum([self.compute_codebleu_score(r, p) for r, p in zip(references, predictions)])/len(references)
|
| 83 |
return {"codebleu_score": average_codebleu_score}
|