Pringled commited on
Commit
1fa0382
·
verified ·
1 Parent(s): 40e569e

add task relevance note below results table

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -62,6 +62,8 @@ Results on the [CoIR benchmark](https://github.com/CoIR-team/coir) (NDCG@10, `mt
62
  | potion-retrieval-32M | 32M | 32.10 | 4.22 | 31.80 | 36.71 | 45.11 | 38.64 | 29.97 | 32.62 | 8.70 | 56.26 | 36.93 |
63
  | potion-base-32M | 32M | 31.42 | 3.37 | 29.58 | 34.77 | 42.69 | 37.88 | 28.51 | 30.55 | 14.61 | 53.36 | 38.88 |
64
 
 
 
65
  ## Model Details
66
 
67
  | Property | Value |
 
62
  | potion-retrieval-32M | 32M | 32.10 | 4.22 | 31.80 | 36.71 | 45.11 | 38.64 | 29.97 | 32.62 | 8.70 | 56.26 | 36.93 |
63
  | potion-base-32M | 32M | 31.42 | 3.37 | 29.58 | 34.77 | 42.69 | 37.88 | 28.51 | 30.55 | 14.61 | 53.36 | 38.88 |
64
 
65
+ Not all CoIR tasks are equally relevant for the primary use case of retrieving code context given a natural language query. **CosQA** and **CodeFeedback (ST/MT)** are the most representative — they match a developer-style NL query against a code corpus with a genuine semantic gap. **COIRCodeSearchNetRetrieval** goes in the opposite direction (code query → text), and the **CodeTransOcean** tasks cover cross-language translation, a distinct problem. The hybrid row uses min-max score normalization with equal weighting (alpha=0.5) between dense and BM25 retrieval.
66
+
67
  ## Model Details
68
 
69
  | Property | Value |