Spaces:
Sleeping
Sleeping
victormiller
commited on
Commit
•
6e2a2df
1
Parent(s):
a55deb2
Update results.py
Browse files- results.py +1 -2
results.py
CHANGED
@@ -5,9 +5,8 @@ intro_div = Div(
|
|
5 |
H2("Perplexity Evaluation on Duplicate Data"),
|
6 |
H3("Model based Quality Estimation"),
|
7 |
quality_est = P("We took one of the model-based data quality evaluation strategies adopted by [DataComp-LM](https://arxiv.org/abs/2406.11794), which used perplexity filtering as a candidate for quality filtering. DataComp-LM followed [CCNet’s](https://arxiv.org/abs/1911.00359) practice to use a 5-gram Kneser-Ney model as implemented in the [KenLM](https://github.com/kpu/kenlm) library for efficient perplexity calculation. Following this practice, we estimated data quality by taking a KenLM model (from [edugp/kenlm](https://huggingface.co/edugp/kenlm)) trained on English Wikipedia data to compute perplexity on data with different duplication patterns. Lower perplexity is regarded as a signal of higher quality."),
|
8 |
-
H3("Sampling Strategy"),
|
9 |
-
sampling_strat = P("We started from a processed Common Crawl (CC) ablation dataset divided by the number of duplicates of each document. For each CC dump, we have different buckets each holding chunks of document with different duplicate count ranges (1-1, 2-5, 6-10, 11-100, 101-1000, 1001-30000000). We sampled the first 10k documents from each chunk with their meta data."),
|
10 |
)
|
|
|
11 |
perp1_div = Div(
|
12 |
H3("Perplexity vs Buckets"),
|
13 |
perp_v_buckets = P("For each bucket, we aggregated all the chunks that belong to a single year and calculated the average perplexity for each (bucket, year) data point."),
|
|
|
5 |
H2("Perplexity Evaluation on Duplicate Data"),
|
6 |
H3("Model based Quality Estimation"),
|
7 |
quality_est = P("We took one of the model-based data quality evaluation strategies adopted by [DataComp-LM](https://arxiv.org/abs/2406.11794), which used perplexity filtering as a candidate for quality filtering. DataComp-LM followed [CCNet’s](https://arxiv.org/abs/1911.00359) practice to use a 5-gram Kneser-Ney model as implemented in the [KenLM](https://github.com/kpu/kenlm) library for efficient perplexity calculation. Following this practice, we estimated data quality by taking a KenLM model (from [edugp/kenlm](https://huggingface.co/edugp/kenlm)) trained on English Wikipedia data to compute perplexity on data with different duplication patterns. Lower perplexity is regarded as a signal of higher quality."),
|
|
|
|
|
8 |
)
|
9 |
+
|
10 |
perp1_div = Div(
|
11 |
H3("Perplexity vs Buckets"),
|
12 |
perp_v_buckets = P("For each bucket, we aggregated all the chunks that belong to a single year and calculated the average perplexity for each (bucket, year) data point."),
|