Daiqing commited on
Commit
f81d842
1 Parent(s): 76c0dd5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -1
README.md CHANGED
@@ -67,6 +67,11 @@ During the user study, we give users instructions to evaluate image pairs based
67
 
68
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63855d851769b7c4b10e1f76/o3Bt62qFsTO9DkeX2yLua.png)
69
 
 
 
 
 
 
70
  We introduce a new benchmark, [MJHQ-30K](https://huggingface.co/datasets/playgroundai/MJHQ30K), for automatic evaluation of a model’s aesthetic quality. The benchmark computes FID on a high-quality dataset to gauge aesthetic quality.
71
 
72
  We curate the high-quality dataset from Midjourney with 10 common categories, each category with 3K samples. Following common practice, we use aesthetic score and CLIP score to ensure high image quality and high image-text alignment. Furthermore, we take extra care to make the data diverse within each category.
@@ -77,6 +82,12 @@ We release this benchmark to the public and encourage the community to adopt it
77
 
78
  ### Base Models for all resolution
79
 
80
- < INSERT TABLE HERE >
 
 
 
 
 
 
81
 
82
  Apart from playground-v2-1024px-aesthetic, we release all intermediate checkpoints at different training stages to the community in order to foster foundation model research in pixels. Here, we report the FID score and CLIP score on the MSCOCO14 evaluation set for the reference purposes. (Note that our reported numbers may differ from the numbers reported in SDXL’s published results, as our prompt list may be different.)
 
67
 
68
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63855d851769b7c4b10e1f76/o3Bt62qFsTO9DkeX2yLua.png)
69
 
70
+ | Model | Overall FID |
71
+ | ------------------------------------- | ----- |
72
+ | SDXL-1-0-refiner | 9.55 |
73
+ | playground-v2-1024px-aesthetic | **7.07** |
74
+
75
  We introduce a new benchmark, [MJHQ-30K](https://huggingface.co/datasets/playgroundai/MJHQ30K), for automatic evaluation of a model’s aesthetic quality. The benchmark computes FID on a high-quality dataset to gauge aesthetic quality.
76
 
77
  We curate the high-quality dataset from Midjourney with 10 common categories, each category with 3K samples. Following common practice, we use aesthetic score and CLIP score to ensure high image quality and high image-text alignment. Furthermore, we take extra care to make the data diverse within each category.
 
82
 
83
  ### Base Models for all resolution
84
 
85
+ | Model | FID | Clip Score |
86
+ | ---------------------------- | ------ | ---------- |
87
+ | SDXL-1-0-refiner | 13.04 | 32.62 |
88
+ | [playground-v2-256px-base](https://huggingface.co/playgroundai/playground-v2-256px-base) | 9.83 | 31.90 |
89
+ | [playground-v2-512px-base](https://huggingface.co/playgroundai/playground-v2-512px-base) | 9.55 | 32.08 |
90
+ | [playground-v2-1024px-base](https://huggingface.co/playgroundai/playground-v2-1024px-base) | 9.97 | 31.90 |
91
+
92
 
93
  Apart from playground-v2-1024px-aesthetic, we release all intermediate checkpoints at different training stages to the community in order to foster foundation model research in pixels. Here, we report the FID score and CLIP score on the MSCOCO14 evaluation set for the reference purposes. (Note that our reported numbers may differ from the numbers reported in SDXL’s published results, as our prompt list may be different.)