Update README.md
Browse files
README.md
CHANGED
|
@@ -50,7 +50,8 @@ First, we introduce a VLM-based preference model **HPSv3**, trained on a "wide s
|
|
| 50 |
|
| 51 |
## β¨ Updates
|
| 52 |
|
| 53 |
-
- **[2025-
|
|
|
|
| 54 |
|
| 55 |
## π Table of Contents
|
| 56 |
1. [π Quick Start](#π-quick-start)
|
|
@@ -123,11 +124,11 @@ The demo will be available at `http://localhost:7860` and provides:
|
|
| 123 |
<img src="assets/gradio.png" alt="Gradio Demo" width="500"/>
|
| 124 |
</p>
|
| 125 |
|
| 126 |
-
## ποΈ Training
|
| 127 |
|
| 128 |
-
### π Dataset
|
| 129 |
|
| 130 |
-
|
|
|
|
|
|
|
| 131 |
|
| 132 |
Human Preference Dataset v3 (HPD v3) comprises 1.08M text-image pairs and 1.17M annotated pairwise data. To modeling the wide spectrum of human preference, we introduce newest state-of-the-art generative models and high quality real photographs while maintaining old models and lower quality real images.
|
| 133 |
|
|
@@ -156,30 +157,80 @@ Human Preference Dataset v3 (HPD v3) comprises 1.08M text-image pairs and 1.17M
|
|
| 156 |
| Curated HPDv2 | - | 327763 | - | Train |
|
| 157 |
</details>
|
| 158 |
|
| 159 |
-
|
| 160 |
-
```
|
| 161 |
HPDv3 is comming soon! Stay tuned!
|
| 162 |
-
```
|
| 163 |
-
<!-- ```bash
|
| 164 |
-
huggingface-cli download --repo-type dataset MizzenAI/HPDv3 --local-dir /your-local-dataset-path
|
| 165 |
``` -->
|
|
|
|
|
|
|
|
|
|
| 166 |
|
| 167 |
-
|
| 168 |
|
| 169 |
**Important Note: For simplicity, path1's image is always the prefered one**
|
| 170 |
|
| 171 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 172 |
[
|
| 173 |
-
|
| 174 |
-
|
| 175 |
-
"
|
| 176 |
-
"
|
| 177 |
-
"confidence": 0.
|
| 178 |
-
|
| 179 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 180 |
]
|
| 181 |
```
|
| 182 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 183 |
### π Training Command
|
| 184 |
|
| 185 |
```bash
|
|
|
|
| 50 |
|
| 51 |
## β¨ Updates
|
| 52 |
|
| 53 |
+
- **[2025-08-08]** π We release [HPDv3](https://huggingface.co/datasets/MizzenAI/HPDv3) dataset!.
|
| 54 |
+
- **[2025-08-06]** π We release HPSv3: inference code, training code, cohp code and [HPSv3 model weights](https://huggingface.co/MizzenAI/HPSv3). And [PyPI Package](https://pypi.org/project/hpsv3/).
|
| 55 |
|
| 56 |
## π Table of Contents
|
| 57 |
1. [π Quick Start](#π-quick-start)
|
|
|
|
| 124 |
<img src="assets/gradio.png" alt="Gradio Demo" width="500"/>
|
| 125 |
</p>
|
| 126 |
|
|
|
|
| 127 |
|
|
|
|
| 128 |
|
| 129 |
+
## π Dataset
|
| 130 |
+
|
| 131 |
+
### Human Preference Dataset v3
|
| 132 |
|
| 133 |
Human Preference Dataset v3 (HPD v3) comprises 1.08M text-image pairs and 1.17M annotated pairwise data. To modeling the wide spectrum of human preference, we introduce newest state-of-the-art generative models and high quality real photographs while maintaining old models and lower quality real images.
|
| 134 |
|
|
|
|
| 157 |
| Curated HPDv2 | - | 327763 | - | Train |
|
| 158 |
</details>
|
| 159 |
|
| 160 |
+
### Download HPDv3
|
| 161 |
+
<!-- ```
|
| 162 |
HPDv3 is comming soon! Stay tuned!
|
|
|
|
|
|
|
|
|
|
| 163 |
``` -->
|
| 164 |
+
```bash
|
| 165 |
+
huggingface-cli download --repo-type dataset MizzenAI/HPDv3 --local-dir /your-local-dataset-path
|
| 166 |
+
```
|
| 167 |
|
| 168 |
+
### Pairwise Training Data Format
|
| 169 |
|
| 170 |
**Important Note: For simplicity, path1's image is always the prefered one**
|
| 171 |
|
| 172 |
+
#### All Annotated Pairs (`all.json`)
|
| 173 |
+
|
| 174 |
+
**Important Notes: In HPDv3, we simply put the preferred sample at the first place (path1)**
|
| 175 |
+
|
| 176 |
+
`all.json` contains **all** annotated pairs except for test.
|
| 177 |
+
|
| 178 |
+
```bash
|
| 179 |
[
|
| 180 |
+
# samples from HPDv3 annotation pipeline
|
| 181 |
+
{
|
| 182 |
+
"prompt": "Description of the visual content or the generation prompt.",
|
| 183 |
+
"choice_dist": [12, 7], # Distribution of votes from annotators (12 votes for image1, 7 votes for image2)
|
| 184 |
+
"confidence": 0.9999907, # Confidence score reflecting preference reliability, based on annotators' capabilities (independent of choice_dist)
|
| 185 |
+
"path1": "images/uuid1.jpg", # File path to the preferred image
|
| 186 |
+
"path2": "images/uuid2.jpg", # File path to the non-preferred image
|
| 187 |
+
"model1": "flux", # Model used to generate the preferred image (path1)
|
| 188 |
+
"model2": "infinity" # Model used to generate the non-preferred image (path2)
|
| 189 |
+
},
|
| 190 |
+
# samples from Midjourney
|
| 191 |
+
{
|
| 192 |
+
"prompt": "Description of the visual content or the generation prompt.",
|
| 193 |
+
"choice_dist": null, # No distribution of votes Information from Discord
|
| 194 |
+
"confidence": null, # No Confidence Information from Discord
|
| 195 |
+
"path1": "images/uuid1.jpg", # File path to the preferred image.
|
| 196 |
+
"path2": "images/uuid2.jpg", # File path to the non-preferred image.
|
| 197 |
+
"model1": "midjourney", # Comparsion between images generated from midjourney
|
| 198 |
+
"model2": "midjourney" # Comparsion between images generated from midjourney
|
| 199 |
+
},
|
| 200 |
+
# samples from Curated HPDv2
|
| 201 |
+
{
|
| 202 |
+
"prompt": "Description of the visual content or the generation prompt.",
|
| 203 |
+
"choice_dist": null, # No distribution of votes Information from the original HPDv2 traindataset
|
| 204 |
+
"confidence": null, # No Confidence Information from the original HPDv2 traindataset
|
| 205 |
+
"path1": "images/uuid1.jpg", # File path to the preferred image.
|
| 206 |
+
"path2": "images/uuid2.jpg", # File path to the non-preferred image.
|
| 207 |
+
"model1": "hpdv2", # No specific model name in the original HPDv2 traindataset, set to hpdv2
|
| 208 |
+
"model2": "hpdv2" # No specific model name in the original HPDv2 traindataset, set to hpdv2
|
| 209 |
+
},
|
| 210 |
]
|
| 211 |
```
|
| 212 |
|
| 213 |
+
|
| 214 |
+
#### Train set (`train.json`)
|
| 215 |
+
We sample part of training data from `all.json` to build training dataset `train.json`. Moreover, to improve robustness, we integrate random sampled part of data from [Pick-a-pic](https://huggingface.co/datasets/pickapic-anonymous/pickapic_v1) and [ImageRewardDB](https://huggingface.co/datasets/zai-org/ImageRewardDB), which is `pickapic.json` and `imagereward.json`. For these two datasets, we only provide the pair infomation, and its corresponding image can be found in their official dataset repository.
|
| 216 |
+
|
| 217 |
+
|
| 218 |
+
#### Test Set (`test.json`)
|
| 219 |
+
```bash
|
| 220 |
+
[
|
| 221 |
+
{
|
| 222 |
+
"prompt": "Description of the visual content",
|
| 223 |
+
"path1": "images/uuid1.jpg", # Preferred sample
|
| 224 |
+
"path2": "images/uuid2.jpg", # Unpreferred sample
|
| 225 |
+
"model1": "flux", # Model used to generate the preferred sample (path1).
|
| 226 |
+
"model2": "infinity", # Model used to generate the non-preferred sample (path2).
|
| 227 |
+
|
| 228 |
+
}
|
| 229 |
+
]
|
| 230 |
+
```
|
| 231 |
+
|
| 232 |
+
## ποΈ Training
|
| 233 |
+
|
| 234 |
### π Training Command
|
| 235 |
|
| 236 |
```bash
|