Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: cc-by-4.0
|
|
5 |
|
6 |
The BHI SISR Dataset's purpose is for training single image super-resolution models and is a result of tests on my BHI filtering method, which I made [a huggingface community blogpost about](https://huggingface.co/blog/Phips/bhi-filtering), which can be extremely summarized by that removing (by filtering) only the worst quality tiles from a training set has a way bigger positive effect on training metrics than keeping only the best quality training tiles.
|
7 |
|
8 |
-
It consists of 390'
|
9 |
|
10 |
<figure>
|
11 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/634e9aa407e669188d3912f9/bV0oaFKJzdsEqRme_lqU8.png" alt="48 first training tiles">
|
@@ -18,7 +18,7 @@ Though if an on-the-fly degradation pipeline is used during training, such a hig
|
|
18 |
Size on disc:
|
19 |
```
|
20 |
du BHI_HR
|
21 |
-
|
22 |
```
|
23 |
|
24 |
Also for the future, I am releasing the full dataset here. But there can of course be attempts in the future to make distilled versions of this dataset that perform better since I might find additional metrics or filtering methods in the future that might help reduce dataset size while achieving better training validation metric performance.
|
@@ -95,7 +95,7 @@ Well, JPEG XL would be the absolute best modern option for lossless compression,
|
|
95 |
|
96 |
TODO put paper stuff etc in here about webp / jpeg xl being superior concerning lossless compression
|
97 |
|
98 |
-
|
99 |
|
100 |
## Upload
|
101 |
|
|
|
5 |
|
6 |
The BHI SISR Dataset's purpose is for training single image super-resolution models and is a result of tests on my BHI filtering method, which I made [a huggingface community blogpost about](https://huggingface.co/blog/Phips/bhi-filtering), which can be extremely summarized by that removing (by filtering) only the worst quality tiles from a training set has a way bigger positive effect on training metrics than keeping only the best quality training tiles.
|
7 |
|
8 |
+
It consists of 390'035 images, which are all 512x512px dimensions and in the webp format.
|
9 |
|
10 |
<figure>
|
11 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/634e9aa407e669188d3912f9/bV0oaFKJzdsEqRme_lqU8.png" alt="48 first training tiles">
|
|
|
18 |
Size on disc:
|
19 |
```
|
20 |
du BHI_HR
|
21 |
+
131148100 BHI_HR/
|
22 |
```
|
23 |
|
24 |
Also for the future, I am releasing the full dataset here. But there can of course be attempts in the future to make distilled versions of this dataset that perform better since I might find additional metrics or filtering methods in the future that might help reduce dataset size while achieving better training validation metric performance.
|
|
|
95 |
|
96 |
TODO put paper stuff etc in here about webp / jpeg xl being superior concerning lossless compression
|
97 |
|
98 |
+
See file list for all the files in the dataset
|
99 |
|
100 |
## Upload
|
101 |
|