Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
amd-shark
/
sdxl-quant-fp8
like
0
Follow
AMD SHARK
20
Model card
Files
Files and versions
xet
Community
1
cfd94d7
sdxl-quant-fp8
Ctrl+K
Ctrl+K
4 contributors
History:
19 commits
nickfraser
Added models that are fully quantized with FP8.
cfd94d7
11 months ago
all_linear_sym_8_calib8
Fix names
about 1 year ago
all_sym_8_calib10
MI250 QKV fused and all layers sym, FP8 attention, guidance scale 8, calib steps 10
about 1 year ago
brevitas
updated quant_params with QKV fusion
about 1 year ago
linear_conv_fp8_sdpa_fp16_eq_bl
Added models that are fully quantized with FP8.
11 months ago
linear_conv_fp8_sdpa_fp16_no_eq_bl
Added models that are fully quantized with FP8.
11 months ago
linear_conv_fp8_sdpa_fp8_eq_bl
Added models that are fully quantized with FP8.
11 months ago
linear_conv_fp8_sdpa_fp8_no_eq_bl
Added models that are fully quantized with FP8.
11 months ago
.gitattributes
Safe
2.08 kB
Added models that are fully quantized with FP8.
11 months ago
attn.py
Safe
6.26 kB
Added SDPA math model & test
about 1 year ago
sdxl.json
Safe
2.19 MB
Upload sdxl.json with huggingface_hub
about 1 year ago
sdxl.safetensors
Safe
5.14 GB
xet
Upload sdxl.safetensors with huggingface_hub
about 1 year ago
test_attn.py
1.29 kB
Added SDPA math model & test
about 1 year ago