Build
Browse files- build/torch24-cxx11-cu118-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx11-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch24-cxx11-cu121-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx11-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch24-cxx11-cu124-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx11-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch24-cxx98-cu118-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx98-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch24-cxx98-cu121-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx98-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch24-cxx98-cu124-x86_64-linux/quantization/__init__.py +2 -0
- build/torch24-cxx98-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx11-cu118-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx11-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx11-cu121-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx11-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx11-cu124-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx11-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx98-cu118-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx98-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx98-cu121-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx98-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
- build/torch25-cxx98-cu124-x86_64-linux/quantization/__init__.py +2 -0
- build/torch25-cxx98-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so +2 -2
build/torch24-cxx11-cu118-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx11-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2b7dbe4a5f68d853c46eab87bff39934ce9fb1695256c0d02c6f90ec7c4b257
|
3 |
+
size 87483216
|
build/torch24-cxx11-cu121-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx11-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80cc601f6714b2978b754a3a0581543042847d510cf4661a2697c04dddf31305
|
3 |
+
size 107014952
|
build/torch24-cxx11-cu124-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx11-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b748bbe0790f44d995713ef86468f361aadce2da8eb0113ba43a8bf96ea2580b
|
3 |
+
size 110873680
|
build/torch24-cxx98-cu118-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx98-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1369202bac8ef7fc82e7f94d66cd8d2b260cbc4e409f10a1b5b5e04f22131eb0
|
3 |
+
size 87453464
|
build/torch24-cxx98-cu121-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx98-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:45117d77739c22a5c65a18d6326c800c8300240b34926e3c5d8079f9d1c62619
|
3 |
+
size 107007424
|
build/torch24-cxx98-cu124-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch24-cxx98-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:399926eaf71ed4ec82c53f7ca5bcf90e024c2590ff44ca6af1b6f74156a6b6f2
|
3 |
+
size 110853632
|
build/torch25-cxx11-cu118-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx11-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df504168ad2d8eec02ddf328cdbbd50c436088e98c2347fd77f2c1b5fd2f1911
|
3 |
+
size 87483216
|
build/torch25-cxx11-cu121-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx11-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5779959f67b0de0b155ebc37142f0e415db5daa4b45281f3837636795150ddac
|
3 |
+
size 107014952
|
build/torch25-cxx11-cu124-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx11-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50de944be9760eeb8d97d8b06fd7e32d6ce4b5f2044aa38bd8171afe4ee3f346
|
3 |
+
size 110873680
|
build/torch25-cxx98-cu118-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx98-cu118-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cb12611bb467ef6bdd2228303a8beebe83ca15230abd1ebd385bcca40201d11
|
3 |
+
size 87453464
|
build/torch25-cxx98-cu121-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx98-cu121-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36df22427383e6c3907fdf83264b6d9a67579c258feba09d5e75c72e5db2ef0e
|
3 |
+
size 107007424
|
build/torch25-cxx98-cu124-x86_64-linux/quantization/__init__.py
CHANGED
@@ -13,6 +13,7 @@ from .marlin import (
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
|
|
16 |
|
17 |
__all__ = [
|
18 |
"awq_marlin_repack",
|
@@ -25,6 +26,7 @@ __all__ = [
|
|
25 |
"gptq_marlin_repack",
|
26 |
"marlin_gemm",
|
27 |
"marlin_qqq_gemm",
|
|
|
28 |
"scaled_fp8_quant",
|
29 |
"scaled_int8_quant",
|
30 |
]
|
|
|
13 |
marlin_qqq_gemm,
|
14 |
marlin_gemm,
|
15 |
)
|
16 |
+
from ._ops import ops
|
17 |
|
18 |
__all__ = [
|
19 |
"awq_marlin_repack",
|
|
|
26 |
"gptq_marlin_repack",
|
27 |
"marlin_gemm",
|
28 |
"marlin_qqq_gemm",
|
29 |
+
"ops",
|
30 |
"scaled_fp8_quant",
|
31 |
"scaled_int8_quant",
|
32 |
]
|
build/torch25-cxx98-cu124-x86_64-linux/quantization/_quantization_0_0_1.abi3.so
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77cbeca8bdda01afe774db01c4a817f20289cd4321e290887ded097c110046d7
|
3 |
+
size 110853464
|