Upload ./README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ Language: **English** | [**中文**](https://huggingface.co/Tencent-Hunyuan/Tens
|
|
12 |
|
13 |
We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
|
14 |
(faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
|
15 |
-
**TensorRT-
|
16 |
|
17 |
> ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
|
18 |
> We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
|
@@ -68,12 +68,22 @@ ln -s ${REMOTE_PATH} ./ckpts/t2i/model_trt/engine/model_onnx.plan
|
|
68 |
|
69 |
If you are using a different GPU, you can build the engine using the following command.
|
70 |
|
|
|
|
|
71 |
```shell
|
72 |
# Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
|
73 |
sh trt/build_engine.sh
|
74 |
```
|
75 |
|
76 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
77 |
|
78 |
### 4. Run the inference using the TensorRT model.
|
79 |
|
|
|
12 |
|
13 |
We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
|
14 |
(faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
|
15 |
+
**TensorRT-10.1.0.27** and **cuda (11.7 or 11.8)**.
|
16 |
|
17 |
> ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
|
18 |
> We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
|
|
|
68 |
|
69 |
If you are using a different GPU, you can build the engine using the following command.
|
70 |
|
71 |
+
##### Hunyuan-DiT v1.2
|
72 |
+
|
73 |
```shell
|
74 |
# Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
|
75 |
sh trt/build_engine.sh
|
76 |
```
|
77 |
|
78 |
+
##### Using Previous versions, Hunyuan-DiT <= v1.1
|
79 |
+
```shell
|
80 |
+
# v1.1
|
81 |
+
sh trt/build_engine.sh 1.1
|
82 |
+
# v1.0
|
83 |
+
sh trt/build_engine.sh 1.0
|
84 |
+
```
|
85 |
+
|
86 |
+
Finally, if you see the output like `&&&& PASSED TensorRT.trtexec [TensorRT v10100]`, the engine is built successfully.
|
87 |
|
88 |
### 4. Run the inference using the TensorRT model.
|
89 |
|