English
Zhiminli commited on
Commit
e54a587
·
verified ·
1 Parent(s): 842e698

Upload ./README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +12 -2
README.md CHANGED
@@ -12,7 +12,7 @@ Language: **English** | [**中文**](https://huggingface.co/Tencent-Hunyuan/Tens
12
 
13
  We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
14
  (faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
15
- **TensorRT-9.2.0.5** and **cuda (11.7 or 11.8)**.
16
 
17
  > ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
18
  > We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
@@ -68,12 +68,22 @@ ln -s ${REMOTE_PATH} ./ckpts/t2i/model_trt/engine/model_onnx.plan
68
 
69
  If you are using a different GPU, you can build the engine using the following command.
70
 
 
 
71
  ```shell
72
  # Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
73
  sh trt/build_engine.sh
74
  ```
75
 
76
- Finally, if you see the output like `&&&& PASSED TensorRT.trtexec [TensorRT v9200]`, the engine is built successfully.
 
 
 
 
 
 
 
 
77
 
78
  ### 4. Run the inference using the TensorRT model.
79
 
 
12
 
13
  We provide a TensorRT version of [HunyuanDiT](https://github.com/Tencent/HunyuanDiT) for inference acceleration
14
  (faster than flash attention). One can convert the torch model to TensorRT model using the following steps based on
15
+ **TensorRT-10.1.0.27** and **cuda (11.7 or 11.8)**.
16
 
17
  > ⚠️ Important Reminder (Suggestion for testing the TensorRT acceleration version):
18
  > We recommend users to test the TensorRT version on NVIDIA GPUs with Compute Capability >= 8.0,(For example, RTX4090,
 
68
 
69
  If you are using a different GPU, you can build the engine using the following command.
70
 
71
+ ##### Hunyuan-DiT v1.2
72
+
73
  ```shell
74
  # Build the TensorRT engine. By default, it will read the `ckpts` folder in the current directory.
75
  sh trt/build_engine.sh
76
  ```
77
 
78
+ ##### Using Previous versions, Hunyuan-DiT <= v1.1
79
+ ```shell
80
+ # v1.1
81
+ sh trt/build_engine.sh 1.1
82
+ # v1.0
83
+ sh trt/build_engine.sh 1.0
84
+ ```
85
+
86
+ Finally, if you see the output like `&&&& PASSED TensorRT.trtexec [TensorRT v10100]`, the engine is built successfully.
87
 
88
  ### 4. Run the inference using the TensorRT model.
89