|
--- |
|
license: mit |
|
language: |
|
- en |
|
pipeline_tag: summarization |
|
library_name: transformers.js |
|
--- |
|
https://huggingface.co/JustinDu/BARTxiv with ONNX weights to be compatible with Transformers.js. |
|
|
|
## Steps to get ONNX quantized weights |
|
1. Use optimum-cli to convert the PyTorch weights to ONNX format. |
|
|
|
```bash |
|
optimum-cli export onnx --model JustinDu/BARTxiv BARTxiv_onnx |
|
``` |
|
|
|
2. Use the `quantize` script in the Transformers.js repo to quantize the ONNX weights. |
|
|
|
```bash |
|
python -m scripts.quantize --input_folder BARTxiv_onnx --output_folder BARTxiv_onnx/onnx |
|
``` |
|
|
|
Move the `.onnx` files in the `BARTxiv_onnx` folder to `BARTxiv_onnx/onnx` (these are the FP32 weights). |