File size: 1,176 Bytes
18ddfe2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 |
# Object Detection TPU Inference Exporter
This package contains SavedModel Exporter for TPU Inference of object detection
models.
## Usage
This Exporter is intended for users who have trained models with CPUs / GPUs,
but would like to use them for inference on TPU without changing their code or
re-training their models.
Users are assumed to have:
+ `PIPELINE_CONFIG`: A pipeline_pb2.TrainEvalPipelineConfig config file;
+ `CHECKPOINT`: A model checkpoint trained on any device;
and need to correctly set:
+ `EXPORT_DIR`: Path to export SavedModel;
+ `INPUT_PLACEHOLDER`: Name of input placeholder in model's signature_def_map;
+ `INPUT_TYPE`: Type of input node, which can be one of 'image_tensor',
'encoded_image_string_tensor', or 'tf_example';
+ `USE_BFLOAT16`: Whether to use bfloat16 instead of float32 on TPU.
The model can be exported with:
```
python object_detection/tpu_exporters/export_saved_model_tpu.py \
--pipeline_config_file=<PIPELINE_CONFIG> \
--ckpt_path=<CHECKPOINT> \
--export_dir=<EXPORT_DIR> \
--input_placeholder_name=<INPUT_PLACEHOLDER> \
--input_type=<INPUT_TYPE> \
--use_bfloat16=<USE_BFLOAT16>
```
|