NCTCMumbai's picture
Upload 2583 files
97b6013 verified

A newer version of the Gradio SDK is available: 5.25.2

Upgrade

Export trained deeplab model to frozen inference graph

After model training finishes, you could export it to a frozen TensorFlow inference graph proto. Your trained model checkpoint usually includes the following files:

  • model.ckpt-${CHECKPOINT_NUMBER}.data-00000-of-00001,
  • model.ckpt-${CHECKPOINT_NUMBER}.index
  • model.ckpt-${CHECKPOINT_NUMBER}.meta

After you have identified a candidate checkpoint to export, you can run the following commandline to export to a frozen graph:

# From tensorflow/models/research/
# Assume all checkpoint files share the same path prefix `${CHECKPOINT_PATH}`.
python deeplab/export_model.py \
    --checkpoint_path=${CHECKPOINT_PATH} \
    --export_path=${OUTPUT_DIR}/frozen_inference_graph.pb

Please also add other model specific flags as you use for training, such as model_variant, add_image_level_feature, etc.