This package contains SavedModel Exporter for TPU Inference of object detection models.
This Exporter is intended for users who have trained models with CPUs / GPUs, but would like to use them for inference on TPU without changing their code or re-training their models.
Users are assumed to have:
PIPELINE_CONFIG
: A pipeline_pb2.TrainEvalPipelineConfig config file;CHECKPOINT
: A model checkpoint trained on any device;
and need to correctly set:
EXPORT_DIR
: Path to export SavedModel;INPUT_PLACEHOLDER
: Name of input placeholder in model's signature_def_map;INPUT_TYPE
: Type of input node, which can be one of 'image_tensor', 'encoded_image_string_tensor', or 'tf_example';USE_BFLOAT16
: Whether to use bfloat16 instead of float32 on TPU.
The model can be exported with:
python object_detection/tpu_exporters/export_saved_model_tpu.py \
--pipeline_config_file=<PIPELINE_CONFIG> \
--ckpt_path=<CHECKPOINT> \
--export_dir=<EXPORT_DIR> \
--input_placeholder_name=<INPUT_PLACEHOLDER> \
--input_type=<INPUT_TYPE> \
--use_bfloat16=<USE_BFLOAT16>