Skip to content

Developer Tools

Advanced utilities for custom model development.

Overview

The inference_models.developer_tools module provides utilities for developers creating custom models that integrate with the inference_models package.

Base functions

Backend-Specific Utilities

CUDA Utilities

Low-level CUDA context management for custom models using CUDA/TensorRT.

ONNX Utilities

Utilities for working with ONNX Runtime in custom models.

PyTorch Utilities

Utilities for PyTorch-based custom models.

TensorRT Utilities

Utilities for TensorRT-based custom models.

Entities

Usage

Basic Usage

from inference_models.developer_tools import (
    get_model_package_contents,
    x_ray_runtime_environment,
    register_model_provider,
)

Backend-Specific Usage

Backend-specific utilities are available as lazy imports:

from inference_models.developer_tools import (
    use_primary_cuda_context,  # CUDA utilities
    set_onnx_execution_provider_defaults,  # ONNX utilities
    generate_batch_chunks,  # PyTorch utilities
    load_trt_model,  # TensorRT utilities
)

Lazy Loading

Backend-specific utilities are lazily loaded only when accessed. This means they won't cause import errors if the required dependencies (e.g., tensorrt, onnxruntime) are not installed, as long as you don't try to use them.