Skip to content

Release 1.5.0, corresponding to NGC container 19.08

Compare
Choose a tag to compare
@dzier dzier released this 03 Sep 23:53

NVIDIA TensorRT Inference Server

The NVIDIA TensorRT Inference Server provides a cloud inferencing solution optimized for NVIDIA GPUs. The server provides an inference service via an HTTP or GRPC endpoint, allowing remote clients to request inferencing for any model being managed by the server.

What's New In 1.5.0

  • Added a new execution mode allows the inference server to start without
    loading any models from the model repository. Model loading and unloading
    is then controlled by a new GRPC/HTTP model control API.

  • Added a new instance-group mode allows TensorFlow models that explicitly
    distribute inferencing across multiple GPUs to run in that manner in the
    inference server.

  • Improved input/output tensor reshape to allow variable-sized dimensions in
    tensors being reshaped.

  • Added a C++ wrapper around the custom backend C API to simplify the creation
    of custom backends. This wrapper is included in the custom backend SDK.

  • Improved the accuracy of the compute statistic reported for inference
    requests. Previously the compute statistic included some additional time
    beyond the actual compute time.

  • The performance client, perf_client, now reports more information for ensemble
    models, including statistics for all contained models and the entire ensemble.

Client Libraries and Examples

Ubuntu 16.04 and Ubuntu 18.04 builds of the client libraries and examples are included in this release in the attached v1.5.0_ubuntu1604.clients.tar.gz and v1.5.0_ubuntu1804.clients.tar.gz files. See the documentation section 'Building the Client Libraries and Examples' for more information on using these files.

Custom Backend SDK

Ubuntu 16.04 and Ubuntu 18.04 builds of the custom backend SDK are included in this release in the attached v1.5.0_ubuntu1604.custombackend.tar.gz and v1.5.0_ubuntu1804.custombackend.tar.gz files. See the documentation section 'Building a Custom Backend' for more information on using these files.