Skip to content

Releases: triton-inference-server/model_navigator

Triton Model Navigator v0.5.5

24 May 13:21
Compare
Choose a tag to compare
  • new: Public nav.utilities module with UnpackedDataloader wrapper
  • new: Added support for strict flag in Torch custom config
  • new: Extended TensorRT custom config to support builder optimization level and hardware compatibility flags
  • fix: Invalid optimal shape calculation for odd values in max batch size

Triton Model Navigator v0.5.4

18 May 11:39
Compare
Choose a tag to compare
  • new: Custom implementation for ONNX and TensorRT runners

  • new: Use CUDA 12 for JAX in unit tests and functional tests

  • new: Step-by-step examples

  • new: Updated documentation

  • new: TensorRTCUDAGraph runner introduced with support for CUDA graphs

  • fix: Optimal shape not set correctly during adaptive conversion

  • fix: Find max batch size command for JAX

  • fix: Save stdout to logfiles in debug mode

  • Version of external components used during testing:

Triton Model Navigator v0.5.3

19 Apr 13:09
Compare
Choose a tag to compare
  • fix: filter outputs using output_metadata in ONNX runners

Triton Model Navigator v0.5.2

11 Apr 12:28
Compare
Choose a tag to compare

Triton Model Navigator v0.5.1

29 Mar 16:12
Compare
Choose a tag to compare

Triton Model Navigator v0.5.0

23 Mar 09:21
Compare
Choose a tag to compare
  • new: Support for PyTriton deployemnt

  • new: Support for Python models with python.optimize API

  • new: PyTorch 2 compile CPU and CUDA runners

  • new: Collect conversion max batch size in status

  • new: PyTorch runners with compile support

  • change: Improved handling CUDA and CPU runners

  • change: Reduced finding device max batch size time by running it once as separate pipeline

  • change: Stored find max batch size result in separate filed in status

  • Version of external components used during testing:

Triton Model Navigator v0.4.4

14 Mar 15:37
Compare
Choose a tag to compare
  • fix: when exporting single input model to saved model, unwrap one element list with inputs

Triton Model Navigator v0.4.3

13 Mar 15:50
Compare
Choose a tag to compare
  • fix: in Keras inference use model.predict(tensor) for single input models

Triton Model Navigator v0.4.2

09 Mar 11:53
Compare
Choose a tag to compare
  • fix: loading configuration for trt_profile from package
  • fix: missing reproduction scripts and logs inside package
  • fix: invalid model path in reproduction script for ONNX to TRT conversion
  • fix: collecting metadata from ONNX model in main thread during ONNX to TRT conversion

Triton Model Navigator v0.4.1

22 Feb 10:52
Compare
Choose a tag to compare
  • fix: when specified use dynamic axes from custom OnnxConfig