Darius Knowledge Hub

Home

❯

TensorRT (+ Triton Inference Server)

TensorRT (+ Triton Inference Server)

Jan 21, 20261 min read

  • knowledge
  • tools
  • TinyML

TensorRT (+ Triton Inference Server)


Created: 03 Jan 2023, 11:23 AM | Modified: =dateformat(this.file.mtime,"dd MMM yyyy, hh:mm a") Tags: knowledge, tools, TinyML


Triton ⇒ serve model over HTTP or gRPC (i.e. over some network, hosted somewhere else?)

https://www.nvidia.com/en-us/on-demand/session/gtcspring21-se2690/

Triton will require a Triton Server and a Triton Client

https://towardsdatascience.com/serving-tensorrt-models-with-nvidia-triton-inference-server-5b68cc141d19


Graph View

Backlinks

  • Darius Knowledge Hub

Created with Quartz v4.5.2 © 2026

  • GitHub
  • Discord Community