Deploying and Scaling AI Applications with the NVIDIA TensorRT Inference Server on Kubernetes

Similar Tracks
Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong
The Linux Foundation
Keynote: Accelerating AI Workloads with GPUs in Kubernetes - Kevin Klues & Sanjay Chatterjee
CNCF [Cloud Native Computing Foundation]
GTC 2020: Deep into Triton Inference Server: BERT Practical Deployment on NVIDIA GPU
Bitcoin Standard
Use Nvidia’s DeepStream and Transfer Learning Toolkit to Deploy Streaming Analytics at Scale
NVIDIA Developer
Unlocking the Full Potential of GPUs for AI Workloads on Kubernetes - Kevin Klues, NVIDIA
CNCF [Cloud Native Computing Foundation]