Similar Tracks
Efficient Inference on MI300X: Our Journey at Microsoft, Rajat Monga, Microsoft, CVP AI Frameworks
AMD Developer Central
Intermediate English Practice | Improve Your Listening & Speaking | Learn English With Podcast
The English Pod Community
Exploring the Latency/Throughput & Cost Space for LLM Inference // Timothée Lacroix // CTO Mistral
MLOps.community
Mastering LLM Inference Optimization From Theory to Cost Effective Deployment: Mark Moyou
AI Engineer