NVIDIA
Published in

NVIDIA

How to deploy (almost) any Hugging face model 🤗 on NVIDIA’s Triton Inference Server with an application to Zero-Shot-Learning for Text Classification

SUMMARY

In this blog post, We examine Nvidia’s Triton Inference Server (formerly known as TensorRT Inference Server) which simplifies the deployment of AI models at scale in production. For the purpose of…

--

--

--

Solving the unsolvable with deep learning. Revolutionizing analytics. Breaking down barriers. Learn more about where AI is creating real impact today.

Recommended from Medium

Introducing the Quaterion: a framework for fine-tuning similarity learning models

“Nano-YOLO” — insights on the multi-part loss function of a simplified YOLO v1

How to run Keras model on Movidius neural compute stick

7 TIPS TO HELP YOU GET STARTED WITH MACHINE LEARNING

Akira’s Machine Learning news — #issue 31

Reach and Limits of the Supermassive Model GPT-3

Are We Ready for Machine Learning?

26 Words About Machine Learning, Every AI-Savvy Leader Must Know

Machine Learning

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Sachin Sharma

Sachin Sharma

Graph Machine Learning Research Engineer @ArangoDB Gmbh | Former AI/Machine Learning Scientist & Engineer @DefineMedia Gmbh | Former Research Intern @DFKI KL

More from Medium

Similarity Index and why ?

Building a Context-based Question Answering System on SQuAD 2.0

How to Summarize Financial News with Abstractive Summarization?

Deep learning in 2022