Video: Deploy models with Hugging Face Inference Endpoints
In this video, I show you how to deploy Transformer models straight from the Hugging Face hub to managed infrastructure on AWS, in just a few clicks.
Starting from a model that I already trained for image classification, I first deploy an endpoint protected by Hugging Face token authentication. Then, I deploy a second endpoint in a private subnet, and I show you how to access it securely from your AWS account thanks to AWS PrivateLink.
Model: https://huggingface.co/juliensimon/autotrain-food101-1471154053
Inference Endpoints: https://huggingface.co/inference-endpoints
Inference Endpoints documentation: https://huggingface.co/docs/inference-endpoints/index
AWS PrivateLink documentation: https://docs.aws.amazon.com/vpc/latest/privatelink/concepts.html