Run LLM inference on Cloud Run GPUs with Hugging Face Transformers.js

The following codelab shows how to run a backend service that runs the Transformers.js package . The Transformers.js package is functionally equivalent to the Hugging Face transformers python library together with Google's Gemma 2 model.

See the entire codelab at How to Run Transformers.js on Cloud Run GPUs .

Design a Mobile Site
View Site in Mobile | Classic
Share by: