Skip to content

This repository aims at setting up a Nvidia's Triton Inference server which simplifies the deployment of AI models at scale in production. It natively supports multiple framework backends like TensorFlow, PyTorch, ONNX Runtime, Python, and even custom backends.

Notifications You must be signed in to change notification settings

arangoml/embeddings-compute

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 

Repository files navigation

Compute Embeddings from ML models

This repository aims at setting up a Nvidia's Triton Inference server which simplifies the deployment of AI models at scale in production. It natively supports multiple framework backends like TensorFlow, PyTorch, ONNX Runtime, Python, and even custom backends.

About

This repository aims at setting up a Nvidia's Triton Inference server which simplifies the deployment of AI models at scale in production. It natively supports multiple framework backends like TensorFlow, PyTorch, ONNX Runtime, Python, and even custom backends.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published