cancel
Showing results for 
Search instead for 
Did you mean: 
TonyD1
Cadet
Cadet
  • 374 Views

vLLM with CPU deployment with kServe

I want to deploy an LLM model ( from huggingface) thst is already in S3 bucket (it is a small model, 1b) Did someone achieve a deployment of LLM with CPU in RHOAI using kServe? There is a vLLM CPU, but i get different errors. I don't have GPU in my environment
Labels (4)
3 Replies
TudorRaduta
Community Manager
Community Manager
  • 347 Views

Chetan_Tiwary_
Community Manager
Community Manager
  • 131 Views

0 Kudos
Join the discussion
You must log in to join this conversation.