Hugging Face puts squeeze on Nvidia's AI microservice play
Briefly

Hugging Face's HUGS provides containerized AI models for easier deployment on various hardware, potentially offering a more cost-effective and flexible alternative to Nvidia's NIMs.
Developers can use Hugging Face's HUGS to deploy large language models effortlessly via Docker or Kubernetes instead of dealing with complex configurations like vLLM or TensorRT.
While HUGS supports multiple hardware platforms, it currently does not support Intel Gaudi, focusing instead on compatibility with Nvidia, AMD GPUs, and other AI accelerators.
HUGS can be more cost-effective compared to NIMs, charging around $1 per hour per container on cloud platforms, making it an appealing option for larger deployments.
Read at Theregister
[
|
]