It's Pre:Invent time! Excited to announce that HUGS is now supported and available on Amazon Web Services (AWS) Inferentia2! HUGS or Hugging Face Generative AI Services are optimized, zero-configuration inference microservices designed to simplify and accelerate the development of AI applications using open models. Built on open-source technologies. 👀 Think on NVIDIA NIMs, but on AWS Neuron Accelerators. No compilation or configuration headaches. Take your open model and run it without any pain on AWS Accelerators with in minutes. 😍 Starting today, HUGS on AWS Neuron supports Meta Llama, Nous Research Hermes, Mistral, and Mixtral models, with more to come! 🚀 Want to get started? Check out our Guide on how to deploy Llama 3.1 on AWS EKS with AWS Inferentia2 instances: https://lnkd.in/e24dYp_z
That is HUGE !! Congrats to the entire team on making HUGS available on Inferentia & Trainium!
Awesome product!
Machine Learning Engineer | GenAI | MLOps | Federated Learning Enthusiast
2wAwesome work Philipp Schmid ! Do you plan to include Multimodal stuff in near future ?