All updates

All updates

icon
Filter
Product updates
Hugging Face launches inference as a service for AI deployment
Generative AI Infrastructure
Jul 29, 2024
This week:
Orijin raises seed funding for product development and expansion
Conservation Tech
Yesterday
Orijin raises seed funding for product development and expansion
Smart Farming
Yesterday
Product updates
Perplexity adds OpenAI o1 model and develops homepage widgets
Foundation Models
Yesterday
Partnerships
Nikola partners with WattEV to supply 22 BEVs
Truck Industry Tech
Yesterday
M&A
Zebra Technologies to acquire Photoneo from Photoneo Brightpick Group for undisclosed sum
Logistics Tech
Yesterday
Funding
Scope Technologies increases private placement offering to CAD 1.8 million
Machine Learning Infrastructure
Yesterday
Funding
Firefly Neuroscience raises USD 12.4 million in growth funding to commercialize technology
AI Drug Discovery
Dec 31, 2024
Listing
Nasdaq affirms delisting of OpGen after failed appeal
Precision Medicine
Dec 31, 2024
Funding
Rumble raises USD 775 million in strategic investment to support growth
Creator Economy
Dec 31, 2024
Product updates
InstaDeep releases open-source genomics AI model Nucleotide Transformers
Foundation Models
Dec 31, 2024
Generative AI Infrastructure

Generative AI Infrastructure

Jul 29, 2024

Hugging Face launches inference as a service for AI deployment

Product updates

  • Hugging Face has launched an inference-as-a-service product for AI deployment on NVIDIA's DGX Cloud. This service leverages NVIDIA NIM microservices to enhance token efficiency and provide access to popular AI models for developers.

  • The new service will deliver up to five times better token efficiency, enable immediate access to NVIDIA NIM microservices, and support leading AI models like Llama 3 and Mistral AI. Developers can prototype and deploy open-source AI models from the Hugging Face Hub, benefiting from serverless inference, increased flexibility, minimal infrastructure overhead, and optimized performance with NVIDIA NIM on the NVIDIA DGX Cloud.

  • Analyst QuickTake: Hugging Face offers integrated MLOps solutions, with a platform akin to GitHub for AI code repositories, models, and datasets. Launching its inference-as-a-service capabilities for foundation models such as LLama 3 and Mistral AI expands its capabilities into LLMOps solutions.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.