Hugging Face has introduced Inference Providers, a new feature that allows developers to run AI models on third-party cloud services. Partnering with companies like SambaNova, Fal, Replicate, and Together AI, the platform now offers users the flexibility to deploy models on different infrastructures directly from their project pages.
Previously, Hugging Face primarily focused on its in-house AI hosting solutions, but the company is shifting towards a more collaborative approach. By integrating with external serverless providers, developers can now scale their models without managing hardware, making deployment easier and more cost-efficient. Users will pay standard provider rates, and Hugging Face Pro subscribers will receive additional free credits.
Since its founding in 2016, Hugging Face has grown into a leading AI model hub, backed by major investors like Google, Amazon, and Nvidia. With its latest move, the company continues to expand its ecosystem, making AI more accessible for developers worldwide.