Please refer to the Inference Providers Documentation for detailed information.
HF-Inference API is one of the many providers available on the Hugging Face Hub. It is deployed by Hugging Face ourselves, using text-generation-inference for LLMs for instance. This service used to be called “Inference API (serverless)” prior to Inference Providers.
For more details about the HF-Inference API, check out it's dedicated page.
The HF-Inference API is powered by Inference Endpoints under the hood.
For some tasks, there might not be support by any Inference Provider, and hence, there is no widget.
To check usage across all providers, check out your billing page.
To check your HF-Inference usage specifically, check out the Inference Dashboard. The dashboard shows both your serverless and dedicated endpoints usage.
Yes! We provide client wrappers in both JS and Python: