How can I turn off the inference API for my model?
Specify inference: false in your model card’s metadata.
Why don’t I see an inference widget or why can’t I use the inference API?
For some tasks, there might not be support in the inference API, and, hence, there is no widget. For all libraries (except 🌍 Transformers), there is a mapping of library to supported tasks in the API. When a model repository has a task that is not supported by the repository library, the repository has inference: false by default.
Can I send large volumes of requests? Can I get accelerated APIs?
If you are interested in accelerated inference, higher volumes of requests, or an SLA, please contact us at api-enterprise at huggingface.co.