[FEEDBACK] Inference Providers
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
Love that I can call DeepSeek R1 directly from the Hub ๐ฅ
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="together",
api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)
messages = [
{
"role": "user",
"content": "What is the capital of France?"
}
]
completion = client.chat.completions.create(
model="deepseek-ai/DeepSeek-R1",
messages=messages,
max_tokens=500
)
print(completion.choices[0].message)
Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
Thanks for your quick reply, good to know!
Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...
Could be good to add featherless.ai
TitanML !!
I would like to see Lambda as a provider available in the list!
https://lambdalabs.com/
OMG: "Model stabilityai/stable-diffusion-3.5-large-turbo does not exist"...
Please include kluster.ai as well! Thanks!
Important note about Pay-as-you-go (= PAYG) for Inference Providers: https://huggingface.co/posts/julien-c/158943939527784
For Inference Providers who have built support for our Billing API (currently: Fal, Novita, HF-Inference โ with more coming soon), we've started enabling Pay as you go (=PAYG)
What this means is that you can use those Inference Providers beyond the free included credits, and they're charged to your HF account.
Would love to have kluster.ai added - thanks!
I've been using kluster.ai and it is great, I would definitely like to see it added, thank you!
Hey folks we were recommended to post that Prodia (https://docs.prodia.com/) should become an inference provider as well because of how strongly we rank on Artifical Analysis for image gen models
Since yesterday, the ZeroGPU version of SD3.5 Large Turbo freezes out with ERROR (All the time...) Please Support, fix this problem as soon as you can...