Inference Providers documentation
Nscale
Nscale
Nscale is a vertically integrated AI cloud that delivers bespoke, sovereign AI infrastructure at scale.
Built on this foundation, Nscale’s inference service empowers developers with a wide range of models and ready-to-use inference services that integrate into workflows without the need to manage the underlying infrastructure.
Supported tasks
Chat Completion (LLM)
Find out more about Chat Completion (LLM) here.
Copied
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="nscale",
api_key="hf_xxxxxxxxxxxxxxxxxxxxxxxx",
)
completion = client.chat.completions.create(
model="Qwen/Qwen3-235B-A22B",
messages=[
{
"role": "user",
"content": "What is the capital of France?"
}
],
)
print(completion.choices[0].message)
Chat Completion (VLM)
Find out more about Chat Completion (VLM) here.
Copied
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="nscale",
api_key="hf_xxxxxxxxxxxxxxxxxxxxxxxx",
)
completion = client.chat.completions.create(
model="meta-llama/Llama-4-Scout-17B-16E-Instruct",
messages=[
{
"role": "user",
"content": [
{
"type": "text",
"text": "Describe this image in one sentence."
},
{
"type": "image_url",
"image_url": {
"url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
}
}
]
}
],
)
print(completion.choices[0].message)
Text To Image
Find out more about Text To Image here.
Copied
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="nscale",
api_key="hf_xxxxxxxxxxxxxxxxxxxxxxxx",
)
# output is a PIL.Image object
image = client.text_to_image(
"Astronaut riding a horse",
model="black-forest-labs/FLUX.1-schnell",
)