Skip to content

Commit 72ae095

Browse files
Update Inference Providers documentation (automated) (#1720)
Co-authored-by: Wauplin <[email protected]>
1 parent 5e616a9 commit 72ae095

File tree

7 files changed

+22
-39
lines changed

7 files changed

+22
-39
lines changed

docs/inference-providers/providers/hf-inference.md

+2-12
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ Find out more about Chat Completion (LLM) [here](../tasks/chat-completion).
5555

5656
<InferenceSnippet
5757
pipeline=text-generation
58-
providersMapping={ {"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"} } }
58+
providersMapping={ {"hf-inference":{"modelId":"Qwen/Qwen3-235B-A22B","providerModelId":"Qwen/Qwen3-235B-A22B"} } }
5959
conversational />
6060

6161

@@ -109,16 +109,6 @@ Find out more about Image Segmentation [here](../tasks/image_segmentation).
109109
/>
110110

111111

112-
### Summarization
113-
114-
Find out more about Summarization [here](../tasks/summarization).
115-
116-
<InferenceSnippet
117-
pipeline=summarization
118-
providersMapping={ {"hf-inference":{"modelId":"pszemraj/led-base-book-summary","providerModelId":"pszemraj/led-base-book-summary"} } }
119-
/>
120-
121-
122112
### Text Classification
123113

124114
Find out more about Text Classification [here](../tasks/text_classification).
@@ -135,7 +125,7 @@ Find out more about Text Generation [here](../tasks/text_generation).
135125

136126
<InferenceSnippet
137127
pipeline=text-generation
138-
providersMapping={ {"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"} } }
128+
providersMapping={ {"hf-inference":{"modelId":"Qwen/Qwen3-235B-A22B","providerModelId":"Qwen/Qwen3-235B-A22B"} } }
139129
/>
140130

141131

docs/inference-providers/tasks/audio-classification.md

-4
Original file line numberDiff line numberDiff line change
@@ -50,10 +50,6 @@ No snippet available for this task.
5050

5151
| Payload | | |
5252
| :--- | :--- | :--- |
53-
| **inputs*** | _string_ | The input audio data as a base64-encoded string. If no `parameters` are provided, you can also provide the audio data as a raw bytes payload. |
54-
| **parameters** | _object_ | |
55-
| **&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;function_to_apply** | _enum_ | Possible values: sigmoid, softmax, none. |
56-
| **&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;top_k** | _integer_ | When specified, limits the output to the top K most probable classes. |
5753

5854

5955
#### Response

docs/inference-providers/tasks/chat-completion.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ The API supports:
5959

6060
<InferenceSnippet
6161
pipeline=text-generation
62-
providersMapping={ {"cerebras":{"modelId":"meta-llama/Llama-3.3-70B-Instruct","providerModelId":"llama-3.3-70b"},"cohere":{"modelId":"CohereLabs/c4ai-command-a-03-2025","providerModelId":"command-a-03-2025"},"fireworks-ai":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"accounts/fireworks/models/deepseek-v3-0324"},"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"},"hyperbolic":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek-ai/DeepSeek-V3-0324"},"nebius":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek-ai/DeepSeek-V3-0324-fast"},"novita":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek/deepseek-v3-0324"},"sambanova":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"DeepSeek-V3-0324"},"together":{"modelId":"deepseek-ai/DeepSeek-R1","providerModelId":"deepseek-ai/DeepSeek-R1"}} }
62+
providersMapping={ {"cerebras":{"modelId":"meta-llama/Llama-3.3-70B-Instruct","providerModelId":"llama-3.3-70b"},"cohere":{"modelId":"CohereLabs/c4ai-command-a-03-2025","providerModelId":"command-a-03-2025"},"fireworks-ai":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"accounts/fireworks/models/deepseek-v3-0324"},"hf-inference":{"modelId":"Qwen/Qwen3-235B-A22B","providerModelId":"Qwen/Qwen3-235B-A22B"},"hyperbolic":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek-ai/DeepSeek-V3-0324"},"nebius":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek-ai/DeepSeek-V3-0324-fast"},"novita":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"deepseek/deepseek-v3-0324"},"sambanova":{"modelId":"deepseek-ai/DeepSeek-V3-0324","providerModelId":"DeepSeek-V3-0324"},"together":{"modelId":"deepseek-ai/DeepSeek-R1","providerModelId":"deepseek-ai/DeepSeek-R1"}} }
6363
conversational />
6464

6565

docs/inference-providers/tasks/summarization.md

+1-4
Original file line numberDiff line numberDiff line change
@@ -30,10 +30,7 @@ Explore all available models and find the one that suits you best [here](https:/
3030
### Using the API
3131

3232

33-
<InferenceSnippet
34-
pipeline=summarization
35-
providersMapping={ {"hf-inference":{"modelId":"pszemraj/led-base-book-summary","providerModelId":"pszemraj/led-base-book-summary"}} }
36-
/>
33+
No snippet available for this task.
3734

3835

3936

docs/inference-providers/tasks/text-generation.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ Explore all available models and find the one that suits you best [here](https:/
3939

4040
<InferenceSnippet
4141
pipeline=text-generation
42-
providersMapping={ {"hf-inference":{"modelId":"Qwen/QwQ-32B","providerModelId":"Qwen/QwQ-32B"},"together":{"modelId":"deepseek-ai/DeepSeek-R1","providerModelId":"deepseek-ai/DeepSeek-R1"}} }
42+
providersMapping={ {"hf-inference":{"modelId":"Qwen/Qwen3-235B-A22B","providerModelId":"Qwen/Qwen3-235B-A22B"},"together":{"modelId":"deepseek-ai/DeepSeek-R1","providerModelId":"deepseek-ai/DeepSeek-R1"}} }
4343
/>
4444

4545

scripts/inference-providers/package.json

+2-2
Original file line numberDiff line numberDiff line change
@@ -14,8 +14,8 @@
1414
"author": "",
1515
"license": "ISC",
1616
"dependencies": {
17-
"@huggingface/inference": "^3.9.2",
18-
"@huggingface/tasks": "^0.18.11",
17+
"@huggingface/inference": "^3.10.0",
18+
"@huggingface/tasks": "^0.19.0",
1919
"@types/node": "^22.5.0",
2020
"handlebars": "^4.7.8",
2121
"node": "^20.17.0",

scripts/inference-providers/pnpm-lock.yaml

+15-15
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)