Skip to content

Commit 793370d

Browse files
committed
Fix snippets CI
gg @Wauplin
1 parent f1a7d20 commit 793370d

File tree

7 files changed

+7
-7
lines changed

7 files changed

+7
-7
lines changed

packages/tasks-gen/snippets-fixtures/conversational-llm-non-stream/1.openai.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
]
1414

1515
completion = client.chat.completions.create(
16-
model="meta-llama/Llama-3.1-8B-Instruct",
16+
model="meta-llama/Llama-3.1-8B-Instruct",
1717
messages=messages,
1818
max_tokens=500
1919
)

packages/tasks-gen/snippets-fixtures/conversational-llm-non-stream/1.openai.together.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
]
1414

1515
completion = client.chat.completions.create(
16-
model="meta-llama/Llama-3.1-8B-Instruct",
16+
model="meta-llama/Llama-3.1-8B-Instruct",
1717
messages=messages,
1818
max_tokens=500
1919
)

packages/tasks-gen/snippets-fixtures/conversational-llm-stream/0.huggingface_hub.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
]
1414

1515
stream = client.chat.completions.create(
16-
model="meta-llama/Llama-3.1-8B-Instruct",
16+
model="meta-llama/Llama-3.1-8B-Instruct",
1717
messages=messages,
1818
max_tokens=500,
1919
stream=True

packages/tasks-gen/snippets-fixtures/conversational-llm-stream/1.openai.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -20,4 +20,4 @@
2020
)
2121

2222
for chunk in stream:
23-
print(chunk.choices[0].delta.content, end="")
23+
print(chunk.choices[0].delta.content, end="")

packages/tasks-gen/snippets-fixtures/conversational-vlm-non-stream/1.openai.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@
2424
]
2525

2626
completion = client.chat.completions.create(
27-
model="meta-llama/Llama-3.2-11B-Vision-Instruct",
27+
model="meta-llama/Llama-3.2-11B-Vision-Instruct",
2828
messages=messages,
2929
max_tokens=500
3030
)

packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/0.huggingface_hub.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@
2424
]
2525

2626
stream = client.chat.completions.create(
27-
model="meta-llama/Llama-3.2-11B-Vision-Instruct",
27+
model="meta-llama/Llama-3.2-11B-Vision-Instruct",
2828
messages=messages,
2929
max_tokens=500,
3030
stream=True

packages/tasks-gen/snippets-fixtures/conversational-vlm-stream/1.openai.hf-inference.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -31,4 +31,4 @@
3131
)
3232

3333
for chunk in stream:
34-
print(chunk.choices[0].delta.content, end="")
34+
print(chunk.choices[0].delta.content, end="")

0 commit comments

Comments
 (0)