Skip to content

Commit 8331875

Browse files
authored
Fix (#12390)
1 parent cb7b089 commit 8331875

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

docker/llm/serving/xpu/docker/vllm_offline_inference.py

+2
Original file line numberDiff line numberDiff line change
@@ -54,6 +54,8 @@
5454
disable_async_output_proc=True,
5555
distributed_executor_backend="ray",
5656
max_model_len=2000,
57+
trust_remote_code=True,
58+
block_size=8,
5759
max_num_batched_tokens=2000)
5860
# Generate texts from the prompts. The output is a list of RequestOutput objects
5961
# that contain the prompt, generated text, and other information.

0 commit comments

Comments
 (0)