We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 815dcb9 commit f787207Copy full SHA for f787207
examples/offline_inference_npu.py
@@ -37,7 +37,7 @@ def main():
37
# Create a sampling params object.
38
sampling_params = SamplingParams(max_tokens=100, temperature=0.0)
39
# Create an LLM.
40
- llm = LLM(model="Qwen/Qwen2.5-0.5B-Instruct", enforce_eager=True)
+ llm = LLM(model="Qwen/Qwen2.5-0.5B-Instruct")
41
42
# Generate texts from the prompts.
43
outputs = llm.generate(prompts, sampling_params)
0 commit comments