批量推理 #1247
-
while True: prompts = [] inputs = tokenizer(prompts, padding=True, truncation=True, return_tensors="pt") response = model.generate( |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
我们有batch推理的demo呀 |
Beta Was this translation helpful? Give feedback.
我们有batch推理的demo呀