Skip to content

Commit 3d30a85

Browse files
committed
Fix result for fms generate inference
Signed-off-by: Flavia Beo <[email protected]>
1 parent da5696c commit 3d30a85

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

scripts/generate_layers_metrics.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -174,13 +174,13 @@ def __infer_layer(model, max_len, device, max_new_tokens, batch_size, tokenizer)
174174
contiguous_cache=True,
175175
extra_kwargs={},
176176
)
177+
result, timings = result
177178
if args.model_loader == "hf":
178179
result = model.generate(ids,
179180
max_length=max_seq_len,
180181
max_new_tokens=max_new_token,
181182
do_sample=do_sample,
182183
use_cache=use_cache)
183-
result, timings = result
184184
logger.info(f"Generation completed: Result len is {len(result)}")
185185
if len(result.shape) == 1:
186186
result = result.unsqueeze(0)

0 commit comments

Comments
 (0)