Model: Add tokens in responses
Signed-off-by: kingbri <bdashore3@proton.me>
This commit is contained in:
parent
7f18ea1d7c
commit
2ad79cb9ea
1 changed files with 1 additions and 1 deletions
2
model.py
2
model.py
|
|
@ -355,4 +355,4 @@ class ModelContainer:
|
|||
if eos or generated_tokens == max_tokens: break
|
||||
|
||||
elapsed_time = last_chunk_time - start_time
|
||||
print(f"Response generated in {round(elapsed_time, 2)} seconds ({round(generated_tokens / elapsed_time, 2)} T/s)")
|
||||
print(f"Response: {round(generated_tokens, 2)} tokens generated in {round(elapsed_time, 2)} seconds ({round(generated_tokens / elapsed_time, 2)} T/s)")
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue