Skip to content

Commit 7dc7e4c

Browse files
committed
Improved output
1 parent b69ce03 commit 7dc7e4c

File tree

2 files changed

+6
-4
lines changed

2 files changed

+6
-4
lines changed

multimode_llm.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -142,8 +142,6 @@ def do_chat(self, user_prompt: str, image: Image, system_prompt: str=None,
142142
new_response = self.tokenizer_stream.decode(new_token)
143143
response += new_response
144144

145-
print(new_response)
146-
147145
inferenceMs = int((time.perf_counter() - start_inference_time) * 1000)
148146

149147
del generator

multimode_llm_adapter.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -105,8 +105,12 @@ def long_process(self, data: RequestData) -> JSON:
105105

106106
generator.compute_logits()
107107
generator.generate_next_token()
108-
new_token = generator.get_next_tokens()[0]
109-
self.reply_text += tokenizer_stream.decode(new_token)
108+
109+
next_tokens = generator.get_next_tokens()
110+
next_token = next_tokens[0]
111+
next_response = tokenizer_stream.decode(next_token)
112+
113+
self.reply_text += next_response
110114

111115
inferenceMs : int = int((time.perf_counter() - start_inference_time) * 1000)
112116

0 commit comments

Comments
 (0)