From 86238fe666f292a5a461b890729d99bdb9a971a2 Mon Sep 17 00:00:00 2001 From: Saifeddine ALOUI Date: Wed, 4 Sep 2024 13:58:11 +0200 Subject: [PATCH] Update personality.py --- lollms/personality.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/lollms/personality.py b/lollms/personality.py index bbd5da4..94677f9 100644 --- a/lollms/personality.py +++ b/lollms/personality.py @@ -749,7 +749,7 @@ def generate_with_images(self, prompt, images, max_size, temperature = None, top top_p=self.model_top_p if top_p is None else top_p, repeat_penalty=self.model_repeat_penalty if repeat_penalty is None else repeat_penalty, repeat_last_n = self.model_repeat_last_n if repeat_last_n is None else repeat_last_n - ).strip() + ) return self.bot_says def generate(self, prompt, max_size = None, temperature = None, top_k = None, top_p=None, repeat_penalty=None, repeat_last_n=None, callback=None, debug=False, show_progress=False ): @@ -769,7 +769,7 @@ def generate(self, prompt, max_size = None, temperature = None, top_k = None, to top_p=self.model_top_p if top_p is None else top_p, repeat_penalty=self.model_repeat_penalty if repeat_penalty is None else repeat_penalty, repeat_last_n = self.model_repeat_last_n if repeat_last_n is None else repeat_last_n, - ).strip() + ) if debug: self.print_prompt("prompt", prompt+self.bot_says) @@ -3935,7 +3935,7 @@ def mix_it_up(self, prompt: str, models, master_model, nb_rounds=2, max_generati return rounds_info - def answer(self, context_details, callback=None, send_full=True, custom_entries = ""): + def answer(self, context_details, custom_entries = "", send_full=True, callback=None): if context_details["is_continue"]: full_prompt = self.build_prompt_from_context_details(context_details, custom_entries=custom_entries, suppress= ["ai_prefix"]) else: