From cb874b137e34c93d9d0ac0c9e704b4f83f1272f2 Mon Sep 17 00:00:00 2001 From: Saifeddine ALOUI Date: Sat, 15 Feb 2025 16:11:46 +0100 Subject: [PATCH] enhanced --- configs/config.yaml | 4 ++-- lollms/configs/config.yaml | 4 ++-- lollms/server/events/lollms_generation_events.py | 2 -- lollms/server/events/lollms_personality_events.py | 1 - setup.py | 2 +- 5 files changed, 5 insertions(+), 8 deletions(-) diff --git a/configs/config.yaml b/configs/config.yaml index af3a317..09ccbaf 100644 --- a/configs/config.yaml +++ b/configs/config.yaml @@ -339,7 +339,7 @@ current_language: english fun_mode: false think_first_mode: false thinking_prompt: "Use a think first process to answer the user: - + Ask yourself about the user's request and answer it with logical details. If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly. If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer. @@ -351,7 +351,7 @@ thinking_prompt: "Use a think first process to answer the user: Make sure you continue thinking until you find a satisfactory answer Assess any potential errors you may make - + After thinking you can answer the user." diff --git a/lollms/configs/config.yaml b/lollms/configs/config.yaml index af3a317..09ccbaf 100644 --- a/lollms/configs/config.yaml +++ b/lollms/configs/config.yaml @@ -339,7 +339,7 @@ current_language: english fun_mode: false think_first_mode: false thinking_prompt: "Use a think first process to answer the user: - + Ask yourself about the user's request and answer it with logical details. If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly. If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer. @@ -351,7 +351,7 @@ thinking_prompt: "Use a think first process to answer the user: Make sure you continue thinking until you find a satisfactory answer Assess any potential errors you may make - + After thinking you can answer the user." diff --git a/lollms/server/events/lollms_generation_events.py b/lollms/server/events/lollms_generation_events.py index 57c2aae..4dc6717 100644 --- a/lollms/server/events/lollms_generation_events.py +++ b/lollms/server/events/lollms_generation_events.py @@ -212,7 +212,6 @@ def add_events(sio:socketio): client.generation_thread = threading.Thread(target=do_generation) client.generation_thread.start() - ASCIIColors.info("Started generation task") lollmsElfServer.busy=True except Exception as ex: @@ -269,7 +268,6 @@ def add_events(sio:socketio): ASCIIColors.green("Starting message generation by "+lollmsElfServer.personality.name) client.generation_thread = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message.id, client_id)) client.generation_thread.start() - ASCIIColors.info("Started generation task") lollmsElfServer.busy=True #tpe = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message_id, client_id)) #tpe.start() diff --git a/lollms/server/events/lollms_personality_events.py b/lollms/server/events/lollms_personality_events.py index 801bd46..37289b0 100644 --- a/lollms/server/events/lollms_personality_events.py +++ b/lollms/server/events/lollms_personality_events.py @@ -183,4 +183,3 @@ def add_events(sio:socketio): client.generation_thread = threading.Thread(target=do_generation) client.generation_thread.start() - ASCIIColors.info("Started generation task") diff --git a/setup.py b/setup.py index aced178..16cd20c 100644 --- a/setup.py +++ b/setup.py @@ -26,7 +26,7 @@ def get_all_files(path): setuptools.setup( name="lollms", - version="10.1.0", + version="11.0.0", author="Saifeddine ALOUI (ParisNeo)", author_email="parisneo_ai@gmail.com", description="A python library for AI personality definition",