This commit is contained in:
Saifeddine ALOUI 2025-02-15 16:11:46 +01:00
parent 9efab84767
commit cb874b137e
5 changed files with 5 additions and 8 deletions

View File

@ -339,7 +339,7 @@ current_language: english
fun_mode: false fun_mode: false
think_first_mode: false think_first_mode: false
thinking_prompt: "Use a think first process to answer the user: thinking_prompt: "Use a think first process to answer the user:
<thinking> <think>
Ask yourself about the user's request and answer it with logical details. Ask yourself about the user's request and answer it with logical details.
If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly. If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly.
If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer. If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer.
@ -351,7 +351,7 @@ thinking_prompt: "Use a think first process to answer the user:
Make sure you continue thinking until you find a satisfactory answer Make sure you continue thinking until you find a satisfactory answer
Assess any potential errors you may make Assess any potential errors you may make
</thinking> </think>
After thinking you can answer the user." After thinking you can answer the user."

View File

@ -339,7 +339,7 @@ current_language: english
fun_mode: false fun_mode: false
think_first_mode: false think_first_mode: false
thinking_prompt: "Use a think first process to answer the user: thinking_prompt: "Use a think first process to answer the user:
<thinking> <think>
Ask yourself about the user's request and answer it with logical details. Ask yourself about the user's request and answer it with logical details.
If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly. If the user is requesting general information that does not require internet search and you are confident about it, then prepare to answer directly.
If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer. If the user is requesting general information that does require internet search and you have in the context enough information to answer, then use that data to answer.
@ -351,7 +351,7 @@ thinking_prompt: "Use a think first process to answer the user:
Make sure you continue thinking until you find a satisfactory answer Make sure you continue thinking until you find a satisfactory answer
Assess any potential errors you may make Assess any potential errors you may make
</thinking> </think>
After thinking you can answer the user." After thinking you can answer the user."

View File

@ -212,7 +212,6 @@ def add_events(sio:socketio):
client.generation_thread = threading.Thread(target=do_generation) client.generation_thread = threading.Thread(target=do_generation)
client.generation_thread.start() client.generation_thread.start()
ASCIIColors.info("Started generation task")
lollmsElfServer.busy=True lollmsElfServer.busy=True
except Exception as ex: except Exception as ex:
@ -269,7 +268,6 @@ def add_events(sio:socketio):
ASCIIColors.green("Starting message generation by "+lollmsElfServer.personality.name) ASCIIColors.green("Starting message generation by "+lollmsElfServer.personality.name)
client.generation_thread = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message.id, client_id)) client.generation_thread = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message.id, client_id))
client.generation_thread.start() client.generation_thread.start()
ASCIIColors.info("Started generation task")
lollmsElfServer.busy=True lollmsElfServer.busy=True
#tpe = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message_id, client_id)) #tpe = threading.Thread(target=lollmsElfServer.start_message_generation, args=(message, message_id, client_id))
#tpe.start() #tpe.start()

View File

@ -183,4 +183,3 @@ def add_events(sio:socketio):
client.generation_thread = threading.Thread(target=do_generation) client.generation_thread = threading.Thread(target=do_generation)
client.generation_thread.start() client.generation_thread.start()
ASCIIColors.info("Started generation task")

View File

@ -26,7 +26,7 @@ def get_all_files(path):
setuptools.setup( setuptools.setup(
name="lollms", name="lollms",
version="10.1.0", version="11.0.0",
author="Saifeddine ALOUI (ParisNeo)", author="Saifeddine ALOUI (ParisNeo)",
author_email="parisneo_ai@gmail.com", author_email="parisneo_ai@gmail.com",
description="A python library for AI personality definition", description="A python library for AI personality definition",