From 132f5c0cf52fe0ef5e73a614471a5def70564b8e Mon Sep 17 00:00:00 2001 From: Saifeddine ALOUI Date: Sun, 14 Apr 2024 23:33:19 +0200 Subject: [PATCH] upgraded configuration --- configs/config.yaml | 6 +- .../personal/configs/lollms_elf_config.yaml | 21 +++- .../personal/configs/lollms_elf_config.yaml | 21 +++- .../configs/lollms_elf_local_config.yaml | 100 ++++++++++++++++-- lollms/configs/config.yaml | 6 +- lollms/server/configs/config.yaml | 21 +++- .../configs/lollms_discord_local_config.yaml | 21 +++- 7 files changed, 168 insertions(+), 28 deletions(-) diff --git a/configs/config.yaml b/configs/config.yaml index c3948b8..cb6278d 100644 --- a/configs/config.yaml +++ b/configs/config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 79 +version: 80 binding_name: null model_name: null model_variant: null @@ -143,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs diff --git a/elf_docker_cfg/personal/configs/lollms_elf_config.yaml b/elf_docker_cfg/personal/configs/lollms_elf_config.yaml index f07ab2c..cb6278d 100644 --- a/elf_docker_cfg/personal/configs/lollms_elf_config.yaml +++ b/elf_docker_cfg/personal/configs/lollms_elf_config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 73 +version: 80 binding_name: null model_name: null model_variant: null @@ -8,9 +8,10 @@ model_type: null show_news_panel: True # Security measures +turn_on_setting_update_validation: True turn_on_code_execution: True turn_on_code_validation: True -turn_on_open_file_validation: True +turn_on_open_file_validation: False turn_on_send_file_validation: False force_accept_remote_access: false @@ -26,14 +27,15 @@ port: 9600 # Genreration parameters discussion_prompt_separator: "!@>" seed: -1 -n_predict: 1024 ctx_size: 4084 +max_n_predict: 4096 min_n_predict: 512 temperature: 0.9 top_k: 50 top_p: 0.95 repeat_last_n: 40 repeat_penalty: 1.2 +num_experts_per_token: 2 n_threads: 8 @@ -47,6 +49,7 @@ extensions: [] user_name: user user_description: "" use_user_name_in_discussions: false +use_model_name_in_discussions: false user_avatar: default_user.svg use_user_informations_in_discussion: false @@ -85,6 +88,14 @@ current_language: en enable_sd_service: false sd_base_url: http://localhost:7860 +# Image generation service comfyui +enable_comfyui_service: false +comfyui_base_url: http://127.0.0.1:8188/ + +# Motion control service +enable_motion_ctrl_service: false +motion_ctrl_base_url: http://localhost:7861 + # ollama service enable_ollama_service: false ollama_base_url: http://localhost:11434 @@ -132,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs diff --git a/elf_test_cfg/personal/configs/lollms_elf_config.yaml b/elf_test_cfg/personal/configs/lollms_elf_config.yaml index f07ab2c..cb6278d 100644 --- a/elf_test_cfg/personal/configs/lollms_elf_config.yaml +++ b/elf_test_cfg/personal/configs/lollms_elf_config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 73 +version: 80 binding_name: null model_name: null model_variant: null @@ -8,9 +8,10 @@ model_type: null show_news_panel: True # Security measures +turn_on_setting_update_validation: True turn_on_code_execution: True turn_on_code_validation: True -turn_on_open_file_validation: True +turn_on_open_file_validation: False turn_on_send_file_validation: False force_accept_remote_access: false @@ -26,14 +27,15 @@ port: 9600 # Genreration parameters discussion_prompt_separator: "!@>" seed: -1 -n_predict: 1024 ctx_size: 4084 +max_n_predict: 4096 min_n_predict: 512 temperature: 0.9 top_k: 50 top_p: 0.95 repeat_last_n: 40 repeat_penalty: 1.2 +num_experts_per_token: 2 n_threads: 8 @@ -47,6 +49,7 @@ extensions: [] user_name: user user_description: "" use_user_name_in_discussions: false +use_model_name_in_discussions: false user_avatar: default_user.svg use_user_informations_in_discussion: false @@ -85,6 +88,14 @@ current_language: en enable_sd_service: false sd_base_url: http://localhost:7860 +# Image generation service comfyui +enable_comfyui_service: false +comfyui_base_url: http://127.0.0.1:8188/ + +# Motion control service +enable_motion_ctrl_service: false +motion_ctrl_base_url: http://localhost:7861 + # ollama service enable_ollama_service: false ollama_base_url: http://localhost:11434 @@ -132,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs diff --git a/elf_test_cfg/personal/configs/lollms_elf_local_config.yaml b/elf_test_cfg/personal/configs/lollms_elf_local_config.yaml index 1d0b041..cb6278d 100644 --- a/elf_test_cfg/personal/configs/lollms_elf_local_config.yaml +++ b/elf_test_cfg/personal/configs/lollms_elf_local_config.yaml @@ -1,9 +1,24 @@ -# =================== Lord Of Large Language Models Configuration file =========================== -version: 40 +# =================== Lord Of Large Language Multimodal Systems Configuration file =========================== +version: 80 binding_name: null model_name: null +model_variant: null +model_type: null +show_news_panel: True +# Security measures +turn_on_setting_update_validation: True +turn_on_code_execution: True +turn_on_code_validation: True +turn_on_open_file_validation: False +turn_on_send_file_validation: False + +force_accept_remote_access: false + +# Server information +headless_server_mode: False +allowed_origins: [] # Host information host: localhost @@ -12,14 +27,15 @@ port: 9600 # Genreration parameters discussion_prompt_separator: "!@>" seed: -1 -n_predict: 1024 ctx_size: 4084 +max_n_predict: 4096 min_n_predict: 512 temperature: 0.9 top_k: 50 top_p: 0.95 repeat_last_n: 40 repeat_penalty: 1.2 +num_experts_per_token: 2 n_threads: 8 @@ -33,15 +49,24 @@ extensions: [] user_name: user user_description: "" use_user_name_in_discussions: false +use_model_name_in_discussions: false user_avatar: default_user.svg use_user_informations_in_discussion: false # UI parameters -discussion_db_name: database.db +discussion_db_name: default # Automatic updates debug: False +debug_log_file_path: "" auto_update: true +auto_sync_personalities: true +auto_sync_extensions: true +auto_sync_bindings: true +auto_sync_models: true + + + auto_save: true auto_title: false # Install mode (cpu, cpu-noavx, nvidia-tensorcores, nvidia, amd-noavx, amd, apple-intel, apple-silicon) @@ -49,16 +74,54 @@ hardware_mode: nvidia-tensorcores # Automatically open the browser auto_show_browser: true +# copy to clipboard +copy_to_clipboard_add_all_details: false + # Voice service enable_voice_service: false -xtts_base_url: http://127.0.0.1:8020 +xtts_base_url: http://localhost:8020 auto_read: false current_voice: null current_language: en # Image generation service enable_sd_service: false -sd_base_url: http://127.0.0.1:7860 +sd_base_url: http://localhost:7860 + +# Image generation service comfyui +enable_comfyui_service: false +comfyui_base_url: http://127.0.0.1:8188/ + +# Motion control service +enable_motion_ctrl_service: false +motion_ctrl_base_url: http://localhost:7861 + +# ollama service +enable_ollama_service: false +ollama_base_url: http://localhost:11434 + +# petals service +enable_petals_service: false +petals_base_url: http://localhost:8064 +petals_model_path: TinyLlama/TinyLlama-1.1B-Chat-v1.0 +petals_device: cuda + +# lollms service +enable_lollms_service: false +lollms_base_url: http://localhost:1234 + +# elastic search service +elastic_search_service: false +elastic_search_url: http://localhost:9200 + +# vll service +enable_vllm_service: false +vllm_url: http://localhost:8000 +vllm_model_path: TinyLlama/TinyLlama-1.1B-Chat-v1.0 +vllm_gpu_memory_utilization: 0.9 +vllm_max_model_len: 4096 +vllm_max_num_seqs: 256 + # Audio media_on: false @@ -71,22 +134,33 @@ audio_silenceTimer: 5000 # Data vectorization activate_skills_lib: false # Activate vectorizing previous conversations +skills_lib_database_name: "default" # Default skills database summerize_discussion: false # activate discussion summary (better but adds computation time) + max_summary_size: 512 # in tokens data_vectorization_visualize_on_vectorization: false use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs - +# Activate internet search +activate_internet_search: false +internet_vectorization_chunk_size: 512 # chunk size +internet_vectorization_overlap_size: 128 # overlap between chunks size +internet_vectorization_nb_chunks: 2 # number of chunks to use +internet_nb_search_pages: 3 # number of pages to select +internet_quick_search: False # If active the search engine will not load and read the webpages +internet_activate_search_decision: False # If active the ai decides by itself if it needs to do search # Helpers pdf_latex_path: null @@ -94,3 +168,13 @@ pdf_latex_path: null positive_boost: null negative_boost: null force_output_language_to_be: null +fun_mode: False + + +# webui configurations +show_code_of_conduct: true +activate_audio_infos: true + + +# whisper configuration +whisper_model: base \ No newline at end of file diff --git a/lollms/configs/config.yaml b/lollms/configs/config.yaml index c3948b8..cb6278d 100644 --- a/lollms/configs/config.yaml +++ b/lollms/configs/config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 79 +version: 80 binding_name: null model_name: null model_variant: null @@ -143,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs diff --git a/lollms/server/configs/config.yaml b/lollms/server/configs/config.yaml index f07ab2c..cb6278d 100644 --- a/lollms/server/configs/config.yaml +++ b/lollms/server/configs/config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 73 +version: 80 binding_name: null model_name: null model_variant: null @@ -8,9 +8,10 @@ model_type: null show_news_panel: True # Security measures +turn_on_setting_update_validation: True turn_on_code_execution: True turn_on_code_validation: True -turn_on_open_file_validation: True +turn_on_open_file_validation: False turn_on_send_file_validation: False force_accept_remote_access: false @@ -26,14 +27,15 @@ port: 9600 # Genreration parameters discussion_prompt_separator: "!@>" seed: -1 -n_predict: 1024 ctx_size: 4084 +max_n_predict: 4096 min_n_predict: 512 temperature: 0.9 top_k: 50 top_p: 0.95 repeat_last_n: 40 repeat_penalty: 1.2 +num_experts_per_token: 2 n_threads: 8 @@ -47,6 +49,7 @@ extensions: [] user_name: user user_description: "" use_user_name_in_discussions: false +use_model_name_in_discussions: false user_avatar: default_user.svg use_user_informations_in_discussion: false @@ -85,6 +88,14 @@ current_language: en enable_sd_service: false sd_base_url: http://localhost:7860 +# Image generation service comfyui +enable_comfyui_service: false +comfyui_base_url: http://127.0.0.1:8188/ + +# Motion control service +enable_motion_ctrl_service: false +motion_ctrl_base_url: http://localhost:7861 + # ollama service enable_ollama_service: false ollama_base_url: http://localhost:11434 @@ -132,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs diff --git a/personal_data/configs/lollms_discord_local_config.yaml b/personal_data/configs/lollms_discord_local_config.yaml index 1d183a9..cb6278d 100644 --- a/personal_data/configs/lollms_discord_local_config.yaml +++ b/personal_data/configs/lollms_discord_local_config.yaml @@ -1,5 +1,5 @@ # =================== Lord Of Large Language Multimodal Systems Configuration file =========================== -version: 73 +version: 80 binding_name: null model_name: null model_variant: null @@ -8,9 +8,10 @@ model_type: null show_news_panel: True # Security measures +turn_on_setting_update_validation: True turn_on_code_execution: True turn_on_code_validation: True -turn_on_open_file_validation: True +turn_on_open_file_validation: False turn_on_send_file_validation: False force_accept_remote_access: false @@ -26,14 +27,15 @@ port: 9600 # Genreration parameters discussion_prompt_separator: "!@>" seed: -1 -n_predict: 1024 ctx_size: 4084 +max_n_predict: 4096 min_n_predict: 512 temperature: 0.9 top_k: 50 top_p: 0.95 repeat_last_n: 40 repeat_penalty: 1.2 +num_experts_per_token: 2 n_threads: 8 @@ -47,6 +49,7 @@ extensions: [] user_name: user user_description: "" use_user_name_in_discussions: false +use_model_name_in_discussions: false user_avatar: default_user.svg use_user_informations_in_discussion: false @@ -85,6 +88,14 @@ current_language: en enable_sd_service: false sd_base_url: http://localhost:7860 +# Image generation service comfyui +enable_comfyui_service: false +comfyui_base_url: http://127.0.0.1:8188/ + +# Motion control service +enable_motion_ctrl_service: false +motion_ctrl_base_url: http://localhost:7861 + # ollama service enable_ollama_service: false ollama_base_url: http://localhost:11434 @@ -132,11 +143,13 @@ use_files: true # Activate using files data_vectorization_activate: true # To activate/deactivate data vectorization data_vectorization_method: "tfidf_vectorizer" #"model_embedding" or "tfidf_vectorizer" data_visualization_method: "PCA" #"PCA" or "TSNE" +data_vectorization_sentense_transformer_model: "all-MiniLM-L6-v2" # you can use another model by setting its name here or its path + data_vectorization_save_db: False # For each new session, new files data_vectorization_chunk_size: 512 # chunk size data_vectorization_overlap_size: 128 # overlap between chunks size data_vectorization_nb_chunks: 2 # number of chunks to use -data_vectorization_build_keys_words: false # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. +data_vectorization_build_keys_words: true # If true, when querrying the database, we use keywords generated from the user prompt instead of the prompt itself. data_vectorization_force_first_chunk: false # If true, the first chunk of the document will systematically be used data_vectorization_make_persistance: false # If true, the data will be persistant webween runs