mirror of
https://github.com/ParisNeo/lollms-webui.git
synced 2025-06-12 20:38:10 +00:00
enhanced ui
This commit is contained in:
@ -5,7 +5,7 @@
|
||||
<link rel="icon" type="image/svg+xml" href="/vite.svg" />
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||
<title>Vite + Vue</title>
|
||||
<script type="module" crossorigin src="/assets/index-gD_iVx44.js"></script>
|
||||
<script type="module" crossorigin src="/assets/index-xU-eXePa.js"></script>
|
||||
<link rel="stylesheet" crossorigin href="/assets/index-tsveoIuX.css">
|
||||
</head>
|
||||
<body>
|
||||
|
@ -1,7 +1,8 @@
|
||||
import { createApp } from 'vue';
|
||||
import { createRouter, createWebHistory } from 'vue-router';
|
||||
import { createStore } from 'vuex'
|
||||
import App from './App.vue';
|
||||
import Install from './components/Install.vue';
|
||||
import Install from './views/Install.vue';
|
||||
import './style.css';
|
||||
const routes = [
|
||||
{
|
||||
@ -16,4 +17,51 @@ const router = createRouter({
|
||||
routes,
|
||||
});
|
||||
|
||||
|
||||
// Create a new store instance.
|
||||
export const store = createStore({
|
||||
state () {
|
||||
return {
|
||||
// count: 0,
|
||||
yesNoDialog:null,
|
||||
universalForm:null,
|
||||
toast:null,
|
||||
messageBox:null,
|
||||
api_get_req:null,
|
||||
startSpeechRecognition:null,
|
||||
ready:false,
|
||||
loading_infos: "",
|
||||
loading_progress: 0,
|
||||
version : "unknown",
|
||||
settingsChanged:false,
|
||||
isConnected: false, // Add the isConnected property
|
||||
isModelOk: false,
|
||||
isGenerating: false,
|
||||
config:null,
|
||||
mountedPers:null,
|
||||
mountedPersArr:[],
|
||||
mountedExtensions:[],
|
||||
bindingsZoo:[],
|
||||
modelsArr:[],
|
||||
selectedModel:null,
|
||||
personalities:[],
|
||||
diskUsage:null,
|
||||
ramUsage:null,
|
||||
vramUsage:null,
|
||||
modelsZoo:[],
|
||||
installedModels:[],
|
||||
currentModel:null,
|
||||
extensionsZoo:[],
|
||||
databases:[],
|
||||
}
|
||||
},
|
||||
mutations: {
|
||||
|
||||
},
|
||||
getters: {
|
||||
},
|
||||
actions: {
|
||||
}
|
||||
})
|
||||
|
||||
createApp(App).use(router).mount('#app');
|
||||
|
@ -1,5 +1,6 @@
|
||||
<template>
|
||||
<div class="flex flex-col items-center justify-center rounded-lg m-2 shadow-lg hover:border-primary dark:hover:border-primary hover:border-solid hover:border-2 border-2 border-transparent even:bg-bg-light-discussion-odd dark:even:bg-bg-dark-discussion-odd flex flex-col flex-grow flex-wrap overflow-visible p-4 pb-2">
|
||||
<img :src = "LoLLMSLogo">
|
||||
<h1 class="text-4xl font-bold mb-4">LOLLMS installation tool</h1>
|
||||
<p class="text-left">
|
||||
Welcome to the installer of lollms. Here you can select your install profile.<br>
|
||||
@ -7,11 +8,11 @@
|
||||
</p>
|
||||
<div class="flex flex-col gap-2">
|
||||
<label class="flex items-center">
|
||||
<input type="radio" value="nvidia" v-model="selectedOption" class="mr-2">
|
||||
<input type="radio" value="cuda" v-model="selectedOption" class="mr-2">
|
||||
Use NVIDIA GPU with CUDA
|
||||
</label>
|
||||
<label class="flex items-center">
|
||||
<input type="radio" value="amd" v-model="selectedOption" class="mr-2">
|
||||
<input type="radio" value="rocm" v-model="selectedOption" class="mr-2">
|
||||
Use AMD GPU with ROCm
|
||||
</label>
|
||||
<label class="flex items-center">
|
||||
@ -27,16 +28,23 @@
|
||||
Install
|
||||
</button>
|
||||
</div>
|
||||
<MessageBox ></MessageBox>
|
||||
</template>
|
||||
|
||||
<script>
|
||||
import MessageBox from '../components/messageBox.vue';
|
||||
import axios from 'axios';
|
||||
import LoLLMSLogo from '../assets/logo.png'
|
||||
export default {
|
||||
data() {
|
||||
return {
|
||||
LoLLMSLogo:LoLLMSLogo,
|
||||
selectedOption: "cpu",
|
||||
};
|
||||
},
|
||||
components:{
|
||||
MessageBox
|
||||
},
|
||||
methods: {
|
||||
install() {
|
||||
axios.post("/start_installing",{mode:this.selectedOption}).then(()=>{
|
@ -1,10 +1,18 @@
|
||||
"""
|
||||
Project: lollms_installer
|
||||
Author: Your Name
|
||||
Description: This tool is designed to install and configure the LoLLMS system on your machine. LoLLMS is a multi-bindings, multi-personalities LLM full-stack system for AI applications in robotics. It provides a user-friendly interface for setting up and managing the system.
|
||||
"""
|
||||
from fastapi import FastAPI
|
||||
from fastapi.staticfiles import StaticFiles
|
||||
from pydantic import BaseModel
|
||||
import uvicorn
|
||||
from lollms.paths import LollmsPaths
|
||||
from lollms.main_config import LOLLMSConfig
|
||||
from lollms.utilities import check_and_install_torch, PackageManager
|
||||
from lollms.utilities import check_and_install_torch, PackageManager, check_torch_version, reinstall_pytorch_with_cuda, reinstall_pytorch_with_cpu, reinstall_pytorch_with_rocm
|
||||
from lollms.com import NotificationType, NotificationDisplayType, LoLLMsCom
|
||||
from lollms.types import MSG_TYPE, SENDER_TYPES
|
||||
from lollms.app import LollmsApplication
|
||||
from pathlib import Path
|
||||
from ascii_colors import ASCIIColors
|
||||
import subprocess
|
||||
@ -13,6 +21,7 @@ from pathlib import Path
|
||||
from starlette.responses import FileResponse
|
||||
from starlette.requests import Request
|
||||
import webbrowser
|
||||
import socketio
|
||||
|
||||
root_path = Path(__file__).parent.parent.parent.parent
|
||||
global_path = root_path/"global_paths_cfg.yaml"
|
||||
@ -34,8 +43,12 @@ ASCIIColors.red("|_____|_____|_____|_____|_|_|_|_____|")
|
||||
ASCIIColors.red(" Configurator ")
|
||||
ASCIIColors.red(" LoLLMS configuratoin tool")
|
||||
ASCIIColors.yellow(f"Root dir : {root_path}")
|
||||
app = FastAPI(debug=True)
|
||||
|
||||
sio = socketio.AsyncServer(async_mode='asgi')
|
||||
app = FastAPI(debug=True)
|
||||
app.mount("/socket.io", socketio.ASGIApp(sio))
|
||||
|
||||
lollms_app = LollmsApplication("lollms_installer",config=config,lollms_paths=lollms_paths, load_binding=False, load_model=False, socketio=sio)
|
||||
|
||||
# Serve the index.html file for all routes
|
||||
@app.get("/{full_path:path}")
|
||||
@ -50,21 +63,60 @@ async def serve_index(request: Request, full_path: Path):
|
||||
|
||||
# app.mount("/", StaticFiles(directory=root_path/"scripts/python/lollms_installer/frontend/dist"), name="static")
|
||||
|
||||
|
||||
class InstallProperties(BaseModel):
|
||||
mode: str
|
||||
|
||||
@app.post("/start_installing")
|
||||
def start_installing(data: InstallProperties):
|
||||
"""
|
||||
Handle the start_installing endpoint.
|
||||
|
||||
Parameters:
|
||||
- **data**: An instance of the `InstallProperties` model containing the installation mode.
|
||||
|
||||
Returns:
|
||||
- A dictionary with a "message" key indicating the success of the installation.
|
||||
"""
|
||||
if data.mode=="cpu":
|
||||
config.enable_gpu=False
|
||||
config.save_config()
|
||||
else:
|
||||
config.enable_gpu=False
|
||||
try:
|
||||
lollms_app.ShowBlockingMessage("Installing pytorch for CPU")
|
||||
reinstall_pytorch_with_cpu()
|
||||
config.save_config()
|
||||
lollms_app.HideBlockingMessage()
|
||||
except:
|
||||
lollms_app.HideBlockingMessage()
|
||||
|
||||
elif data.mode=="cuda":
|
||||
config.enable_gpu=True
|
||||
config.save_config()
|
||||
try:
|
||||
lollms_app.ShowBlockingMessage("Installing pytorch for nVidia GPU (cuda)")
|
||||
reinstall_pytorch_with_cuda()
|
||||
config.save_config()
|
||||
lollms_app.HideBlockingMessage()
|
||||
except:
|
||||
lollms_app.HideBlockingMessage()
|
||||
elif data.mode=="rocm":
|
||||
config.enable_gpu=True
|
||||
try:
|
||||
lollms_app.ShowBlockingMessage("Installing pytorch for AMD GPU (rocm)")
|
||||
reinstall_pytorch_with_rocm()
|
||||
config.save_config()
|
||||
lollms_app.HideBlockingMessage()
|
||||
except:
|
||||
lollms_app.HideBlockingMessage()
|
||||
elif data.mode=="metal":
|
||||
try:
|
||||
lollms_app.ShowBlockingMessage("Installing pytorch for Apple Silicon (Metal)")
|
||||
config.enable_gpu=False
|
||||
reinstall_pytorch_with_cpu()
|
||||
config.save_config()
|
||||
lollms_app.HideBlockingMessage()
|
||||
except:
|
||||
lollms_app.HideBlockingMessage()
|
||||
# Your code here
|
||||
return {"message": "Item created successfully"}
|
||||
|
||||
if __name__ == "__main__":
|
||||
webbrowser.open(f"http://localhost:8000")
|
||||
uvicorn.run(app, host="0.0.0.0", port=8000)
|
||||
uvicorn.run(app, host="localhost", port=8000)
|
Reference in New Issue
Block a user