mirror of
https://github.com/mudler/LocalAI.git
synced 2024-12-24 14:56:41 +00:00
143 lines
3.0 KiB
Go
143 lines
3.0 KiB
Go
|
package main
|
||
|
|
||
|
// A simple program demonstrating the text area component from the Bubbles
|
||
|
// component library.
|
||
|
|
||
|
import (
|
||
|
"fmt"
|
||
|
"strings"
|
||
|
|
||
|
"github.com/charmbracelet/bubbles/textarea"
|
||
|
"github.com/charmbracelet/bubbles/viewport"
|
||
|
tea "github.com/charmbracelet/bubbletea"
|
||
|
"github.com/charmbracelet/lipgloss"
|
||
|
llama "github.com/go-skynet/llama/go"
|
||
|
)
|
||
|
|
||
|
func startInteractive(l *llama.LLama, opts ...llama.PredictOption) error {
|
||
|
p := tea.NewProgram(initialModel(l, opts...))
|
||
|
|
||
|
_, err := p.Run()
|
||
|
return err
|
||
|
}
|
||
|
|
||
|
type (
|
||
|
errMsg error
|
||
|
)
|
||
|
|
||
|
type model struct {
|
||
|
viewport viewport.Model
|
||
|
messages *[]string
|
||
|
textarea textarea.Model
|
||
|
senderStyle lipgloss.Style
|
||
|
err error
|
||
|
l *llama.LLama
|
||
|
opts []llama.PredictOption
|
||
|
|
||
|
predictC chan string
|
||
|
}
|
||
|
|
||
|
func initialModel(l *llama.LLama, opts ...llama.PredictOption) model {
|
||
|
ta := textarea.New()
|
||
|
ta.Placeholder = "Send a message..."
|
||
|
ta.Focus()
|
||
|
|
||
|
ta.Prompt = "┃ "
|
||
|
ta.CharLimit = 280
|
||
|
|
||
|
ta.SetWidth(200)
|
||
|
ta.SetHeight(3)
|
||
|
|
||
|
// Remove cursor line styling
|
||
|
ta.FocusedStyle.CursorLine = lipgloss.NewStyle()
|
||
|
|
||
|
ta.ShowLineNumbers = false
|
||
|
|
||
|
vp := viewport.New(200, 5)
|
||
|
vp.SetContent(`Welcome to llama-cli. Type a message and press Enter to send. Alpaca doesn't keep context of the whole chat (yet).`)
|
||
|
|
||
|
ta.KeyMap.InsertNewline.SetEnabled(false)
|
||
|
|
||
|
predictChannel := make(chan string)
|
||
|
messages := []string{}
|
||
|
m := model{
|
||
|
textarea: ta,
|
||
|
messages: &messages,
|
||
|
viewport: vp,
|
||
|
senderStyle: lipgloss.NewStyle().Foreground(lipgloss.Color("5")),
|
||
|
err: nil,
|
||
|
l: l,
|
||
|
opts: opts,
|
||
|
predictC: predictChannel,
|
||
|
}
|
||
|
go func() {
|
||
|
for p := range predictChannel {
|
||
|
str, _ := templateString(emptyInput, struct {
|
||
|
Instruction string
|
||
|
Input string
|
||
|
}{Instruction: p})
|
||
|
res, _ := l.Predict(
|
||
|
str,
|
||
|
opts...,
|
||
|
)
|
||
|
|
||
|
mm := *m.messages
|
||
|
*m.messages = mm[:len(mm)-1]
|
||
|
*m.messages = append(*m.messages, m.senderStyle.Render("llama: ")+res)
|
||
|
m.viewport.SetContent(strings.Join(*m.messages, "\n"))
|
||
|
ta.Reset()
|
||
|
m.viewport.GotoBottom()
|
||
|
}
|
||
|
}()
|
||
|
|
||
|
return m
|
||
|
}
|
||
|
|
||
|
func (m model) Init() tea.Cmd {
|
||
|
return textarea.Blink
|
||
|
}
|
||
|
|
||
|
func (m model) Update(msg tea.Msg) (tea.Model, tea.Cmd) {
|
||
|
var (
|
||
|
tiCmd tea.Cmd
|
||
|
vpCmd tea.Cmd
|
||
|
)
|
||
|
|
||
|
m.textarea, tiCmd = m.textarea.Update(msg)
|
||
|
m.viewport, vpCmd = m.viewport.Update(msg)
|
||
|
|
||
|
switch msg := msg.(type) {
|
||
|
case tea.WindowSizeMsg:
|
||
|
|
||
|
// m.viewport.Width = msg.Width
|
||
|
// m.viewport.Height = msg.Height
|
||
|
case tea.KeyMsg:
|
||
|
switch msg.Type {
|
||
|
case tea.KeyCtrlC, tea.KeyEsc:
|
||
|
fmt.Println(m.textarea.Value())
|
||
|
return m, tea.Quit
|
||
|
case tea.KeyEnter:
|
||
|
*m.messages = append(*m.messages, m.senderStyle.Render("You: ")+m.textarea.Value(), m.senderStyle.Render("Loading response..."))
|
||
|
m.predictC <- m.textarea.Value()
|
||
|
m.viewport.SetContent(strings.Join(*m.messages, "\n"))
|
||
|
m.textarea.Reset()
|
||
|
m.viewport.GotoBottom()
|
||
|
}
|
||
|
|
||
|
// We handle errors just like any other message
|
||
|
case errMsg:
|
||
|
m.err = msg
|
||
|
return m, nil
|
||
|
}
|
||
|
|
||
|
return m, tea.Batch(tiCmd, vpCmd)
|
||
|
}
|
||
|
|
||
|
func (m model) View() string {
|
||
|
return fmt.Sprintf(
|
||
|
"%s\n\n%s",
|
||
|
m.viewport.View(),
|
||
|
m.textarea.View(),
|
||
|
) + "\n\n"
|
||
|
}
|