Compare commits
7 Commits
e9fde37201
...
657416780d
Author | SHA1 | Date | |
---|---|---|---|
657416780d | |||
c143d863cb | |||
3aff5514e4 | |||
5acdbb5675 | |||
c53e952acc | |||
3d8d3b61b3 | |||
4fb059c850 |
@ -94,7 +94,11 @@ func buildRequest(params model.RequestParameters, messages []model.Message) Requ
|
||||
if err != nil {
|
||||
panic("Could not serialize []ToolCall to XMLFunctionCall")
|
||||
}
|
||||
message.Content += xmlString
|
||||
if len(message.Content) > 0 {
|
||||
message.Content += fmt.Sprintf("\n\n%s", xmlString)
|
||||
} else {
|
||||
message.Content = xmlString
|
||||
}
|
||||
case model.MessageRoleToolResult:
|
||||
xmlFuncResults := convertToolResultsToXMLFunctionResult(msg.ToolResults)
|
||||
xmlString, err := xmlFuncResults.XMLString()
|
||||
@ -197,6 +201,8 @@ func (c *AnthropicClient) CreateChatCompletionStream(
|
||||
scanner := bufio.NewScanner(resp.Body)
|
||||
sb := strings.Builder{}
|
||||
|
||||
isToolCall := false
|
||||
|
||||
for scanner.Scan() {
|
||||
line := scanner.Text()
|
||||
line = strings.TrimSpace(line)
|
||||
@ -271,6 +277,8 @@ func (c *AnthropicClient) CreateChatCompletionStream(
|
||||
return content, fmt.Errorf("reached </function_calls> stop sequence but no opening tag found")
|
||||
}
|
||||
|
||||
isToolCall = true
|
||||
|
||||
funcCallXml := content[start:]
|
||||
funcCallXml += FUNCTION_STOP_SEQUENCE
|
||||
|
||||
@ -316,10 +324,12 @@ func (c *AnthropicClient) CreateChatCompletionStream(
|
||||
case "message_stop":
|
||||
// return the completed message
|
||||
if callback != nil {
|
||||
callback(model.Message{
|
||||
Role: model.MessageRoleAssistant,
|
||||
Content: sb.String(),
|
||||
})
|
||||
if !isToolCall {
|
||||
callback(model.Message{
|
||||
Role: model.MessageRoleAssistant,
|
||||
Content: sb.String(),
|
||||
})
|
||||
}
|
||||
}
|
||||
return sb.String(), nil
|
||||
case "error":
|
||||
|
@ -204,7 +204,7 @@ func (c *OpenAIClient) CreateChatCompletionStream(
|
||||
ctx context.Context,
|
||||
params model.RequestParameters,
|
||||
messages []model.Message,
|
||||
callbback provider.ReplyCallback,
|
||||
callback provider.ReplyCallback,
|
||||
output chan<- string,
|
||||
) (string, error) {
|
||||
client := openai.NewClient(c.APIKey)
|
||||
@ -256,22 +256,22 @@ func (c *OpenAIClient) CreateChatCompletionStream(
|
||||
return content.String(), err
|
||||
}
|
||||
|
||||
if callbback != nil {
|
||||
if callback != nil {
|
||||
for _, result := range results {
|
||||
callbback(result)
|
||||
callback(result)
|
||||
}
|
||||
}
|
||||
|
||||
// Recurse into CreateChatCompletionStream with the tool call replies
|
||||
messages = append(messages, results...)
|
||||
return c.CreateChatCompletionStream(ctx, params, messages, callbback, output)
|
||||
}
|
||||
|
||||
if callbback != nil {
|
||||
callbback(model.Message{
|
||||
Role: model.MessageRoleAssistant,
|
||||
Content: content.String(),
|
||||
})
|
||||
return c.CreateChatCompletionStream(ctx, params, messages, callback, output)
|
||||
} else {
|
||||
if callback != nil {
|
||||
callback(model.Message{
|
||||
Role: model.MessageRoleAssistant,
|
||||
Content: content.String(),
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
return content.String(), err
|
||||
|
263
pkg/tui/tui.go
263
pkg/tui/tui.go
@ -2,8 +2,6 @@ package tui
|
||||
|
||||
// The terminal UI for lmcli, launched from the `lmcli chat` command
|
||||
// TODO:
|
||||
// - binding to open selected message/input in $EDITOR
|
||||
// - ability to continue or retry previous response
|
||||
// - conversation list view
|
||||
// - change model
|
||||
// - rename conversation
|
||||
@ -24,6 +22,7 @@ import (
|
||||
"github.com/charmbracelet/bubbles/viewport"
|
||||
tea "github.com/charmbracelet/bubbletea"
|
||||
"github.com/charmbracelet/lipgloss"
|
||||
"github.com/muesli/reflow/wordwrap"
|
||||
)
|
||||
|
||||
type focusState int
|
||||
@ -33,6 +32,13 @@ const (
|
||||
focusMessages
|
||||
)
|
||||
|
||||
type editorTarget int
|
||||
|
||||
const (
|
||||
input editorTarget = iota
|
||||
selectedMessage
|
||||
)
|
||||
|
||||
type model struct {
|
||||
width int
|
||||
height int
|
||||
@ -44,16 +50,20 @@ type model struct {
|
||||
conversation *models.Conversation
|
||||
messages []models.Message
|
||||
waitingForReply bool
|
||||
editorTarget editorTarget
|
||||
stopSignal chan interface{}
|
||||
replyChan chan models.Message
|
||||
replyChunkChan chan string
|
||||
err error
|
||||
persistence bool // whether we will save new messages in the conversation
|
||||
err error
|
||||
|
||||
// ui state
|
||||
focus focusState
|
||||
status string // a general status message
|
||||
highlightCache []string // a cache of syntax highlighted message content
|
||||
focus focusState
|
||||
wrap bool // whether message content is wrapped to viewport width
|
||||
status string // a general status message
|
||||
highlightCache []string // a cache of syntax highlighted message content
|
||||
messageOffsets []int
|
||||
selectedMessage int
|
||||
|
||||
// ui elements
|
||||
content viewport.Model
|
||||
@ -90,11 +100,12 @@ type (
|
||||
var (
|
||||
userStyle = lipgloss.NewStyle().Faint(true).Bold(true).Foreground(lipgloss.Color("10"))
|
||||
assistantStyle = lipgloss.NewStyle().Faint(true).Bold(true).Foreground(lipgloss.Color("12"))
|
||||
messageStyle = lipgloss.NewStyle().PaddingLeft(1)
|
||||
messageStyle = lipgloss.NewStyle().PaddingLeft(2).PaddingRight(2)
|
||||
headerStyle = lipgloss.NewStyle().
|
||||
Background(lipgloss.Color("0"))
|
||||
contentStyle = lipgloss.NewStyle().
|
||||
Padding(1)
|
||||
conversationStyle = lipgloss.NewStyle().
|
||||
MarginTop(1).
|
||||
MarginBottom(1)
|
||||
footerStyle = lipgloss.NewStyle().
|
||||
BorderTop(true).
|
||||
BorderStyle(lipgloss.NormalBorder())
|
||||
@ -120,16 +131,35 @@ func (m model) Update(msg tea.Msg) (tea.Model, tea.Cmd) {
|
||||
var cmds []tea.Cmd
|
||||
|
||||
switch msg := msg.(type) {
|
||||
case msgTempfileEditorClosed:
|
||||
contents := string(msg)
|
||||
switch m.editorTarget {
|
||||
case input:
|
||||
m.input.SetValue(contents)
|
||||
case selectedMessage:
|
||||
m.setMessageContents(m.selectedMessage, contents)
|
||||
if m.persistence && m.messages[m.selectedMessage].ID > 0 {
|
||||
// update persisted message
|
||||
err := m.ctx.Store.UpdateMessage(&m.messages[m.selectedMessage])
|
||||
if err != nil {
|
||||
cmds = append(cmds, wrapError(fmt.Errorf("Could not save edited message: %v", err)))
|
||||
}
|
||||
}
|
||||
m.updateContent()
|
||||
}
|
||||
case tea.KeyMsg:
|
||||
switch msg.String() {
|
||||
case "ctrl+c":
|
||||
if m.waitingForReply {
|
||||
m.stopSignal <- "stahp!"
|
||||
m.stopSignal <- ""
|
||||
} else {
|
||||
return m, tea.Quit
|
||||
}
|
||||
case "ctrl+p":
|
||||
m.persistence = !m.persistence
|
||||
case "ctrl+w":
|
||||
m.wrap = !m.wrap
|
||||
m.updateContent()
|
||||
case "q":
|
||||
if m.focus != focusInput {
|
||||
return m, tea.Quit
|
||||
@ -177,11 +207,16 @@ func (m model) Update(msg tea.Msg) (tea.Model, tea.Cmd) {
|
||||
reply := models.Message(msg)
|
||||
last := len(m.messages) - 1
|
||||
if last < 0 {
|
||||
panic("Unexpected messages length handling msgReply")
|
||||
panic("Unexpected empty messages handling msgReply")
|
||||
}
|
||||
if reply.Role == models.MessageRoleToolCall && m.messages[last].Role == models.MessageRoleAssistant {
|
||||
m.setMessage(last, reply)
|
||||
} else if reply.Role != models.MessageRoleAssistant {
|
||||
m.setMessageContents(last, strings.TrimSpace(m.messages[last].Content))
|
||||
if m.messages[last].Role == models.MessageRoleAssistant {
|
||||
// the last message was an assistant message, so this is a continuation
|
||||
if reply.Role == models.MessageRoleToolCall {
|
||||
// update last message rrole to tool call
|
||||
m.messages[last].Role = models.MessageRoleToolCall
|
||||
}
|
||||
} else {
|
||||
m.addMessage(reply)
|
||||
}
|
||||
|
||||
@ -193,7 +228,7 @@ func (m model) Update(msg tea.Msg) (tea.Model, tea.Cmd) {
|
||||
if err != nil {
|
||||
cmds = append(cmds, wrapError(err))
|
||||
} else {
|
||||
cmds = append(cmds, m.persistRecentMessages())
|
||||
cmds = append(cmds, m.persistConversation())
|
||||
}
|
||||
}
|
||||
|
||||
@ -205,6 +240,12 @@ func (m model) Update(msg tea.Msg) (tea.Model, tea.Cmd) {
|
||||
cmds = append(cmds, m.waitForReply())
|
||||
case msgResponseEnd:
|
||||
m.waitingForReply = false
|
||||
last := len(m.messages) - 1
|
||||
if last < 0 {
|
||||
panic("Unexpected empty messages handling msgResponseEnd")
|
||||
}
|
||||
m.setMessageContents(last, strings.TrimSpace(m.messages[last].Content))
|
||||
m.updateContent()
|
||||
m.status = "Press ctrl+s to send"
|
||||
case msgResponseError:
|
||||
m.waitingForReply = false
|
||||
@ -376,11 +417,15 @@ func initialModel(ctx *lmcli.Context, convShortname string) model {
|
||||
stopSignal: make(chan interface{}),
|
||||
replyChan: make(chan models.Message),
|
||||
replyChunkChan: make(chan string),
|
||||
|
||||
wrap: true,
|
||||
selectedMessage: -1,
|
||||
}
|
||||
|
||||
m.content = viewport.New(0, 0)
|
||||
|
||||
m.input = textarea.New()
|
||||
m.input.CharLimit = 0
|
||||
m.input.Placeholder = "Enter a message"
|
||||
|
||||
m.input.FocusedStyle.CursorLine = lipgloss.NewStyle()
|
||||
@ -407,11 +452,60 @@ func initialModel(ctx *lmcli.Context, convShortname string) model {
|
||||
return m
|
||||
}
|
||||
|
||||
// fraction is the fraction of the total screen height into view the offset
|
||||
// should be scrolled into view. 0.5 = items will be snapped to middle of
|
||||
// view
|
||||
func scrollIntoView(vp *viewport.Model, offset int, fraction float32) {
|
||||
currentOffset := vp.YOffset
|
||||
if offset >= currentOffset && offset < currentOffset+vp.Height {
|
||||
return
|
||||
}
|
||||
distance := currentOffset - offset
|
||||
if distance < 0 {
|
||||
// we should scroll down until it just comes into view
|
||||
vp.SetYOffset(currentOffset - (distance + (vp.Height - int(float32(vp.Height)*fraction))) + 1)
|
||||
} else {
|
||||
// we should scroll up
|
||||
vp.SetYOffset(currentOffset - distance - int(float32(vp.Height)*fraction))
|
||||
}
|
||||
}
|
||||
|
||||
func (m *model) handleMessagesKey(msg tea.KeyMsg) tea.Cmd {
|
||||
switch msg.String() {
|
||||
case "tab":
|
||||
m.focus = focusInput
|
||||
m.updateContent()
|
||||
m.input.Focus()
|
||||
case "e":
|
||||
message := m.messages[m.selectedMessage]
|
||||
cmd := openTempfileEditor("message.*.md", message.Content, "# Edit the message below\n")
|
||||
m.editorTarget = selectedMessage
|
||||
return cmd
|
||||
case "ctrl+k":
|
||||
if m.selectedMessage > 0 && len(m.messages) == len(m.messageOffsets) {
|
||||
m.selectedMessage--
|
||||
m.updateContent()
|
||||
offset := m.messageOffsets[m.selectedMessage]
|
||||
scrollIntoView(&m.content, offset, 0.1)
|
||||
}
|
||||
case "ctrl+j":
|
||||
if m.selectedMessage < len(m.messages)-1 && len(m.messages) == len(m.messageOffsets) {
|
||||
m.selectedMessage++
|
||||
m.updateContent()
|
||||
offset := m.messageOffsets[m.selectedMessage]
|
||||
scrollIntoView(&m.content, offset, 0.1)
|
||||
}
|
||||
case "ctrl+r":
|
||||
// resubmit the conversation with all messages up until and including
|
||||
// the selected message
|
||||
if len(m.messages) == 0 {
|
||||
return nil
|
||||
}
|
||||
m.messages = m.messages[:m.selectedMessage+1]
|
||||
m.highlightCache = m.highlightCache[:m.selectedMessage+1]
|
||||
m.updateContent()
|
||||
m.content.GotoBottom()
|
||||
return m.promptLLM()
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@ -420,6 +514,10 @@ func (m *model) handleInputKey(msg tea.KeyMsg) tea.Cmd {
|
||||
switch msg.String() {
|
||||
case "esc":
|
||||
m.focus = focusMessages
|
||||
if m.selectedMessage < 0 || m.selectedMessage >= len(m.messages) {
|
||||
m.selectedMessage = len(m.messages) - 1
|
||||
}
|
||||
m.updateContent()
|
||||
m.input.Blur()
|
||||
case "ctrl+s":
|
||||
userInput := strings.TrimSpace(m.input.Value())
|
||||
@ -447,7 +545,7 @@ func (m *model) handleInputKey(msg tea.KeyMsg) tea.Cmd {
|
||||
|
||||
// ensure all messages up to the one we're about to add are
|
||||
// persistent
|
||||
cmd := m.persistRecentMessages()
|
||||
cmd := m.persistConversation()
|
||||
if cmd != nil {
|
||||
return cmd
|
||||
}
|
||||
@ -464,19 +562,11 @@ func (m *model) handleInputKey(msg tea.KeyMsg) tea.Cmd {
|
||||
|
||||
m.updateContent()
|
||||
m.content.GotoBottom()
|
||||
|
||||
return m.promptLLM()
|
||||
case "ctrl+r":
|
||||
if len(m.messages) == 0 {
|
||||
return nil
|
||||
}
|
||||
// TODO: retry from selected message
|
||||
if m.messages[len(m.messages)-1].Role != models.MessageRoleUser {
|
||||
m.messages = m.messages[:len(m.messages)-1]
|
||||
m.updateContent()
|
||||
}
|
||||
m.content.GotoBottom()
|
||||
return m.promptLLM()
|
||||
case "ctrl+e":
|
||||
cmd := openTempfileEditor("message.*.md", m.input.Value(), "# Edit your input below\n")
|
||||
m.editorTarget = input
|
||||
return cmd
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@ -573,16 +663,55 @@ func (m *model) promptLLM() tea.Cmd {
|
||||
}
|
||||
}
|
||||
|
||||
func (m *model) persistRecentMessages() tea.Cmd {
|
||||
func (m *model) persistConversation() tea.Cmd {
|
||||
existingMessages, err := m.ctx.Store.Messages(m.conversation)
|
||||
if err != nil {
|
||||
return wrapError(fmt.Errorf("Could not retrieve existing conversation messages while trying to save: %v", err))
|
||||
}
|
||||
|
||||
existingById := make(map[uint]*models.Message, len(existingMessages))
|
||||
for _, msg := range existingMessages {
|
||||
existingById[msg.ID] = &msg
|
||||
}
|
||||
|
||||
currentById := make(map[uint]*models.Message, len(m.messages))
|
||||
for _, msg := range m.messages {
|
||||
currentById[msg.ID] = &msg
|
||||
}
|
||||
|
||||
for _, msg := range existingMessages {
|
||||
_, ok := currentById[msg.ID]
|
||||
if !ok {
|
||||
err := m.ctx.Store.DeleteMessage(&msg)
|
||||
if err != nil {
|
||||
return wrapError(fmt.Errorf("Failed to remove messages: %v", err))
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
for i, msg := range m.messages {
|
||||
if msg.ID > 0 {
|
||||
continue
|
||||
exist, ok := existingById[msg.ID]
|
||||
if ok {
|
||||
if msg.Content == exist.Content {
|
||||
continue
|
||||
}
|
||||
// update message when contents don't match that of store
|
||||
err := m.ctx.Store.UpdateMessage(&msg)
|
||||
if err != nil {
|
||||
return wrapError(err)
|
||||
}
|
||||
} else {
|
||||
// this would be quite odd... and I'm not sure how to handle
|
||||
// it at the time of writing this
|
||||
}
|
||||
} else {
|
||||
newMessage, err := m.ctx.Store.AddReply(m.conversation, msg)
|
||||
if err != nil {
|
||||
return wrapError(err)
|
||||
}
|
||||
m.setMessage(i, *newMessage)
|
||||
}
|
||||
newMessage, err := m.ctx.Store.AddReply(m.conversation, msg)
|
||||
if err != nil {
|
||||
return wrapError(err)
|
||||
}
|
||||
m.setMessage(i, *newMessage)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@ -620,11 +749,25 @@ func (m *model) setMessageContents(i int, content string) {
|
||||
m.highlightCache[i] = highlighted
|
||||
}
|
||||
|
||||
// render the conversation into the main viewport
|
||||
func (m *model) updateContent() {
|
||||
atBottom := m.content.AtBottom()
|
||||
m.content.SetContent(m.conversationView())
|
||||
if atBottom {
|
||||
// if we were at bottom before the update, scroll with the output
|
||||
m.content.GotoBottom()
|
||||
}
|
||||
}
|
||||
|
||||
// render the conversation into a string
|
||||
func (m *model) conversationView() string {
|
||||
sb := strings.Builder{}
|
||||
msgCnt := len(m.messages)
|
||||
|
||||
m.messageOffsets = make([]int, len(m.messages))
|
||||
lineCnt := conversationStyle.GetMarginTop()
|
||||
for i, message := range m.messages {
|
||||
m.messageOffsets[i] = lineCnt
|
||||
|
||||
icon := "⚙️"
|
||||
friendly := message.Role.FriendlyRole()
|
||||
style := lipgloss.NewStyle().Bold(true).Faint(true)
|
||||
@ -640,37 +783,53 @@ func (m *model) updateContent() {
|
||||
icon = "🔧"
|
||||
}
|
||||
|
||||
// write message heading with space for content
|
||||
user := style.Render(icon + friendly)
|
||||
|
||||
var saved string
|
||||
if message.ID == 0 {
|
||||
saved = lipgloss.NewStyle().Faint(true).Render(" (not saved)")
|
||||
}
|
||||
|
||||
// write message heading with space for content
|
||||
header := fmt.Sprintf("%s\n\n", style.Render(icon+friendly)+saved)
|
||||
var selectedPrefix string
|
||||
if m.focus == focusMessages && i == m.selectedMessage {
|
||||
selectedPrefix = "> "
|
||||
}
|
||||
|
||||
header := lipgloss.NewStyle().PaddingLeft(1).Render(selectedPrefix + user + saved)
|
||||
sb.WriteString(header)
|
||||
lineCnt += lipgloss.Height(header)
|
||||
|
||||
// TODO: special rendering for tool calls/results?
|
||||
if message.Content != "" {
|
||||
sb.WriteString("\n\n")
|
||||
lineCnt += 1
|
||||
|
||||
// write message contents
|
||||
var highlighted string
|
||||
if m.highlightCache[i] == "" {
|
||||
highlighted = message.Content
|
||||
} else {
|
||||
highlighted = m.highlightCache[i]
|
||||
// write message contents
|
||||
var highlighted string
|
||||
if m.highlightCache[i] == "" {
|
||||
highlighted = message.Content
|
||||
} else {
|
||||
highlighted = m.highlightCache[i]
|
||||
}
|
||||
var contents string
|
||||
if m.wrap {
|
||||
wrapWidth := m.content.Width - messageStyle.GetHorizontalPadding() - 2
|
||||
wrapped := wordwrap.String(highlighted, wrapWidth)
|
||||
contents = wrapped
|
||||
} else {
|
||||
contents = highlighted
|
||||
}
|
||||
sb.WriteString(messageStyle.Width(0).Render(contents))
|
||||
lineCnt += lipgloss.Height(contents)
|
||||
}
|
||||
contents := messageStyle.Width(m.content.Width - 5).Render(highlighted)
|
||||
sb.WriteString(contents)
|
||||
|
||||
if i < msgCnt-1 {
|
||||
sb.WriteString("\n\n")
|
||||
lineCnt += 1
|
||||
}
|
||||
}
|
||||
atBottom := m.content.AtBottom()
|
||||
m.content.SetContent(contentStyle.Render(sb.String()))
|
||||
if atBottom {
|
||||
// if we were at bottom before the update, scroll with the output
|
||||
m.content.GotoBottom()
|
||||
}
|
||||
return conversationStyle.Render(sb.String())
|
||||
}
|
||||
|
||||
func Launch(ctx *lmcli.Context, convShortname string) error {
|
||||
|
42
pkg/tui/util.go
Normal file
42
pkg/tui/util.go
Normal file
@ -0,0 +1,42 @@
|
||||
package tui
|
||||
|
||||
import (
|
||||
"os"
|
||||
"os/exec"
|
||||
"strings"
|
||||
|
||||
tea "github.com/charmbracelet/bubbletea"
|
||||
)
|
||||
|
||||
type msgTempfileEditorClosed string
|
||||
|
||||
// openTempfileEditor opens an $EDITOR on a new temporary file with the given
|
||||
// content. Upon closing, the contents of the file are read back returned
|
||||
// wrapped in a msgTempfileEditorClosed returned by the tea.Cmd
|
||||
func openTempfileEditor(pattern string, content string, placeholder string) tea.Cmd {
|
||||
msgFile, _ := os.CreateTemp("/tmp", pattern)
|
||||
|
||||
err := os.WriteFile(msgFile.Name(), []byte(placeholder+content), os.ModeAppend)
|
||||
if err != nil {
|
||||
return wrapError(err)
|
||||
}
|
||||
|
||||
editor := os.Getenv("EDITOR")
|
||||
if editor == "" {
|
||||
editor = "vim"
|
||||
}
|
||||
|
||||
c := exec.Command(editor, msgFile.Name())
|
||||
return tea.ExecProcess(c, func(err error) tea.Msg {
|
||||
bytes, err := os.ReadFile(msgFile.Name())
|
||||
if err != nil {
|
||||
return msgError(err)
|
||||
}
|
||||
fileContents := string(bytes)
|
||||
if strings.HasPrefix(fileContents, placeholder) {
|
||||
fileContents = fileContents[len(placeholder):]
|
||||
}
|
||||
stripped := strings.Trim(fileContents, "\n \t")
|
||||
return msgTempfileEditorClosed(stripped)
|
||||
})
|
||||
}
|
Loading…
Reference in New Issue
Block a user