Fixed handling of long (slash separated) and short model identifiers
Renamed `GetCompletionProvider` to `GetModelProvider` and update it to return the model's short name (the one to use when making requests)
This commit is contained in:
@@ -1051,13 +1051,13 @@ func (m *Model) promptLLM() tea.Cmd {
|
||||
m.elapsed = 0
|
||||
|
||||
return func() tea.Msg {
|
||||
completionProvider, err := m.State.Ctx.GetCompletionProvider(*m.State.Ctx.Config.Defaults.Model)
|
||||
model, provider, err := m.State.Ctx.GetModelProvider(*m.State.Ctx.Config.Defaults.Model)
|
||||
if err != nil {
|
||||
return shared.MsgError(err)
|
||||
}
|
||||
|
||||
requestParams := models.RequestParameters{
|
||||
Model: *m.State.Ctx.Config.Defaults.Model,
|
||||
Model: model,
|
||||
MaxTokens: *m.State.Ctx.Config.Defaults.MaxTokens,
|
||||
Temperature: *m.State.Ctx.Config.Defaults.Temperature,
|
||||
ToolBag: m.State.Ctx.EnabledTools,
|
||||
@@ -1078,7 +1078,7 @@ func (m *Model) promptLLM() tea.Cmd {
|
||||
}
|
||||
}()
|
||||
|
||||
resp, err := completionProvider.CreateChatCompletionStream(
|
||||
resp, err := provider.CreateChatCompletionStream(
|
||||
ctx, requestParams, toPrompt, replyHandler, m.replyChunkChan,
|
||||
)
|
||||
|
||||
|
||||
Reference in New Issue
Block a user