Private
Public Access
1
0

Properly support per-model maxTokens/temperature

This commit is contained in:
2025-06-25 07:49:29 +00:00
parent 259648f699
commit 3cd897d494
3 changed files with 76 additions and 32 deletions

View File

@@ -39,9 +39,9 @@ func NewAppModel(ctx *lmcli.Context, initialConversation *conversation.Conversat
}
model, provider, _, _ := ctx.GetModelProvider(*ctx.Config.Defaults.Model, "")
app.Model = model
app.ProviderName = provider
modelConfig, _ := ctx.GetModelProvider(*ctx.Config.Defaults.Model, "")
app.Model = modelConfig.Model
app.ProviderName = modelConfig.Provider
app.Agent = ctx.GetAgent(ctx.Config.Defaults.Agent)
return app
}
@@ -257,15 +257,16 @@ func (a *AppModel) Prompt(
chatReplyChunks chan provider.Chunk,
stopSignal chan struct{},
) (*conversation.Message, error) {
model, _, p, err := a.Ctx.GetModelProvider(a.Model, a.ProviderName)
modelConfig, err := a.Ctx.GetModelProvider(a.Model, a.ProviderName)
if err != nil {
return nil, err
}
p := modelConfig.Client
params := provider.RequestParameters{
Model: model,
MaxTokens: *a.Ctx.Config.Defaults.MaxTokens,
Temperature: *a.Ctx.Config.Defaults.Temperature,
Model: modelConfig.Model,
MaxTokens: modelConfig.MaxTokens,
Temperature: modelConfig.Temperature,
}
if a.Agent != nil {