Properly support per-model maxTokens/temperature
This commit is contained in:
@@ -39,9 +39,9 @@ func NewAppModel(ctx *lmcli.Context, initialConversation *conversation.Conversat
|
||||
|
||||
}
|
||||
|
||||
model, provider, _, _ := ctx.GetModelProvider(*ctx.Config.Defaults.Model, "")
|
||||
app.Model = model
|
||||
app.ProviderName = provider
|
||||
modelConfig, _ := ctx.GetModelProvider(*ctx.Config.Defaults.Model, "")
|
||||
app.Model = modelConfig.Model
|
||||
app.ProviderName = modelConfig.Provider
|
||||
app.Agent = ctx.GetAgent(ctx.Config.Defaults.Agent)
|
||||
return app
|
||||
}
|
||||
@@ -257,15 +257,16 @@ func (a *AppModel) Prompt(
|
||||
chatReplyChunks chan provider.Chunk,
|
||||
stopSignal chan struct{},
|
||||
) (*conversation.Message, error) {
|
||||
model, _, p, err := a.Ctx.GetModelProvider(a.Model, a.ProviderName)
|
||||
modelConfig, err := a.Ctx.GetModelProvider(a.Model, a.ProviderName)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
p := modelConfig.Client
|
||||
|
||||
params := provider.RequestParameters{
|
||||
Model: model,
|
||||
MaxTokens: *a.Ctx.Config.Defaults.MaxTokens,
|
||||
Temperature: *a.Ctx.Config.Defaults.Temperature,
|
||||
Model: modelConfig.Model,
|
||||
MaxTokens: modelConfig.MaxTokens,
|
||||
Temperature: modelConfig.Temperature,
|
||||
}
|
||||
|
||||
if a.Agent != nil {
|
||||
|
||||
Reference in New Issue
Block a user