diff --git a/go.mod b/go.mod index 0a5fe897..bb049890 100644 --- a/go.mod +++ b/go.mod @@ -22,7 +22,7 @@ require ( github.com/disintegration/imaging v1.6.2 github.com/fumiama/ahsai v0.1.0 github.com/fumiama/cron v1.3.0 - github.com/fumiama/deepinfra v0.0.0-20250910022828-8cde75e137f4 + github.com/fumiama/deepinfra v0.0.0-20250910144855-27a4e697106d github.com/fumiama/go-base16384 v1.7.0 github.com/fumiama/go-registry v0.2.7 github.com/fumiama/gotracemoe v0.0.3 diff --git a/go.sum b/go.sum index 54224894..9a54a965 100644 --- a/go.sum +++ b/go.sum @@ -59,8 +59,8 @@ github.com/fumiama/ahsai v0.1.0 h1:LXD61Kaj6kJHa3AEGsLIfKNzcgaVxg7JB72OR4yNNZ4= github.com/fumiama/ahsai v0.1.0/go.mod h1:fFeNnqgo44i8FIaguK659aQryuZeFy+4klYLQu/rfdk= github.com/fumiama/cron v1.3.0 h1:ZWlwuexF+HQHl3cYytEE5HNwD99q+3vNZF1GrEiXCFo= github.com/fumiama/cron v1.3.0/go.mod h1:bz5Izvgi/xEUI8tlBN8BI2jr9Moo8N4or0KV8xXuPDY= -github.com/fumiama/deepinfra v0.0.0-20250910022828-8cde75e137f4 h1:cV3HXXLNudIL9rIEYt1RCgl6H4703nE3+jL4pJNsRtc= -github.com/fumiama/deepinfra v0.0.0-20250910022828-8cde75e137f4/go.mod h1:wW05PQSn8mo1mZIoa6LBUE+3xIBjkoONvnfPTV5ZOhY= +github.com/fumiama/deepinfra v0.0.0-20250910144855-27a4e697106d h1:iGxnST620IHrJ47DXkjzrZJ2rskBogWze+UyvnAxT6g= +github.com/fumiama/deepinfra v0.0.0-20250910144855-27a4e697106d/go.mod h1:wW05PQSn8mo1mZIoa6LBUE+3xIBjkoONvnfPTV5ZOhY= github.com/fumiama/go-base16384 v1.7.0 h1:6fep7XPQWxRlh4Hu+KsdH+6+YdUp+w6CwRXtMWSsXCA= github.com/fumiama/go-base16384 v1.7.0/go.mod h1:OEn+947GV5gsbTAnyuUW/SrfxJYUdYupSIQXOuGOcXM= github.com/fumiama/go-registry v0.2.7 h1:tLEqgEpsiybQMqBv0dLHm5leia/z1DhajMupwnOHeNs= diff --git a/plugin/aichat/main.go b/plugin/aichat/main.go index 50c875a9..fc78af58 100644 --- a/plugin/aichat/main.go +++ b/plugin/aichat/main.go @@ -126,9 +126,7 @@ func init() { mod = model.NewOpenAI( cfg.ModelName, cfg.Separator, temperature, topp, maxn, - ).SetExtra(&map[string]bool{ - "enable_thinking": false, - }) + ) case 1: mod = model.NewOLLaMA( cfg.ModelName, cfg.Separator, @@ -518,9 +516,7 @@ func llmchat(prompt string, temp int64) (string, error) { mod = model.NewOpenAI( cfg.ModelName, cfg.Separator, temperature, topp, maxn, - ).SetExtra(&map[string]bool{ - "enable_thinking": false, - }) + ) case 1: mod = model.NewOLLaMA( cfg.ModelName, cfg.Separator,