summaryrefslogtreecommitdiff
path: root/internal/hexaicli/run_model_override_test.go
blob: b32b172f3c5495cdea385fc63646da0caad8d075 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
package hexaicli

import (
	"bytes"
	"context"
	"strings"
	"testing"

	"codeberg.org/snonux/hexai/internal/appconfig"
	"codeberg.org/snonux/hexai/internal/llm"
)

type fakeClientModelEnv struct{ name, model string }

func (f fakeClientModelEnv) Chat(_ context.Context, _ []llm.Message, _ ...llm.RequestOption) (string, error) {
	return "ok", nil
}
func (f fakeClientModelEnv) Name() string         { return f.name }
func (f fakeClientModelEnv) DefaultModel() string { return f.model }

// Ensure that HEXAI_MODEL overrides config for CLI runs.
func TestRun_ModelEnvOverride_FlowsIntoClient(t *testing.T) {
	t.Setenv("XDG_CONFIG_HOME", t.TempDir())
	t.Setenv("HEXAI_MODEL", "gpt-5-codex")
	t.Setenv("HEXAI_PROVIDER", "openai")
	// Replace client constructor to assert model was overridden
	oldNew := newClientFromApp
	defer func() { newClientFromApp = oldNew }()
	var seenModel string
	newClientFromApp = func(cfg appconfig.App) (llm.Client, error) {
		seenModel = strings.TrimSpace(cfg.OpenAIModel)
		return fakeClientModelEnv{name: "openai", model: cfg.OpenAIModel}, nil
	}

	var out, errb bytes.Buffer
	if err := Run(context.Background(), []string{"hello"}, strings.NewReader(""), &out, &errb); err != nil {
		t.Fatalf("run error: %v", err)
	}
	if seenModel != "gpt-5-codex" {
		t.Fatalf("expected cfg.OpenAIModel=gpt-5-codex, got %q", seenModel)
	}
	if !strings.Contains(errb.String(), "model=gpt-5-codex") {
		t.Fatalf("stderr should print effective model, got: %s", errb.String())
	}
}