summaryrefslogtreecommitdiff
path: root/hexai/config.toml
diff options
context:
space:
mode:
authorPaul Buetow <paul@buetow.org>2026-05-02 17:55:52 +0300
committerPaul Buetow <paul@buetow.org>2026-05-02 17:55:52 +0300
commit62913b518d0fb99913e93314d9d4407730b7f32e (patch)
treeec89b6c394435afc805aa9b463bae4bfa5c482be /hexai/config.toml
parente5cd3949c477308709f0052bbd0fae68a970bfa7 (diff)
Update
Diffstat (limited to 'hexai/config.toml')
-rw-r--r--hexai/config.toml16
1 files changed, 13 insertions, 3 deletions
diff --git a/hexai/config.toml b/hexai/config.toml
index 1a6dff1..5954eae 100644
--- a/hexai/config.toml
+++ b/hexai/config.toml
@@ -17,12 +17,22 @@ name = "ollama"
[ollama]
model = "gemma4:31b-cloud"
+# Smaller Gemma variants on Ollama Cloud — uncomment one if gemma4:31b-cloud
+# becomes too slow:
+# model = "gemma3:27b"
+# model = "gemma3:12b"
+# model = "gemma3:4b"
base_url = "https://ollama.com"
temperature = 0.2
-# In-code auto-completion uses gemma4:31b-cloud (the dense Gemma 4 hosted on
-# Ollama Cloud). Latency-sensitive completions use the same model as chat.
+# In-code auto-completion uses gemma3:12b on Ollama Cloud — a good balance of
+# speed and quality. Switch to a larger or smaller Gemma below if needed.
[[models.completion]]
provider = "ollama"
-model = "gemma4:31b-cloud"
+model = "gemma3:12b"
+# Other Gemma variants on Ollama Cloud — uncomment one if gemma3:12b is not
+# the right fit:
+# model = "gemma4:31b-cloud"
+# model = "gemma3:27b"
+# model = "gemma3:4b"
temperature = 0.2