From 036a55effc894b9937fab5e1685effcedb521b58 Mon Sep 17 00:00:00 2001 From: Christian Nieves Date: Tue, 17 Jun 2025 21:38:27 +0000 Subject: [PATCH] goose --- vim/.vim/lua/plugins/avante.lua | 11 +---------- 1 file changed, 1 insertion(+), 10 deletions(-) diff --git a/vim/.vim/lua/plugins/avante.lua b/vim/.vim/lua/plugins/avante.lua index 4f59c1d..383ecf5 100644 --- a/vim/.vim/lua/plugins/avante.lua +++ b/vim/.vim/lua/plugins/avante.lua @@ -18,16 +18,7 @@ return { url = "sso://user/vintharas/avante-goose.nvim", cond = use_google(), opts = { - -- Add your options here - -- These are the defaults - auto_start_backend = true, -- Whether to automatically start go/devai-api-http-proxy. If false you can use :AvanteGooseServerStart to start the server - auto_start_silent = true, -- Whether to have a silent auto start (don't log status messages) - model = "goose-v3.5-s", -- Select model from go/goose-models. - temperature = 0.1, -- Model temperature - max_decoder_steps = 8192, -- Maximum decoder steps (This affects the token limit of the output. More decoder steps -> higher limit in output tokens) - endpoint = "http://localhost:8080/predict", -- Endpoint to start/listen to go/devai-api-http-proxy - debug = false, -- Enables debug mode (outputs lots of logs for troubleshooting issues) - debug_backend = false, -- Whether to start the backend in debug mode. This logs backend output information under stdpath('cache')/devai-http-wrapper.log + model = "gemini-for-google-2.5-pro", -- Select model from go/goose-models. }, }, },