mirror of
https://github.com/ryan4yin/nix-config.git
synced 2026-04-27 03:08:38 +02:00
feat: neovim - update avante.nvim & ollama
This commit is contained in:
@@ -19,26 +19,27 @@ return {
|
||||
},
|
||||
version = false, -- Never set this value to "*"! Never!
|
||||
opts = {
|
||||
-- add any opts here
|
||||
provider = "deepseek_coder",
|
||||
provider = "deepseek_reasoner",
|
||||
cursor_applying_provider = "deepseek_reasoner", -- In this example, use Groq for applying, but you can also use any provider you want.
|
||||
behaviour = {
|
||||
-- auto_suggestions = true,
|
||||
enable_cursor_planning_mode = true, -- enable cursor planning mode!
|
||||
},
|
||||
-- WARNING: Since auto-suggestions are a high-frequency operation and therefore expensive,
|
||||
-- currently designating it as `copilot` provider is dangerous because: https://github.com/yetone/avante.nvim/issues/1048
|
||||
-- Of course, you can reduce the request frequency by increasing `suggestion.debounce`.
|
||||
auto_suggestions_provider = "aliyun_qwen3",
|
||||
suggestion = {
|
||||
debounce = 750, -- wait for x ms before suggestion
|
||||
throttle = 1200, -- wait for at least x ms before the next suggestion
|
||||
},
|
||||
|
||||
ollama = {
|
||||
endpoint = "http://127.0.0.1:11434", -- Note that there is no /v1 at the end.
|
||||
endpoint = "http://192.168.5.100:11434", -- Note that there is no /v1 at the end.
|
||||
model = "modelscope.cn/unsloth/Qwen3-30B-A3B-GGUF",
|
||||
-- model = "modelscope.cn/unsloth/Qwen3-235B-A22B-GGUF",
|
||||
},
|
||||
vendors = {
|
||||
openrouter_claude_4_7_sonnet = {
|
||||
__inherited_from = "openai",
|
||||
endpoint = "https://openrouter.ai/api/v1",
|
||||
api_key_name = "OPENROUTER_API_KEY",
|
||||
model = "anthropic/claude-3.7-sonnet",
|
||||
},
|
||||
openrouter_gemini_2_flash = {
|
||||
__inherited_from = "openai",
|
||||
endpoint = "https://openrouter.ai/api/v1",
|
||||
api_key_name = "OPENROUTER_API_KEY",
|
||||
model = "google/gemini-2.0-flash-001",
|
||||
},
|
||||
deepseek_coder = {
|
||||
__inherited_from = "openai",
|
||||
api_key_name = "DEEPSEEK_API_KEY",
|
||||
@@ -67,14 +68,6 @@ return {
|
||||
model = "qwen3-235b-a22b",
|
||||
},
|
||||
},
|
||||
-- WARNING: Since auto-suggestions are a high-frequency operation and therefore expensive,
|
||||
-- currently designating it as `copilot` provider is dangerous because: https://github.com/yetone/avante.nvim/issues/1048
|
||||
-- Of course, you can reduce the request frequency by increasing `suggestion.debounce`.
|
||||
auto_suggestions_provider = "openrouter_gemini_2_flash",
|
||||
suggestion = {
|
||||
debounce = 750, -- wait for x ms before suggestion
|
||||
throttle = 1200, -- wait for at least x ms before the next suggestion
|
||||
},
|
||||
},
|
||||
-- if you want to build from source then do `make BUILD_FROM_SOURCE=true`
|
||||
build = "make",
|
||||
|
||||
Reference in New Issue
Block a user