feat(neovim): configure local llm models on avante
This commit is contained in:
parent
9fe182d987
commit
416c8e953a
@ -435,6 +435,24 @@ return {
|
||||
return hub and hub:get_active_servers_prompt() or ''
|
||||
end,
|
||||
providers = {
|
||||
['deepseek-r1-local'] = {
|
||||
__inherited_from = 'ollama',
|
||||
api_key_name = '',
|
||||
endpoint = 'http://172.30.64.1:10000',
|
||||
model = 'deepseek-r1:14b',
|
||||
},
|
||||
['gemma3-local'] = {
|
||||
__inherited_from = 'ollama',
|
||||
api_key_name = '',
|
||||
endpoint = 'http://172.30.64.1:10000',
|
||||
model = 'gemma3:12b',
|
||||
},
|
||||
['qwen3-local'] = {
|
||||
__inherited_from = 'ollama',
|
||||
api_key_name = '',
|
||||
endpoint = 'http://172.30.64.1:10000',
|
||||
model = 'qwen3:14b',
|
||||
},
|
||||
['gemini-pro'] = {
|
||||
__inherited_from = 'gemini',
|
||||
model = 'gemini-2.5-pro-preview-05-06',
|
||||
|
Loading…
x
Reference in New Issue
Block a user