if true then return {} else return { 'gerazov/ollama-chat.nvim', dependencies = { 'nvim-lua/plenary.nvim', 'stevearc/dressing.nvim', 'nvim-telescope/telescope.nvim', }, -- lazy load on command cmd = { 'OllamaQuickChat', 'OllamaCreateNewChat', 'OllamaContinueChat', 'OllamaChat', 'OllamaChatCode', 'OllamaModel', 'OllamaServe', 'OllamaServeStop', }, keys = { { 'ocq', 'OllamaQuickChat', desc = 'Ollama Quick Chat', mode = { 'n', 'x' }, silent = true, }, { 'ocn', 'OllamaCreateNewChat', desc = 'Create Ollama Chat', mode = { 'n', 'x' }, silent = true, }, { 'occ', 'OllamaContinueChat', desc = 'Continue Ollama Chat', mode = { 'n', 'x' }, silent = true, }, { 'och', 'OllamaChat', desc = 'Chat', mode = { 'n' }, silent = true, }, { 'ocd', 'OllamaChatCode', desc = 'Chat Code', mode = { 'n' }, silent = true, }, }, opts = { chats_folder = vim.fn.stdpath 'data', -- data folder is ~/.local/share/nvim -- you can also choose "current" and "tmp" quick_chat_file = 'ollama-chat.md', animate_spinner = true, -- set this to false to disable spinner animation model = 'openhermes2-mistral', model_code = 'codellama', url = 'http://127.0.0.1:11434', serve = { on_start = false, command = 'ollama', args = { 'serve' }, stop_command = 'pkill', stop_args = { '-SIGTERM', 'ollama' }, }, }, } end