* Add config option for controlling Ollama think parameter
Allows enabling or disable thinking for supported models. Neither option
will dislay thinking content in the chat. Future support for displaying
think content will require frontend changes for formatting.
* Add thinking strings
* Update ollama to use the ChatLog/ChatSession APIs
* Add documentation about history trimming.
* Revert changes to chat_log.py
* Explicitly check for SystemContent when converting system messages
* Remove half of a comment
* Improve quality of ollama function calling by repairing function call arguments
* Fix formatting of the tests
* Run ruff format on ollama conversation
* Add test for non-string arguments
* Add LLM tools support for Ollama
* fix tests
* coverage
* Separate call for tool parameters
* Fix example
* hint on parameters schema if LLM forgot to request it
* Switch to native tool call functionality
* Fix tests
* Fix tools list
* update strings and default model
* Ignore mypy error until fixed upstream
* Ignore mypy error until fixed upstream
* Add missing prompt part
* Update default model