* refactor: Move http client into provider
* refactor: Rework ollama provider for work with internal http client
* refactor: Rework LM Studio provider to work with internal http client
* refactor: Rework Mistral AI to work with internal http client
* fix: Replace url and header to QNetworkRequest
* refactor: Rework Google provider to use internal http client
* refactor: OpenAI compatible providers switch to use internal http client
* fix: Remove m_requestHandler from tests
* refactor: Remove old handleData method
* fix: Remove LLMClientInterfaceTest
This commit is a result of the following commands:
clang-format-19 --style=file -i $(git ls-files | fgrep .cpp)
clang-format-19 --style=file -i $(git ls-files | fgrep .hpp)
* feat: Add attachments for message
* feat: Support QtC color palette for chat view
* feat: Improve code completion from non-FIM models
* refactor: Removed trimming messages
* chore: Bump version to 0.4.4
* feat: Add provider settings
* feat: Add Claude provider
* feat: Add Claude templates
* refactor: Setting input sensitivity
* fix: Back text after read code block
* fix: Add missing system message for ollama fim
* feat: Add automatic template handling for Ollama models
- Add OllamaAutoFim
- Use native Ollama API format when possible
- Remove need for manual template selection for most Ollama models
- Default to model-specific format from Ollama modelfile
- Fallback to manual template selection if needed
This change simplifies configuration by automatically using
the correct template format for each Ollama model.