From 6b86637dcb48c09b6a96684a398226a4e84a30fa Mon Sep 17 00:00:00 2001
From: Petr Mironychev <9195189+Palm1r@users.noreply.github.com>
Date: Mon, 10 Mar 2025 12:07:39 +0100
Subject: [PATCH] doc: Add llama.cpp description to README.md
---
README.md | 36 +++++++++++++++++++++++++-----------
1 file changed, 25 insertions(+), 11 deletions(-)
diff --git a/README.md b/README.md
index 5207551..9055f93 100644
--- a/README.md
+++ b/README.md
@@ -27,17 +27,18 @@
2. [Install plugin to QtCreator](#install-plugin-to-qtcreator)
3. [Configure for Anthropic Claude](#configure-for-anthropic-claude)
4. [Configure for OpenAI](#configure-for-openai)
-4. [Configure for Mistral AI](#configure-for-mistral-ai)
-4. [Configure for Google AI](#configure-for-google-ai)
-5. [Configure for Ollama](#configure-for-ollama)
-6. [System Prompt Configuration](#system-prompt-configuration)
-7. [File Context Features](#file-context-features)
-9. [QtCreator Version Compatibility](#qtcreator-version-compatibility)
-10. [Development Progress](#development-progress)
-11. [Hotkeys](#hotkeys)
-12. [Troubleshooting](#troubleshooting)
-13. [Support the Development](#support-the-development-of-qodeassist)
-14. [How to Build](#how-to-build)
+5. [Configure for Mistral AI](#configure-for-mistral-ai)
+6. [Configure for Google AI](#configure-for-google-ai)
+7. [Configure for Ollama](#configure-for-ollama)
+8. [Configure for llama.cpp](#configure-for-llamacpp)
+9. [System Prompt Configuration](#system-prompt-configuration)
+10. [File Context Features](#file-context-features)
+11. [QtCreator Version Compatibility](#qtcreator-version-compatibility)
+12. [Development Progress](#development-progress)
+13. [Hotkeys](#hotkeys)
+14. [Troubleshooting](#troubleshooting)
+15. [Support the Development](#support-the-development-of-qodeassist)
+16. [How to Build](#how-to-build)
## Overview
@@ -51,6 +52,7 @@
- Automatic syncing with open editor files (optional)
- Support for multiple LLM providers:
- Ollama
+ - llama.cpp
- OpenAI
- Anthropic Claude
- LM Studio
@@ -184,6 +186,18 @@ You're all set! QodeAssist is now ready to use in Qt Creator.
+## Configure for llama.cpp
+1. Open Qt Creator settings and navigate to the QodeAssist section
+2. Go to General tab and configure:
+ - Set "llama.cpp" as the provider for code completion or/and chat assistant
+ - Set the llama.cpp URL (e.g. http://localhost:8080)
+ - Fill in model name
+ - Choose template for model(e.g. llama.cpp FIM for any model with FIM support)
+
+ Example of llama.cpp settings: (click to expand)
+
+
+
## System Prompt Configuration
The plugin comes with default system prompts optimized for chat and instruct models, as these currently provide better results for code assistance. If you prefer using FIM (Fill-in-Middle) models, you can easily customize the system prompt in the settings.