/* * Copyright (C) 2024-2025 Petr Mironychev * * This file is part of QodeAssist. * * QodeAssist is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * QodeAssist is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with QodeAssist. If not, see . */ #include "OllamaProvider.hpp" #include #include #include #include #include #include "llmcore/OllamaMessage.hpp" #include "llmcore/ValidationUtils.hpp" #include "logger/Logger.hpp" #include "settings/ChatAssistantSettings.hpp" #include "settings/CodeCompletionSettings.hpp" #include "settings/ProviderSettings.hpp" namespace QodeAssist::Providers { QString OllamaProvider::name() const { return "Ollama"; } QString OllamaProvider::url() const { return "http://localhost:11434"; } QString OllamaProvider::completionEndpoint() const { return "/api/generate"; } QString OllamaProvider::chatEndpoint() const { return "/api/chat"; } bool OllamaProvider::supportsModelListing() const { return true; } void OllamaProvider::prepareRequest( QJsonObject &request, LLMCore::PromptTemplate *prompt, LLMCore::ContextData context, LLMCore::RequestType type) { if (!prompt->isSupportProvider(providerID())) { LOG_MESSAGE(QString("Template %1 doesn't support %2 provider").arg(name(), prompt->name())); } prompt->prepareRequest(request, context); auto applySettings = [&request](const auto &settings) { QJsonObject options; options["num_predict"] = settings.maxTokens(); options["temperature"] = settings.temperature(); options["stop"] = request.take("stop"); if (settings.useTopP()) options["top_p"] = settings.topP(); if (settings.useTopK()) options["top_k"] = settings.topK(); if (settings.useFrequencyPenalty()) options["frequency_penalty"] = settings.frequencyPenalty(); if (settings.usePresencePenalty()) options["presence_penalty"] = settings.presencePenalty(); request["options"] = options; request["keep_alive"] = settings.ollamaLivetime(); }; if (type == LLMCore::RequestType::CodeCompletion) { applySettings(Settings::codeCompletionSettings()); } else { applySettings(Settings::chatAssistantSettings()); } } QList OllamaProvider::getInstalledModels(const QString &url) { QList models; QNetworkAccessManager manager; QNetworkRequest request(QString("%1%2").arg(url, "/api/tags")); prepareNetworkRequest(request); QNetworkReply *reply = manager.get(request); QEventLoop loop; QObject::connect(reply, &QNetworkReply::finished, &loop, &QEventLoop::quit); loop.exec(); if (reply->error() == QNetworkReply::NoError) { QByteArray responseData = reply->readAll(); QJsonDocument jsonResponse = QJsonDocument::fromJson(responseData); QJsonObject jsonObject = jsonResponse.object(); QJsonArray modelArray = jsonObject["models"].toArray(); for (const QJsonValue &value : modelArray) { QJsonObject modelObject = value.toObject(); QString modelName = modelObject["name"].toString(); models.append(modelName); } } else { LOG_MESSAGE(QString("Error fetching models: %1").arg(reply->errorString())); } reply->deleteLater(); return models; } QList OllamaProvider::validateRequest(const QJsonObject &request, LLMCore::TemplateType type) { const auto fimReq = QJsonObject{ {"keep_alive", {}}, {"model", {}}, {"stream", {}}, {"prompt", {}}, {"suffix", {}}, {"system", {}}, {"options", QJsonObject{ {"temperature", {}}, {"stop", {}}, {"top_p", {}}, {"top_k", {}}, {"num_predict", {}}, {"frequency_penalty", {}}, {"presence_penalty", {}}}}}; const auto messageReq = QJsonObject{ {"keep_alive", {}}, {"model", {}}, {"stream", {}}, {"messages", QJsonArray{{QJsonObject{{"role", {}}, {"content", {}}}}}}, {"options", QJsonObject{ {"temperature", {}}, {"stop", {}}, {"top_p", {}}, {"top_k", {}}, {"num_predict", {}}, {"frequency_penalty", {}}, {"presence_penalty", {}}}}}; return LLMCore::ValidationUtils::validateRequestFields( request, type == LLMCore::TemplateType::FIM ? fimReq : messageReq); } QString OllamaProvider::apiKey() const { return {}; } void OllamaProvider::prepareNetworkRequest(QNetworkRequest &networkRequest) const { networkRequest.setHeader(QNetworkRequest::ContentTypeHeader, "application/json"); const auto key = Settings::providerSettings().ollamaBasicAuthApiKey(); if (!key.isEmpty()) { networkRequest.setRawHeader("Authorization", "Basic " + key.toLatin1()); } } LLMCore::ProviderID OllamaProvider::providerID() const { return LLMCore::ProviderID::Ollama; } void OllamaProvider::sendRequest( const QString &requestId, const QUrl &url, const QJsonObject &payload) { m_dataBuffers[requestId].clear(); m_requestUrls[requestId] = url; QNetworkRequest networkRequest(url); prepareNetworkRequest(networkRequest); LLMCore::HttpRequest request{.networkRequest = networkRequest, .requestId = requestId, .payload = payload}; LOG_MESSAGE(QString("OllamaProvider: Sending request %1 to %2").arg(requestId, url.toString())); emit httpClient()->sendRequest(request); } void OllamaProvider::onDataReceived(const QString &requestId, const QByteArray &data) { LLMCore::DataBuffers &buffers = m_dataBuffers[requestId]; QStringList lines = buffers.rawStreamBuffer.processData(data); if (data.isEmpty()) { return; } bool isDone = false; QString tempResponse; for (const QString &line : lines) { if (line.trimmed().isEmpty()) { continue; } QJsonParseError error; QJsonDocument doc = QJsonDocument::fromJson(line.toUtf8(), &error); if (doc.isNull()) { continue; } QJsonObject obj = doc.object(); if (obj.contains("error") && !obj["error"].toString().isEmpty()) { LOG_MESSAGE("Error in Ollama response: " + obj["error"].toString()); continue; } QString content; if (obj.contains("response")) { content = obj["response"].toString(); } else if (obj.contains("message")) { QJsonObject messageObj = obj["message"].toObject(); content = messageObj["content"].toString(); } if (!content.isEmpty()) { tempResponse += content; } if (obj["done"].toBool()) { isDone = true; } } if (!tempResponse.isEmpty()) { buffers.responseContent += tempResponse; emit partialResponseReceived(requestId, tempResponse); } if (isDone) { emit fullResponseReceived(requestId, buffers.responseContent); m_dataBuffers.remove(requestId); } } void OllamaProvider::onRequestFinished(const QString &requestId, bool success, const QString &error) { if (!success) { LOG_MESSAGE(QString("OllamaProvider request %1 failed: %2").arg(requestId, error)); emit requestFailed(requestId, error); } else { if (m_dataBuffers.contains(requestId)) { const LLMCore::DataBuffers &buffers = m_dataBuffers[requestId]; if (!buffers.responseContent.isEmpty()) { emit fullResponseReceived(requestId, buffers.responseContent); } } } m_dataBuffers.remove(requestId); m_requestUrls.remove(requestId); } } // namespace QodeAssist::Providers