/* * Copyright (C) 2024 Petr Mironychev * * This file is part of QodeAssist. * * QodeAssist is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * QodeAssist is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with QodeAssist. If not, see . */ #include "OllamaProvider.hpp" #include #include #include #include #include #include "OllamaMessage.hpp" #include "logger/Logger.hpp" #include "settings/ChatAssistantSettings.hpp" #include "settings/CodeCompletionSettings.hpp" namespace QodeAssist::Providers { OllamaProvider::OllamaProvider() {} QString OllamaProvider::name() const { return "Ollama"; } QString OllamaProvider::url() const { return "http://localhost:11434"; } QString OllamaProvider::completionEndpoint() const { return "/api/generate"; } QString OllamaProvider::chatEndpoint() const { return "/api/chat"; } bool OllamaProvider::supportsModelListing() const { return true; } void OllamaProvider::prepareRequest(QJsonObject &request, LLMCore::RequestType type) { auto applySettings = [&request](const auto &settings) { QJsonObject options; options["num_predict"] = settings.maxTokens(); options["temperature"] = settings.temperature(); if (settings.useTopP()) options["top_p"] = settings.topP(); if (settings.useTopK()) options["top_k"] = settings.topK(); if (settings.useFrequencyPenalty()) options["frequency_penalty"] = settings.frequencyPenalty(); if (settings.usePresencePenalty()) options["presence_penalty"] = settings.presencePenalty(); request["options"] = options; request["keep_alive"] = settings.ollamaLivetime(); }; if (type == LLMCore::RequestType::Fim) { applySettings(Settings::codeCompletionSettings()); } else { applySettings(Settings::chatAssistantSettings()); } } bool OllamaProvider::handleResponse(QNetworkReply *reply, QString &accumulatedResponse) { const QString endpoint = reply->url().path(); auto messageType = endpoint == completionEndpoint() ? OllamaMessage::Type::Generate : OllamaMessage::Type::Chat; auto processMessage = [&accumulatedResponse](const QJsonDocument &doc, OllamaMessage::Type messageType) { if (doc.isNull()) { LOG_MESSAGE("Invalid JSON response from Ollama"); return false; } auto message = OllamaMessage::fromJson(doc.object(), messageType); if (message.hasError()) { LOG_MESSAGE("Error in Ollama response: " + message.error); return false; } accumulatedResponse += message.getContent(); return message.done; }; if (reply->canReadLine()) { while (reply->canReadLine()) { QByteArray line = reply->readLine().trimmed(); if (line.isEmpty()) continue; if (processMessage(QJsonDocument::fromJson(line), messageType)) { return true; } } return false; } else { return processMessage(QJsonDocument::fromJson(reply->readAll()), messageType); } } QList OllamaProvider::getInstalledModels(const QString &url) { QList models; QNetworkAccessManager manager; QNetworkRequest request(QString("%1%2").arg(url, "/api/tags")); QNetworkReply *reply = manager.get(request); QEventLoop loop; QObject::connect(reply, &QNetworkReply::finished, &loop, &QEventLoop::quit); loop.exec(); if (reply->error() == QNetworkReply::NoError) { QByteArray responseData = reply->readAll(); QJsonDocument jsonResponse = QJsonDocument::fromJson(responseData); QJsonObject jsonObject = jsonResponse.object(); QJsonArray modelArray = jsonObject["models"].toArray(); for (const QJsonValue &value : modelArray) { QJsonObject modelObject = value.toObject(); QString modelName = modelObject["name"].toString(); models.append(modelName); } } else { LOG_MESSAGE(QString("Error fetching models: %1").arg(reply->errorString())); } reply->deleteLater(); return models; } } // namespace QodeAssist::Providers