/* * Copyright (C) 2024-2025 Petr Mironychev * * This file is part of QodeAssist. * * QodeAssist is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * QodeAssist is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with QodeAssist. If not, see . */ #include "OpenAIProvider.hpp" #include "llmcore/ValidationUtils.hpp" #include "logger/Logger.hpp" #include "settings/ChatAssistantSettings.hpp" #include "settings/CodeCompletionSettings.hpp" #include "settings/ProviderSettings.hpp" #include #include #include #include #include namespace QodeAssist::Providers { OpenAIProvider::OpenAIProvider(QObject *parent) : LLMCore::Provider(parent) , m_toolsManager(new Tools::ToolsManager(this)) { connect( m_toolsManager, &Tools::ToolsManager::toolExecutionComplete, this, &OpenAIProvider::onToolExecutionComplete); } QString OpenAIProvider::name() const { return "OpenAI"; } QString OpenAIProvider::url() const { return "https://api.openai.com"; } QString OpenAIProvider::completionEndpoint() const { return "/v1/chat/completions"; } QString OpenAIProvider::chatEndpoint() const { return "/v1/chat/completions"; } bool OpenAIProvider::supportsModelListing() const { return true; } void OpenAIProvider::prepareRequest( QJsonObject &request, LLMCore::PromptTemplate *prompt, LLMCore::ContextData context, LLMCore::RequestType type) { if (!prompt->isSupportProvider(providerID())) { LOG_MESSAGE(QString("Template %1 doesn't support %2 provider").arg(name(), prompt->name())); } prompt->prepareRequest(request, context); auto applyModelParams = [&request](const auto &settings) { request["max_tokens"] = settings.maxTokens(); request["temperature"] = settings.temperature(); if (settings.useTopP()) request["top_p"] = settings.topP(); if (settings.useTopK()) request["top_k"] = settings.topK(); if (settings.useFrequencyPenalty()) request["frequency_penalty"] = settings.frequencyPenalty(); if (settings.usePresencePenalty()) request["presence_penalty"] = settings.presencePenalty(); }; if (type == LLMCore::RequestType::CodeCompletion) { applyModelParams(Settings::codeCompletionSettings()); } else { applyModelParams(Settings::chatAssistantSettings()); } if (supportsTools() && type == LLMCore::RequestType::Chat && Settings::chatAssistantSettings().useTools()) { auto toolsDefinitions = m_toolsManager->getToolsDefinitions(Tools::ToolSchemaFormat::OpenAI); if (!toolsDefinitions.isEmpty()) { request["tools"] = toolsDefinitions; LOG_MESSAGE(QString("Added %1 tools to OpenAI request").arg(toolsDefinitions.size())); } } } QList OpenAIProvider::getInstalledModels(const QString &url) { QList models; QNetworkAccessManager manager; QNetworkRequest request(QString("%1/v1/models").arg(url)); request.setHeader(QNetworkRequest::ContentTypeHeader, "application/json"); if (!apiKey().isEmpty()) { request.setRawHeader("Authorization", QString("Bearer %1").arg(apiKey()).toUtf8()); } QNetworkReply *reply = manager.get(request); QEventLoop loop; QObject::connect(reply, &QNetworkReply::finished, &loop, &QEventLoop::quit); loop.exec(); if (reply->error() == QNetworkReply::NoError) { QByteArray responseData = reply->readAll(); QJsonDocument jsonResponse = QJsonDocument::fromJson(responseData); QJsonObject jsonObject = jsonResponse.object(); if (jsonObject.contains("data")) { QJsonArray modelArray = jsonObject["data"].toArray(); for (const QJsonValue &value : modelArray) { QJsonObject modelObject = value.toObject(); if (modelObject.contains("id")) { QString modelId = modelObject["id"].toString(); if (!modelId.contains("dall-e") && !modelId.contains("whisper") && !modelId.contains("tts") && !modelId.contains("davinci") && !modelId.contains("babbage") && !modelId.contains("omni")) { models.append(modelId); } } } } } else { LOG_MESSAGE(QString("Error fetching OpenAI models: %1").arg(reply->errorString())); } reply->deleteLater(); return models; } QList OpenAIProvider::validateRequest(const QJsonObject &request, LLMCore::TemplateType type) { const auto templateReq = QJsonObject{ {"model", {}}, {"messages", QJsonArray{{QJsonObject{{"role", {}}, {"content", {}}}}}}, {"temperature", {}}, {"max_tokens", {}}, {"top_p", {}}, {"top_k", {}}, {"frequency_penalty", {}}, {"presence_penalty", {}}, {"stop", QJsonArray{}}, {"stream", {}}, {"tools", {}}}; return LLMCore::ValidationUtils::validateRequestFields(request, templateReq); } QString OpenAIProvider::apiKey() const { return Settings::providerSettings().openAiApiKey(); } void OpenAIProvider::prepareNetworkRequest(QNetworkRequest &networkRequest) const { networkRequest.setHeader(QNetworkRequest::ContentTypeHeader, "application/json"); if (!apiKey().isEmpty()) { networkRequest.setRawHeader("Authorization", QString("Bearer %1").arg(apiKey()).toUtf8()); } } LLMCore::ProviderID OpenAIProvider::providerID() const { return LLMCore::ProviderID::OpenAI; } void OpenAIProvider::sendRequest( const LLMCore::RequestID &requestId, const QUrl &url, const QJsonObject &payload) { if (!m_messages.contains(requestId)) { m_dataBuffers[requestId].clear(); } m_requestUrls[requestId] = url; m_originalRequests[requestId] = payload; QNetworkRequest networkRequest(url); prepareNetworkRequest(networkRequest); LLMCore::HttpRequest request{.networkRequest = networkRequest, .requestId = requestId, .payload = payload}; LOG_MESSAGE(QString("OpenAIProvider: Sending request %1 to %2").arg(requestId, url.toString())); emit httpClient()->sendRequest(request); } bool OpenAIProvider::supportsTools() const { return true; } void OpenAIProvider::cancelRequest(const LLMCore::RequestID &requestId) { LOG_MESSAGE(QString("OpenAIProvider: Cancelling request %1").arg(requestId)); LLMCore::Provider::cancelRequest(requestId); cleanupRequest(requestId); } void OpenAIProvider::onDataReceived( const QodeAssist::LLMCore::RequestID &requestId, const QByteArray &data) { LLMCore::DataBuffers &buffers = m_dataBuffers[requestId]; QStringList lines = buffers.rawStreamBuffer.processData(data); for (const QString &line : lines) { if (line.trimmed().isEmpty() || line == "data: [DONE]") { continue; } QJsonObject chunk = parseEventLine(line); if (chunk.isEmpty()) continue; processStreamChunk(requestId, chunk); } } void OpenAIProvider::onRequestFinished( const QodeAssist::LLMCore::RequestID &requestId, bool success, const QString &error) { if (!success) { LOG_MESSAGE(QString("OpenAIProvider request %1 failed: %2").arg(requestId, error)); emit requestFailed(requestId, error); cleanupRequest(requestId); return; } if (m_messages.contains(requestId)) { OpenAIMessage *message = m_messages[requestId]; if (message->state() == LLMCore::MessageState::RequiresToolExecution) { LOG_MESSAGE(QString("Waiting for tools to complete for %1").arg(requestId)); m_dataBuffers.remove(requestId); return; } } if (m_dataBuffers.contains(requestId)) { const LLMCore::DataBuffers &buffers = m_dataBuffers[requestId]; if (!buffers.responseContent.isEmpty()) { LOG_MESSAGE(QString("Emitting full response for %1").arg(requestId)); emit fullResponseReceived(requestId, buffers.responseContent); } } cleanupRequest(requestId); } void OpenAIProvider::onToolExecutionComplete( const QString &requestId, const QHash &toolResults) { if (!m_messages.contains(requestId) || !m_requestUrls.contains(requestId)) { LOG_MESSAGE(QString("ERROR: Missing data for continuation request %1").arg(requestId)); cleanupRequest(requestId); return; } LOG_MESSAGE(QString("Tool execution complete for OpenAI request %1").arg(requestId)); OpenAIMessage *message = m_messages[requestId]; QJsonObject continuationRequest = m_originalRequests[requestId]; QJsonArray messages = continuationRequest["messages"].toArray(); messages.append(message->toProviderFormat()); QJsonArray toolResultMessages = message->createToolResultMessages(toolResults); for (const auto &toolMsg : toolResultMessages) { messages.append(toolMsg); } continuationRequest["messages"] = messages; LOG_MESSAGE(QString("Sending continuation request for %1 with %2 tool results") .arg(requestId) .arg(toolResults.size())); sendRequest(requestId, m_requestUrls[requestId], continuationRequest); } void OpenAIProvider::processStreamChunk(const QString &requestId, const QJsonObject &chunk) { QJsonArray choices = chunk["choices"].toArray(); if (choices.isEmpty()) { return; } QJsonObject choice = choices[0].toObject(); QJsonObject delta = choice["delta"].toObject(); QString finishReason = choice["finish_reason"].toString(); OpenAIMessage *message = m_messages.value(requestId); if (!message) { message = new OpenAIMessage(this); m_messages[requestId] = message; LOG_MESSAGE(QString("Created NEW OpenAIAPIMessage for request %1").arg(requestId)); } if (delta.contains("content") && !delta["content"].isNull()) { QString content = delta["content"].toString(); message->handleContentDelta(content); LLMCore::DataBuffers &buffers = m_dataBuffers[requestId]; buffers.responseContent += content; emit partialResponseReceived(requestId, content); } if (delta.contains("tool_calls")) { QJsonArray toolCalls = delta["tool_calls"].toArray(); for (const auto &toolCallValue : toolCalls) { QJsonObject toolCall = toolCallValue.toObject(); int index = toolCall["index"].toInt(); if (toolCall.contains("id")) { QString id = toolCall["id"].toString(); QJsonObject function = toolCall["function"].toObject(); QString name = function["name"].toString(); message->handleToolCallStart(index, id, name); } if (toolCall.contains("function")) { QJsonObject function = toolCall["function"].toObject(); if (function.contains("arguments")) { QString args = function["arguments"].toString(); message->handleToolCallDelta(index, args); } } } } if (!finishReason.isEmpty() && finishReason != "null") { for (int i = 0; i < 10; ++i) { message->handleToolCallComplete(i); } message->handleFinishReason(finishReason); handleMessageComplete(requestId); } } void OpenAIProvider::handleMessageComplete(const QString &requestId) { if (!m_messages.contains(requestId)) return; OpenAIMessage *message = m_messages[requestId]; if (message->state() == LLMCore::MessageState::RequiresToolExecution) { LOG_MESSAGE(QString("OpenAI message requires tool execution for %1").arg(requestId)); auto toolUseContent = message->getCurrentToolUseContent(); if (toolUseContent.isEmpty()) { LOG_MESSAGE(QString("No tools to execute for %1").arg(requestId)); return; } for (auto toolContent : toolUseContent) { m_toolsManager->executeToolCall( requestId, toolContent->id(), toolContent->name(), toolContent->input()); } } else { LOG_MESSAGE(QString("OpenAI message marked as complete for %1").arg(requestId)); } } void OpenAIProvider::cleanupRequest(const LLMCore::RequestID &requestId) { LOG_MESSAGE(QString("Cleaning up OpenAI request %1").arg(requestId)); if (m_messages.contains(requestId)) { OpenAIMessage *message = m_messages.take(requestId); message->deleteLater(); } m_dataBuffers.remove(requestId); m_requestUrls.remove(requestId); m_originalRequests.remove(requestId); m_toolsManager->cleanupRequest(requestId); } } // namespace QodeAssist::Providers