mirror of
https://invent.kde.org/frameworks/kimageformats.git
synced 2025-07-18 03:54:18 -04:00
Run clang-format on all cpp/h files
NO_CHANGELOG
This commit is contained in:
@ -6,23 +6,22 @@
|
||||
SPDX-License-Identifier: BSD-2-Clause
|
||||
*/
|
||||
|
||||
#include <QtGlobal>
|
||||
#include <QThread>
|
||||
#include <QtGlobal>
|
||||
|
||||
#include <QColorSpace>
|
||||
|
||||
#include "avif_p.h"
|
||||
#include <cfloat>
|
||||
|
||||
|
||||
QAVIFHandler::QAVIFHandler() :
|
||||
m_parseState(ParseAvifNotParsed),
|
||||
m_quality(52),
|
||||
m_container_width(0),
|
||||
m_container_height(0),
|
||||
m_rawAvifData(AVIF_DATA_EMPTY),
|
||||
m_decoder(nullptr),
|
||||
m_must_jump_to_next_image(false)
|
||||
QAVIFHandler::QAVIFHandler()
|
||||
: m_parseState(ParseAvifNotParsed)
|
||||
, m_quality(52)
|
||||
, m_container_width(0)
|
||||
, m_container_height(0)
|
||||
, m_rawAvifData(AVIF_DATA_EMPTY)
|
||||
, m_decoder(nullptr)
|
||||
, m_must_jump_to_next_image(false)
|
||||
{
|
||||
}
|
||||
|
||||
@ -57,7 +56,7 @@ bool QAVIFHandler::canRead(QIODevice *device)
|
||||
}
|
||||
|
||||
avifROData input;
|
||||
input.data = (const uint8_t *) header.constData();
|
||||
input.data = (const uint8_t *)header.constData();
|
||||
input.size = header.size();
|
||||
|
||||
if (avifPeekCompatibleFileType(&input)) {
|
||||
@ -88,7 +87,7 @@ bool QAVIFHandler::ensureDecoder()
|
||||
|
||||
m_rawData = device()->readAll();
|
||||
|
||||
m_rawAvifData.data = (const uint8_t *) m_rawData.constData();
|
||||
m_rawAvifData.data = (const uint8_t *)m_rawData.constData();
|
||||
m_rawAvifData.size = m_rawData.size();
|
||||
|
||||
if (avifPeekCompatibleFileType(&m_rawAvifData) == AVIF_FALSE) {
|
||||
@ -96,7 +95,6 @@ bool QAVIFHandler::ensureDecoder()
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
m_decoder = avifDecoderCreate();
|
||||
|
||||
avifResult decodeResult;
|
||||
@ -124,7 +122,6 @@ bool QAVIFHandler::ensureDecoder()
|
||||
decodeResult = avifDecoderNextImage(m_decoder);
|
||||
|
||||
if (decodeResult == AVIF_RESULT_OK) {
|
||||
|
||||
m_container_width = m_decoder->image->width;
|
||||
m_container_height = m_decoder->image->height;
|
||||
|
||||
@ -194,12 +191,12 @@ bool QAVIFHandler::decode_one_frame()
|
||||
}
|
||||
|
||||
if (m_decoder->image->icc.data && (m_decoder->image->icc.size > 0)) {
|
||||
result.setColorSpace(QColorSpace::fromIccProfile(QByteArray::fromRawData((const char *) m_decoder->image->icc.data, (int) m_decoder->image->icc.size)));
|
||||
if (! result.colorSpace().isValid()) {
|
||||
result.setColorSpace(QColorSpace::fromIccProfile(QByteArray::fromRawData((const char *)m_decoder->image->icc.data, (int)m_decoder->image->icc.size)));
|
||||
if (!result.colorSpace().isValid()) {
|
||||
qWarning("Invalid QColorSpace created from ICC!");
|
||||
}
|
||||
} else {
|
||||
float prim[8] = { 0.64f, 0.33f, 0.3f, 0.6f, 0.15f, 0.06f, 0.3127f, 0.329f };
|
||||
float prim[8] = {0.64f, 0.33f, 0.3f, 0.6f, 0.15f, 0.06f, 0.3127f, 0.329f};
|
||||
// outPrimaries: rX, rY, gX, gY, bX, bY, wX, wY
|
||||
avifColorPrimariesGetValues(m_decoder->image->colorPrimaries, prim);
|
||||
|
||||
@ -208,7 +205,6 @@ bool QAVIFHandler::decode_one_frame()
|
||||
const QPointF bluePoint(QAVIFHandler::CompatibleChromacity(prim[4], prim[5]));
|
||||
const QPointF whitePoint(QAVIFHandler::CompatibleChromacity(prim[6], prim[7]));
|
||||
|
||||
|
||||
QColorSpace::TransferFunction q_trc = QColorSpace::TransferFunction::Custom;
|
||||
float q_trc_gamma = 0.0f;
|
||||
|
||||
@ -231,16 +227,17 @@ bool QAVIFHandler::decode_one_frame()
|
||||
case 0:
|
||||
case 2: /* AVIF_TRANSFER_CHARACTERISTICS_UNSPECIFIED */
|
||||
case 13:
|
||||
q_trc = QColorSpace::TransferFunction::SRgb;
|
||||
q_trc = QColorSpace::TransferFunction::SRgb;
|
||||
break;
|
||||
default:
|
||||
qWarning("CICP colorPrimaries: %d, transferCharacteristics: %d\nThe colorspace is unsupported by this plug-in yet.",
|
||||
m_decoder->image->colorPrimaries, m_decoder->image->transferCharacteristics);
|
||||
m_decoder->image->colorPrimaries,
|
||||
m_decoder->image->transferCharacteristics);
|
||||
q_trc = QColorSpace::TransferFunction::SRgb;
|
||||
break;
|
||||
}
|
||||
|
||||
if (q_trc != QColorSpace::TransferFunction::Custom) { //we create new colorspace using Qt
|
||||
if (q_trc != QColorSpace::TransferFunction::Custom) { // we create new colorspace using Qt
|
||||
switch (m_decoder->image->colorPrimaries) {
|
||||
/* AVIF_COLOR_PRIMARIES_BT709 */
|
||||
case 0:
|
||||
@ -258,7 +255,7 @@ bool QAVIFHandler::decode_one_frame()
|
||||
}
|
||||
}
|
||||
|
||||
if (! result.colorSpace().isValid()) {
|
||||
if (!result.colorSpace().isValid()) {
|
||||
qWarning("Invalid QColorSpace created from NCLX/CICP!");
|
||||
}
|
||||
}
|
||||
@ -303,11 +300,11 @@ bool QAVIFHandler::decode_one_frame()
|
||||
}
|
||||
|
||||
if (m_decoder->image->transformFlags & AVIF_TRANSFORM_CLAP) {
|
||||
if ((m_decoder->image->clap.widthD > 0) && (m_decoder->image->clap.heightD > 0) &&
|
||||
(m_decoder->image->clap.horizOffD > 0) && (m_decoder->image->clap.vertOffD > 0)) {
|
||||
int new_width, new_height, offx, offy;
|
||||
if ((m_decoder->image->clap.widthD > 0) && (m_decoder->image->clap.heightD > 0) && (m_decoder->image->clap.horizOffD > 0)
|
||||
&& (m_decoder->image->clap.vertOffD > 0)) {
|
||||
int new_width, new_height, offx, offy;
|
||||
|
||||
new_width = (int)((double)(m_decoder->image->clap.widthN) / (m_decoder->image->clap.widthD) + 0.5);
|
||||
new_width = (int)((double)(m_decoder->image->clap.widthN) / (m_decoder->image->clap.widthD) + 0.5);
|
||||
if (new_width > result.width()) {
|
||||
new_width = result.width();
|
||||
}
|
||||
@ -318,17 +315,14 @@ bool QAVIFHandler::decode_one_frame()
|
||||
}
|
||||
|
||||
if (new_width > 0 && new_height > 0) {
|
||||
|
||||
offx = ((double)((int32_t) m_decoder->image->clap.horizOffN)) / (m_decoder->image->clap.horizOffD) +
|
||||
(result.width() - new_width) / 2.0 + 0.5;
|
||||
offx = ((double)((int32_t)m_decoder->image->clap.horizOffN)) / (m_decoder->image->clap.horizOffD) + (result.width() - new_width) / 2.0 + 0.5;
|
||||
if (offx < 0) {
|
||||
offx = 0;
|
||||
} else if (offx > (result.width() - new_width)) {
|
||||
offx = result.width() - new_width;
|
||||
}
|
||||
|
||||
offy = ((double)((int32_t) m_decoder->image->clap.vertOffN)) / (m_decoder->image->clap.vertOffD) +
|
||||
(result.height() - new_height) / 2.0 + 0.5;
|
||||
offy = ((double)((int32_t)m_decoder->image->clap.vertOffN)) / (m_decoder->image->clap.vertOffD) + (result.height() - new_height) / 2.0 + 0.5;
|
||||
if (offy < 0) {
|
||||
offy = 0;
|
||||
} else if (offy > (result.height() - new_height)) {
|
||||
@ -339,7 +333,7 @@ bool QAVIFHandler::decode_one_frame()
|
||||
}
|
||||
}
|
||||
|
||||
else { //Zero values, we need to avoid 0 divide.
|
||||
else { // Zero values, we need to avoid 0 divide.
|
||||
qWarning("ERROR: Wrong values in avifCleanApertureBox");
|
||||
}
|
||||
}
|
||||
@ -364,10 +358,10 @@ bool QAVIFHandler::decode_one_frame()
|
||||
|
||||
if (m_decoder->image->transformFlags & AVIF_TRANSFORM_IMIR) {
|
||||
switch (m_decoder->image->imir.axis) {
|
||||
case 0: //vertical
|
||||
case 0: // vertical
|
||||
result = result.mirrored(false, true);
|
||||
break;
|
||||
case 1: //horizontal
|
||||
case 1: // horizontal
|
||||
result = result.mirrored(true, false);
|
||||
break;
|
||||
}
|
||||
@ -417,13 +411,13 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
int maxQuantizerAlpha = 0;
|
||||
avifResult res;
|
||||
|
||||
bool save_grayscale; //true - monochrome, false - colors
|
||||
int save_depth; //8 or 10bit per channel
|
||||
QImage::Format tmpformat; //format for temporary image
|
||||
bool save_grayscale; // true - monochrome, false - colors
|
||||
int save_depth; // 8 or 10bit per channel
|
||||
QImage::Format tmpformat; // format for temporary image
|
||||
|
||||
avifImage *avif = nullptr;
|
||||
|
||||
//grayscale detection
|
||||
// grayscale detection
|
||||
switch (image.format()) {
|
||||
case QImage::Format_Mono:
|
||||
case QImage::Format_MonoLSB:
|
||||
@ -439,7 +433,7 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
break;
|
||||
}
|
||||
|
||||
//depth detection
|
||||
// depth detection
|
||||
switch (image.format()) {
|
||||
case QImage::Format_BGR30:
|
||||
case QImage::Format_A2BGR30_Premultiplied:
|
||||
@ -460,15 +454,15 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
break;
|
||||
}
|
||||
|
||||
//quality settings
|
||||
// quality settings
|
||||
if (maxQuantizer > 20) {
|
||||
minQuantizer = maxQuantizer - 20;
|
||||
if (maxQuantizer > 40) { //we decrease quality of alpha channel here
|
||||
if (maxQuantizer > 40) { // we decrease quality of alpha channel here
|
||||
maxQuantizerAlpha = maxQuantizer - 40;
|
||||
}
|
||||
}
|
||||
|
||||
if (save_grayscale && !image.hasAlphaChannel()) { //we are going to save grayscale image without alpha channel
|
||||
if (save_grayscale && !image.hasAlphaChannel()) { // we are going to save grayscale image without alpha channel
|
||||
if (save_depth > 8) {
|
||||
tmpformat = QImage::Format_Grayscale16;
|
||||
} else {
|
||||
@ -496,7 +490,6 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
/* AVIF_TRANSFER_CHARACTERISTICS_UNSPECIFIED */
|
||||
break;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
if (save_depth > 8) { // QImage::Format_Grayscale16
|
||||
@ -504,7 +497,7 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
const uint16_t *src16bit = reinterpret_cast<const uint16_t *>(tmpgrayimage.constScanLine(y));
|
||||
uint16_t *dest16bit = reinterpret_cast<uint16_t *>(avif->yuvPlanes[0] + y * avif->yuvRowBytes[0]);
|
||||
for (int x = 0; x < tmpgrayimage.width(); x++) {
|
||||
int tmp_pixelval = (int)(((float)(*src16bit) / 65535.0f) * 1023.0f + 0.5f); //downgrade to 10 bits
|
||||
int tmp_pixelval = (int)(((float)(*src16bit) / 65535.0f) * 1023.0f + 0.5f); // downgrade to 10 bits
|
||||
*dest16bit = qBound(0, tmp_pixelval, 1023);
|
||||
dest16bit++;
|
||||
src16bit++;
|
||||
@ -522,14 +515,14 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
}
|
||||
}
|
||||
|
||||
} else { //we are going to save color image
|
||||
} else { // we are going to save color image
|
||||
if (save_depth > 8) {
|
||||
if (image.hasAlphaChannel()) {
|
||||
tmpformat = QImage::Format_RGBA64;
|
||||
} else {
|
||||
tmpformat = QImage::Format_RGBX64;
|
||||
}
|
||||
} else { //8bit depth
|
||||
} else { // 8bit depth
|
||||
if (image.hasAlphaChannel()) {
|
||||
tmpformat = QImage::Format_RGBA8888;
|
||||
} else {
|
||||
@ -542,13 +535,13 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
avifPixelFormat pixel_format = AVIF_PIXEL_FORMAT_YUV420;
|
||||
if (maxQuantizer < 20) {
|
||||
if (maxQuantizer < 10) {
|
||||
pixel_format = AVIF_PIXEL_FORMAT_YUV444; //best quality
|
||||
pixel_format = AVIF_PIXEL_FORMAT_YUV444; // best quality
|
||||
} else {
|
||||
pixel_format = AVIF_PIXEL_FORMAT_YUV422; //high quality
|
||||
pixel_format = AVIF_PIXEL_FORMAT_YUV422; // high quality
|
||||
}
|
||||
}
|
||||
|
||||
avifMatrixCoefficients matrix_to_save = (avifMatrixCoefficients)1; //default for Qt 5.12 and 5.13;
|
||||
avifMatrixCoefficients matrix_to_save = (avifMatrixCoefficients)1; // default for Qt 5.12 and 5.13;
|
||||
|
||||
avifColorPrimaries primaries_to_save = (avifColorPrimaries)2;
|
||||
avifTransferCharacteristics transfer_to_save = (avifTransferCharacteristics)2;
|
||||
@ -602,11 +595,9 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
break;
|
||||
}
|
||||
|
||||
//in case primaries or trc were not identified
|
||||
if ((primaries_to_save == 2) ||
|
||||
(transfer_to_save == 2)) {
|
||||
|
||||
//upgrade image to higher bit depth
|
||||
// in case primaries or trc were not identified
|
||||
if ((primaries_to_save == 2) || (transfer_to_save == 2)) {
|
||||
// upgrade image to higher bit depth
|
||||
if (save_depth == 8) {
|
||||
save_depth = 10;
|
||||
if (tmpcolorimage.hasAlphaChannel()) {
|
||||
@ -616,8 +607,7 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
}
|
||||
}
|
||||
|
||||
if ((primaries_to_save == 2) &&
|
||||
(transfer_to_save != 2)) { //other primaries but known trc
|
||||
if ((primaries_to_save == 2) && (transfer_to_save != 2)) { // other primaries but known trc
|
||||
primaries_to_save = (avifColorPrimaries)1; // AVIF_COLOR_PRIMARIES_BT709
|
||||
matrix_to_save = (avifMatrixCoefficients)1; // AVIF_MATRIX_COEFFICIENTS_BT709
|
||||
|
||||
@ -636,11 +626,10 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
transfer_to_save = (avifTransferCharacteristics)13;
|
||||
break;
|
||||
}
|
||||
} else if ((primaries_to_save != 2) &&
|
||||
(transfer_to_save == 2)) { //recognized primaries but other trc
|
||||
} else if ((primaries_to_save != 2) && (transfer_to_save == 2)) { // recognized primaries but other trc
|
||||
transfer_to_save = (avifTransferCharacteristics)13;
|
||||
tmpcolorimage.convertToColorSpace(tmpcolorimage.colorSpace().withTransferFunction(QColorSpace::TransferFunction::SRgb));
|
||||
} else { //unrecognized profile
|
||||
} else { // unrecognized profile
|
||||
primaries_to_save = (avifColorPrimaries)1; // AVIF_COLOR_PRIMARIES_BT709
|
||||
transfer_to_save = (avifTransferCharacteristics)13;
|
||||
matrix_to_save = (avifMatrixCoefficients)1; // AVIF_MATRIX_COEFFICIENTS_BT709
|
||||
@ -659,7 +648,7 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
rgb.rowBytes = tmpcolorimage.bytesPerLine();
|
||||
rgb.pixels = const_cast<uint8_t *>(tmpcolorimage.constBits());
|
||||
|
||||
if (save_depth > 8) { //10bit depth
|
||||
if (save_depth > 8) { // 10bit depth
|
||||
rgb.depth = 16;
|
||||
|
||||
if (tmpcolorimage.hasAlphaChannel()) {
|
||||
@ -669,7 +658,7 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
}
|
||||
|
||||
rgb.format = AVIF_RGB_FORMAT_RGBA;
|
||||
} else { //8bit depth
|
||||
} else { // 8bit depth
|
||||
rgb.depth = 8;
|
||||
|
||||
if (tmpcolorimage.hasAlphaChannel()) {
|
||||
@ -721,7 +710,6 @@ bool QAVIFHandler::write(const QImage &image)
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
QVariant QAVIFHandler::option(ImageOption option) const
|
||||
{
|
||||
if (option == Quality) {
|
||||
@ -765,9 +753,7 @@ void QAVIFHandler::setOption(ImageOption option, const QVariant &value)
|
||||
|
||||
bool QAVIFHandler::supportsOption(ImageOption option) const
|
||||
{
|
||||
return option == Quality
|
||||
|| option == Size
|
||||
|| option == Animation;
|
||||
return option == Quality || option == Size || option == Animation;
|
||||
}
|
||||
|
||||
int QAVIFHandler::imageCount() const
|
||||
@ -805,7 +791,7 @@ bool QAVIFHandler::jumpToNextImage()
|
||||
return true;
|
||||
}
|
||||
|
||||
if (m_decoder->imageIndex >= m_decoder->imageCount - 1) { //start from begining
|
||||
if (m_decoder->imageIndex >= m_decoder->imageCount - 1) { // start from begining
|
||||
avifDecoderReset(m_decoder);
|
||||
}
|
||||
|
||||
@ -817,11 +803,12 @@ bool QAVIFHandler::jumpToNextImage()
|
||||
return false;
|
||||
}
|
||||
|
||||
if ((m_container_width != m_decoder->image->width) ||
|
||||
(m_container_height != m_decoder->image->height)) {
|
||||
if ((m_container_width != m_decoder->image->width) || (m_container_height != m_decoder->image->height)) {
|
||||
qWarning("Decoded image sequence size (%dx%d) do not match first image size (%dx%d)!",
|
||||
m_decoder->image->width, m_decoder->image->height,
|
||||
m_container_width, m_container_height);
|
||||
m_decoder->image->width,
|
||||
m_decoder->image->height,
|
||||
m_container_width,
|
||||
m_container_height);
|
||||
|
||||
m_parseState = ParseAvifError;
|
||||
return false;
|
||||
@ -833,7 +820,6 @@ bool QAVIFHandler::jumpToNextImage()
|
||||
m_parseState = ParseAvifError;
|
||||
return false;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
bool QAVIFHandler::jumpToImage(int imageNumber)
|
||||
@ -842,7 +828,7 @@ bool QAVIFHandler::jumpToImage(int imageNumber)
|
||||
return false;
|
||||
}
|
||||
|
||||
if (m_decoder->imageCount < 2) { //not an animation
|
||||
if (m_decoder->imageCount < 2) { // not an animation
|
||||
if (imageNumber == 0) {
|
||||
return true;
|
||||
} else {
|
||||
@ -850,11 +836,11 @@ bool QAVIFHandler::jumpToImage(int imageNumber)
|
||||
}
|
||||
}
|
||||
|
||||
if (imageNumber < 0 || imageNumber >= m_decoder->imageCount) { //wrong index
|
||||
if (imageNumber < 0 || imageNumber >= m_decoder->imageCount) { // wrong index
|
||||
return false;
|
||||
}
|
||||
|
||||
if (imageNumber == m_decoder->imageCount) { // we are here already
|
||||
if (imageNumber == m_decoder->imageCount) { // we are here already
|
||||
return true;
|
||||
}
|
||||
|
||||
@ -866,11 +852,12 @@ bool QAVIFHandler::jumpToImage(int imageNumber)
|
||||
return false;
|
||||
}
|
||||
|
||||
if ((m_container_width != m_decoder->image->width) ||
|
||||
(m_container_height != m_decoder->image->height)) {
|
||||
if ((m_container_width != m_decoder->image->width) || (m_container_height != m_decoder->image->height)) {
|
||||
qWarning("Decoded image sequence size (%dx%d) do not match declared container size (%dx%d)!",
|
||||
m_decoder->image->width, m_decoder->image->height,
|
||||
m_container_width, m_container_height);
|
||||
m_decoder->image->width,
|
||||
m_decoder->image->height,
|
||||
m_container_width,
|
||||
m_container_height);
|
||||
|
||||
m_parseState = ParseAvifError;
|
||||
return false;
|
||||
|
Reference in New Issue
Block a user