Data (or broader context) required for post processing of a camera request is saved via Camera3RequestDescriptor. Instead of passing individual arguments to CameraStream::process(), pass the Camera3RequestDescriptor pointer to it. All the arguments necessary to run the post-processor can be accessed from the descriptor. Signed-off-by: Umang Jain <umang.jain@ideasonboard.com> Reviewed-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com> Reviewed-by: Jacopo Mondi <jacopo@jmondi.org> Reviewed-by: Hirokazu Honda <hiroh@chromium.org>
216 lines
6.3 KiB
C++
216 lines
6.3 KiB
C++
/* SPDX-License-Identifier: LGPL-2.1-or-later */
|
|
/*
|
|
* Copyright (C) 2020, Google Inc.
|
|
*
|
|
* post_processor_jpeg.cpp - JPEG Post Processor
|
|
*/
|
|
|
|
#include "post_processor_jpeg.h"
|
|
|
|
#include <chrono>
|
|
|
|
#include "../camera_device.h"
|
|
#include "../camera_metadata.h"
|
|
#include "../camera_request.h"
|
|
#include "encoder_libjpeg.h"
|
|
#include "exif.h"
|
|
|
|
#include <libcamera/base/log.h>
|
|
|
|
#include <libcamera/formats.h>
|
|
|
|
using namespace libcamera;
|
|
using namespace std::chrono_literals;
|
|
|
|
LOG_DEFINE_CATEGORY(JPEG)
|
|
|
|
PostProcessorJpeg::PostProcessorJpeg(CameraDevice *const device)
|
|
: cameraDevice_(device)
|
|
{
|
|
}
|
|
|
|
int PostProcessorJpeg::configure(const StreamConfiguration &inCfg,
|
|
const StreamConfiguration &outCfg)
|
|
{
|
|
if (inCfg.size != outCfg.size) {
|
|
LOG(JPEG, Error) << "Mismatch of input and output stream sizes";
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (outCfg.pixelFormat != formats::MJPEG) {
|
|
LOG(JPEG, Error) << "Output stream pixel format is not JPEG";
|
|
return -EINVAL;
|
|
}
|
|
|
|
streamSize_ = outCfg.size;
|
|
|
|
thumbnailer_.configure(inCfg.size, inCfg.pixelFormat);
|
|
|
|
encoder_ = std::make_unique<EncoderLibJpeg>();
|
|
|
|
return encoder_->configure(inCfg);
|
|
}
|
|
|
|
void PostProcessorJpeg::generateThumbnail(const FrameBuffer &source,
|
|
const Size &targetSize,
|
|
unsigned int quality,
|
|
std::vector<unsigned char> *thumbnail)
|
|
{
|
|
/* Stores the raw scaled-down thumbnail bytes. */
|
|
std::vector<unsigned char> rawThumbnail;
|
|
|
|
thumbnailer_.createThumbnail(source, targetSize, &rawThumbnail);
|
|
|
|
StreamConfiguration thCfg;
|
|
thCfg.size = targetSize;
|
|
thCfg.pixelFormat = thumbnailer_.pixelFormat();
|
|
int ret = thumbnailEncoder_.configure(thCfg);
|
|
|
|
if (!rawThumbnail.empty() && !ret) {
|
|
/*
|
|
* \todo Avoid value-initialization of all elements of the
|
|
* vector.
|
|
*/
|
|
thumbnail->resize(rawThumbnail.size());
|
|
|
|
/*
|
|
* Split planes manually as the encoder expects a vector of
|
|
* planes.
|
|
*
|
|
* \todo Pass a vector of planes directly to
|
|
* Thumbnailer::createThumbnailer above and remove the manual
|
|
* planes split from here.
|
|
*/
|
|
std::vector<Span<uint8_t>> thumbnailPlanes;
|
|
const PixelFormatInfo &formatNV12 = PixelFormatInfo::info(formats::NV12);
|
|
size_t yPlaneSize = formatNV12.planeSize(targetSize, 0);
|
|
size_t uvPlaneSize = formatNV12.planeSize(targetSize, 1);
|
|
thumbnailPlanes.push_back({ rawThumbnail.data(), yPlaneSize });
|
|
thumbnailPlanes.push_back({ rawThumbnail.data() + yPlaneSize, uvPlaneSize });
|
|
|
|
int jpeg_size = thumbnailEncoder_.encode(thumbnailPlanes,
|
|
*thumbnail, {}, quality);
|
|
thumbnail->resize(jpeg_size);
|
|
|
|
LOG(JPEG, Debug)
|
|
<< "Thumbnail compress returned "
|
|
<< jpeg_size << " bytes";
|
|
}
|
|
}
|
|
|
|
int PostProcessorJpeg::process(const FrameBuffer &source,
|
|
CameraBuffer *destination,
|
|
Camera3RequestDescriptor *request)
|
|
{
|
|
if (!encoder_)
|
|
return 0;
|
|
|
|
ASSERT(destination->numPlanes() == 1);
|
|
|
|
const CameraMetadata &requestMetadata = request->settings_;
|
|
CameraMetadata *resultMetadata = request->resultMetadata_.get();
|
|
camera_metadata_ro_entry_t entry;
|
|
int ret;
|
|
|
|
/* Set EXIF metadata for various tags. */
|
|
Exif exif;
|
|
exif.setMake(cameraDevice_->maker());
|
|
exif.setModel(cameraDevice_->model());
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_ORIENTATION, &entry);
|
|
|
|
const uint32_t jpegOrientation = ret ? *entry.data.i32 : 0;
|
|
resultMetadata->addEntry(ANDROID_JPEG_ORIENTATION, jpegOrientation);
|
|
exif.setOrientation(jpegOrientation);
|
|
|
|
exif.setSize(streamSize_);
|
|
/*
|
|
* We set the frame's EXIF timestamp as the time of encode.
|
|
* Since the precision we need for EXIF timestamp is only one
|
|
* second, it is good enough.
|
|
*/
|
|
exif.setTimestamp(std::time(nullptr), 0ms);
|
|
|
|
ret = resultMetadata->getEntry(ANDROID_SENSOR_EXPOSURE_TIME, &entry);
|
|
exif.setExposureTime(ret ? *entry.data.i64 : 0);
|
|
ret = requestMetadata.getEntry(ANDROID_LENS_APERTURE, &entry);
|
|
if (ret)
|
|
exif.setAperture(*entry.data.f);
|
|
|
|
ret = resultMetadata->getEntry(ANDROID_SENSOR_SENSITIVITY, &entry);
|
|
exif.setISO(ret ? *entry.data.i32 : 100);
|
|
|
|
exif.setFlash(Exif::Flash::FlashNotPresent);
|
|
exif.setWhiteBalance(Exif::WhiteBalance::Auto);
|
|
|
|
exif.setFocalLength(1.0);
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_TIMESTAMP, &entry);
|
|
if (ret) {
|
|
exif.setGPSDateTimestamp(*entry.data.i64);
|
|
resultMetadata->addEntry(ANDROID_JPEG_GPS_TIMESTAMP,
|
|
*entry.data.i64);
|
|
}
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_THUMBNAIL_SIZE, &entry);
|
|
if (ret) {
|
|
const int32_t *data = entry.data.i32;
|
|
Size thumbnailSize = { static_cast<uint32_t>(data[0]),
|
|
static_cast<uint32_t>(data[1]) };
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_THUMBNAIL_QUALITY, &entry);
|
|
uint8_t quality = ret ? *entry.data.u8 : 95;
|
|
resultMetadata->addEntry(ANDROID_JPEG_THUMBNAIL_QUALITY, quality);
|
|
|
|
if (thumbnailSize != Size(0, 0)) {
|
|
std::vector<unsigned char> thumbnail;
|
|
generateThumbnail(source, thumbnailSize, quality, &thumbnail);
|
|
if (!thumbnail.empty())
|
|
exif.setThumbnail(thumbnail, Exif::Compression::JPEG);
|
|
}
|
|
|
|
resultMetadata->addEntry(ANDROID_JPEG_THUMBNAIL_SIZE, data, 2);
|
|
}
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_COORDINATES, &entry);
|
|
if (ret) {
|
|
exif.setGPSLocation(entry.data.d);
|
|
resultMetadata->addEntry(ANDROID_JPEG_GPS_COORDINATES,
|
|
entry.data.d, 3);
|
|
}
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_PROCESSING_METHOD, &entry);
|
|
if (ret) {
|
|
std::string method(entry.data.u8, entry.data.u8 + entry.count);
|
|
exif.setGPSMethod(method);
|
|
resultMetadata->addEntry(ANDROID_JPEG_GPS_PROCESSING_METHOD,
|
|
entry.data.u8, entry.count);
|
|
}
|
|
|
|
if (exif.generate() != 0)
|
|
LOG(JPEG, Error) << "Failed to generate valid EXIF data";
|
|
|
|
ret = requestMetadata.getEntry(ANDROID_JPEG_QUALITY, &entry);
|
|
const uint8_t quality = ret ? *entry.data.u8 : 95;
|
|
resultMetadata->addEntry(ANDROID_JPEG_QUALITY, quality);
|
|
|
|
int jpeg_size = encoder_->encode(source, destination->plane(0),
|
|
exif.data(), quality);
|
|
if (jpeg_size < 0) {
|
|
LOG(JPEG, Error) << "Failed to encode stream image";
|
|
return jpeg_size;
|
|
}
|
|
|
|
/* Fill in the JPEG blob header. */
|
|
uint8_t *resultPtr = destination->plane(0).data()
|
|
+ destination->jpegBufferSize(cameraDevice_->maxJpegBufferSize())
|
|
- sizeof(struct camera3_jpeg_blob);
|
|
auto *blob = reinterpret_cast<struct camera3_jpeg_blob *>(resultPtr);
|
|
blob->jpeg_blob_id = CAMERA3_JPEG_BLOB_ID;
|
|
blob->jpeg_size = jpeg_size;
|
|
|
|
/* Update the JPEG result Metadata. */
|
|
resultMetadata->addEntry(ANDROID_JPEG_SIZE, jpeg_size);
|
|
|
|
return 0;
|
|
}
|