libcamera/src/android/jpeg/post_processor_jpeg.cpp
Umang Jain 1976179623 android: camera_stream: Plumb process() with Camera3RequestDescriptor
Data (or broader context) required for post processing of a camera request
is saved via Camera3RequestDescriptor. Instead of passing individual
arguments to CameraStream::process(), pass the Camera3RequestDescriptor
pointer to it. All the arguments necessary to run the post-processor can
be accessed from the descriptor.

Signed-off-by: Umang Jain <umang.jain@ideasonboard.com>
Reviewed-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Reviewed-by: Jacopo Mondi <jacopo@jmondi.org>
Reviewed-by: Hirokazu Honda <hiroh@chromium.org>
2021-10-19 19:15:27 +05:30

216 lines
6.3 KiB
C++

/* SPDX-License-Identifier: LGPL-2.1-or-later */
/*
* Copyright (C) 2020, Google Inc.
*
* post_processor_jpeg.cpp - JPEG Post Processor
*/
#include "post_processor_jpeg.h"
#include <chrono>
#include "../camera_device.h"
#include "../camera_metadata.h"
#include "../camera_request.h"
#include "encoder_libjpeg.h"
#include "exif.h"
#include <libcamera/base/log.h>
#include <libcamera/formats.h>
using namespace libcamera;
using namespace std::chrono_literals;
LOG_DEFINE_CATEGORY(JPEG)
PostProcessorJpeg::PostProcessorJpeg(CameraDevice *const device)
: cameraDevice_(device)
{
}
int PostProcessorJpeg::configure(const StreamConfiguration &inCfg,
const StreamConfiguration &outCfg)
{
if (inCfg.size != outCfg.size) {
LOG(JPEG, Error) << "Mismatch of input and output stream sizes";
return -EINVAL;
}
if (outCfg.pixelFormat != formats::MJPEG) {
LOG(JPEG, Error) << "Output stream pixel format is not JPEG";
return -EINVAL;
}
streamSize_ = outCfg.size;
thumbnailer_.configure(inCfg.size, inCfg.pixelFormat);
encoder_ = std::make_unique<EncoderLibJpeg>();
return encoder_->configure(inCfg);
}
void PostProcessorJpeg::generateThumbnail(const FrameBuffer &source,
const Size &targetSize,
unsigned int quality,
std::vector<unsigned char> *thumbnail)
{
/* Stores the raw scaled-down thumbnail bytes. */
std::vector<unsigned char> rawThumbnail;
thumbnailer_.createThumbnail(source, targetSize, &rawThumbnail);
StreamConfiguration thCfg;
thCfg.size = targetSize;
thCfg.pixelFormat = thumbnailer_.pixelFormat();
int ret = thumbnailEncoder_.configure(thCfg);
if (!rawThumbnail.empty() && !ret) {
/*
* \todo Avoid value-initialization of all elements of the
* vector.
*/
thumbnail->resize(rawThumbnail.size());
/*
* Split planes manually as the encoder expects a vector of
* planes.
*
* \todo Pass a vector of planes directly to
* Thumbnailer::createThumbnailer above and remove the manual
* planes split from here.
*/
std::vector<Span<uint8_t>> thumbnailPlanes;
const PixelFormatInfo &formatNV12 = PixelFormatInfo::info(formats::NV12);
size_t yPlaneSize = formatNV12.planeSize(targetSize, 0);
size_t uvPlaneSize = formatNV12.planeSize(targetSize, 1);
thumbnailPlanes.push_back({ rawThumbnail.data(), yPlaneSize });
thumbnailPlanes.push_back({ rawThumbnail.data() + yPlaneSize, uvPlaneSize });
int jpeg_size = thumbnailEncoder_.encode(thumbnailPlanes,
*thumbnail, {}, quality);
thumbnail->resize(jpeg_size);
LOG(JPEG, Debug)
<< "Thumbnail compress returned "
<< jpeg_size << " bytes";
}
}
int PostProcessorJpeg::process(const FrameBuffer &source,
CameraBuffer *destination,
Camera3RequestDescriptor *request)
{
if (!encoder_)
return 0;
ASSERT(destination->numPlanes() == 1);
const CameraMetadata &requestMetadata = request->settings_;
CameraMetadata *resultMetadata = request->resultMetadata_.get();
camera_metadata_ro_entry_t entry;
int ret;
/* Set EXIF metadata for various tags. */
Exif exif;
exif.setMake(cameraDevice_->maker());
exif.setModel(cameraDevice_->model());
ret = requestMetadata.getEntry(ANDROID_JPEG_ORIENTATION, &entry);
const uint32_t jpegOrientation = ret ? *entry.data.i32 : 0;
resultMetadata->addEntry(ANDROID_JPEG_ORIENTATION, jpegOrientation);
exif.setOrientation(jpegOrientation);
exif.setSize(streamSize_);
/*
* We set the frame's EXIF timestamp as the time of encode.
* Since the precision we need for EXIF timestamp is only one
* second, it is good enough.
*/
exif.setTimestamp(std::time(nullptr), 0ms);
ret = resultMetadata->getEntry(ANDROID_SENSOR_EXPOSURE_TIME, &entry);
exif.setExposureTime(ret ? *entry.data.i64 : 0);
ret = requestMetadata.getEntry(ANDROID_LENS_APERTURE, &entry);
if (ret)
exif.setAperture(*entry.data.f);
ret = resultMetadata->getEntry(ANDROID_SENSOR_SENSITIVITY, &entry);
exif.setISO(ret ? *entry.data.i32 : 100);
exif.setFlash(Exif::Flash::FlashNotPresent);
exif.setWhiteBalance(Exif::WhiteBalance::Auto);
exif.setFocalLength(1.0);
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_TIMESTAMP, &entry);
if (ret) {
exif.setGPSDateTimestamp(*entry.data.i64);
resultMetadata->addEntry(ANDROID_JPEG_GPS_TIMESTAMP,
*entry.data.i64);
}
ret = requestMetadata.getEntry(ANDROID_JPEG_THUMBNAIL_SIZE, &entry);
if (ret) {
const int32_t *data = entry.data.i32;
Size thumbnailSize = { static_cast<uint32_t>(data[0]),
static_cast<uint32_t>(data[1]) };
ret = requestMetadata.getEntry(ANDROID_JPEG_THUMBNAIL_QUALITY, &entry);
uint8_t quality = ret ? *entry.data.u8 : 95;
resultMetadata->addEntry(ANDROID_JPEG_THUMBNAIL_QUALITY, quality);
if (thumbnailSize != Size(0, 0)) {
std::vector<unsigned char> thumbnail;
generateThumbnail(source, thumbnailSize, quality, &thumbnail);
if (!thumbnail.empty())
exif.setThumbnail(thumbnail, Exif::Compression::JPEG);
}
resultMetadata->addEntry(ANDROID_JPEG_THUMBNAIL_SIZE, data, 2);
}
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_COORDINATES, &entry);
if (ret) {
exif.setGPSLocation(entry.data.d);
resultMetadata->addEntry(ANDROID_JPEG_GPS_COORDINATES,
entry.data.d, 3);
}
ret = requestMetadata.getEntry(ANDROID_JPEG_GPS_PROCESSING_METHOD, &entry);
if (ret) {
std::string method(entry.data.u8, entry.data.u8 + entry.count);
exif.setGPSMethod(method);
resultMetadata->addEntry(ANDROID_JPEG_GPS_PROCESSING_METHOD,
entry.data.u8, entry.count);
}
if (exif.generate() != 0)
LOG(JPEG, Error) << "Failed to generate valid EXIF data";
ret = requestMetadata.getEntry(ANDROID_JPEG_QUALITY, &entry);
const uint8_t quality = ret ? *entry.data.u8 : 95;
resultMetadata->addEntry(ANDROID_JPEG_QUALITY, quality);
int jpeg_size = encoder_->encode(source, destination->plane(0),
exif.data(), quality);
if (jpeg_size < 0) {
LOG(JPEG, Error) << "Failed to encode stream image";
return jpeg_size;
}
/* Fill in the JPEG blob header. */
uint8_t *resultPtr = destination->plane(0).data()
+ destination->jpegBufferSize(cameraDevice_->maxJpegBufferSize())
- sizeof(struct camera3_jpeg_blob);
auto *blob = reinterpret_cast<struct camera3_jpeg_blob *>(resultPtr);
blob->jpeg_blob_id = CAMERA3_JPEG_BLOB_ID;
blob->jpeg_size = jpeg_size;
/* Update the JPEG result Metadata. */
resultMetadata->addEntry(ANDROID_JPEG_SIZE, jpeg_size);
return 0;
}