libcamera/src/libcamera/pipeline/vimc.cpp
Laurent Pinchart a2dddf7c26 libcamera: Use the Size class through libcamera
Several of our structures include width and height fields that model a
size while we have a Size class for that purpose. Use the Size class
through libcamera, and give it a toString() method like other geometry
and format classes.

Signed-off-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com>
Reviewed-by: Niklas Söderlund <niklas.soderlund@ragnatech.se>
2019-04-30 22:18:42 +03:00

226 lines
5 KiB
C++

/* SPDX-License-Identifier: LGPL-2.1-or-later */
/*
* Copyright (C) 2018, Google Inc.
*
* vimc.cpp - Pipeline handler for the vimc device
*/
#include <libcamera/camera.h>
#include <libcamera/request.h>
#include <libcamera/stream.h>
#include "device_enumerator.h"
#include "log.h"
#include "media_device.h"
#include "pipeline_handler.h"
#include "utils.h"
#include "v4l2_device.h"
namespace libcamera {
LOG_DEFINE_CATEGORY(VIMC)
class PipelineHandlerVimc : public PipelineHandler
{
public:
PipelineHandlerVimc(CameraManager *manager);
~PipelineHandlerVimc();
CameraConfiguration
streamConfiguration(Camera *camera,
const std::vector<StreamUsage> &usages) override;
int configureStreams(Camera *camera,
const CameraConfiguration &config) override;
int allocateBuffers(Camera *camera,
const std::set<Stream *> &streams) override;
int freeBuffers(Camera *camera,
const std::set<Stream *> &streams) override;
int start(Camera *camera) override;
void stop(Camera *camera) override;
int queueRequest(Camera *camera, Request *request) override;
bool match(DeviceEnumerator *enumerator) override;
private:
class VimcCameraData : public CameraData
{
public:
VimcCameraData(PipelineHandler *pipe)
: CameraData(pipe)
{
}
~VimcCameraData()
{
delete video_;
}
void bufferReady(Buffer *buffer);
V4L2Device *video_;
Stream stream_;
};
VimcCameraData *cameraData(const Camera *camera)
{
return static_cast<VimcCameraData *>(
PipelineHandler::cameraData(camera));
}
std::shared_ptr<MediaDevice> media_;
};
PipelineHandlerVimc::PipelineHandlerVimc(CameraManager *manager)
: PipelineHandler(manager), media_(nullptr)
{
}
PipelineHandlerVimc::~PipelineHandlerVimc()
{
if (media_)
media_->release();
}
CameraConfiguration
PipelineHandlerVimc::streamConfiguration(Camera *camera,
const std::vector<StreamUsage> &usages)
{
VimcCameraData *data = cameraData(camera);
CameraConfiguration configs;
StreamConfiguration config{};
config.pixelFormat = V4L2_PIX_FMT_RGB24;
config.size = { 640, 480 };
config.bufferCount = 4;
configs[&data->stream_] = config;
return configs;
}
int PipelineHandlerVimc::configureStreams(Camera *camera,
const CameraConfiguration &config)
{
VimcCameraData *data = cameraData(camera);
const StreamConfiguration *cfg = &config[&data->stream_];
int ret;
V4L2DeviceFormat format = {};
format.fourcc = cfg->pixelFormat;
format.size = cfg->size;
ret = data->video_->setFormat(&format);
if (ret)
return ret;
if (format.size != cfg->size ||
format.fourcc != cfg->pixelFormat)
return -EINVAL;
return 0;
}
int PipelineHandlerVimc::allocateBuffers(Camera *camera,
const std::set<Stream *> &streams)
{
VimcCameraData *data = cameraData(camera);
Stream *stream = *streams.begin();
const StreamConfiguration &cfg = stream->configuration();
LOG(VIMC, Debug) << "Requesting " << cfg.bufferCount << " buffers";
return data->video_->exportBuffers(&stream->bufferPool());
}
int PipelineHandlerVimc::freeBuffers(Camera *camera,
const std::set<Stream *> &streams)
{
VimcCameraData *data = cameraData(camera);
return data->video_->releaseBuffers();
}
int PipelineHandlerVimc::start(Camera *camera)
{
VimcCameraData *data = cameraData(camera);
return data->video_->streamOn();
}
void PipelineHandlerVimc::stop(Camera *camera)
{
VimcCameraData *data = cameraData(camera);
data->video_->streamOff();
PipelineHandler::stop(camera);
}
int PipelineHandlerVimc::queueRequest(Camera *camera, Request *request)
{
VimcCameraData *data = cameraData(camera);
Buffer *buffer = request->findBuffer(&data->stream_);
if (!buffer) {
LOG(VIMC, Error)
<< "Attempt to queue request with invalid stream";
return -ENOENT;
}
int ret = data->video_->queueBuffer(buffer);
if (ret < 0)
return ret;
PipelineHandler::queueRequest(camera, request);
return 0;
}
bool PipelineHandlerVimc::match(DeviceEnumerator *enumerator)
{
DeviceMatch dm("vimc");
dm.add("Raw Capture 0");
dm.add("Raw Capture 1");
dm.add("RGB/YUV Capture");
dm.add("Sensor A");
dm.add("Sensor B");
dm.add("Debayer A");
dm.add("Debayer B");
dm.add("RGB/YUV Input");
dm.add("Scaler");
media_ = enumerator->search(dm);
if (!media_)
return false;
media_->acquire();
std::unique_ptr<VimcCameraData> data = utils::make_unique<VimcCameraData>(this);
/* Locate and open the capture video node. */
data->video_ = new V4L2Device(media_->getEntityByName("Raw Capture 1"));
if (data->video_->open())
return false;
data->video_->bufferReady.connect(data.get(), &VimcCameraData::bufferReady);
/* Create and register the camera. */
std::set<Stream *> streams{ &data->stream_ };
std::shared_ptr<Camera> camera = Camera::create(this, "VIMC Sensor B",
streams);
registerCamera(std::move(camera), std::move(data));
return true;
}
void PipelineHandlerVimc::VimcCameraData::bufferReady(Buffer *buffer)
{
Request *request = queuedRequests_.front();
pipe_->completeBuffer(camera_, request, buffer);
pipe_->completeRequest(camera_, request);
}
REGISTER_PIPELINE_HANDLER(PipelineHandlerVimc);
} /* namespace libcamera */