module_param(debug, int, 0644);
MODULE_PARM_DESC(debug, "Debug level 0-3");
+static int media_controller;
+module_param(media_controller, int, 0644);
+MODULE_PARM_DESC(media_controller, "Use media controller API");
+
#define unicam_dbg(level, dev, fmt, arg...) \
v4l2_dbg(level, debug, &(dev)->v4l2_dev, fmt, ##arg)
#define unicam_info(dev, fmt, arg...) \
#define MIN_WIDTH 16
#define MIN_HEIGHT 16
/* Default size of the embedded buffer */
-#define UNICAM_EMBEDDED_SIZE 8192
+#define UNICAM_EMBEDDED_SIZE 16384
/*
* Size of the dummy buffer. Can be any size really, but the DMA
MAX_NODES
};
+#define MASK_CS_DEFAULT BIT(V4L2_COLORSPACE_DEFAULT)
+#define MASK_CS_SMPTE170M BIT(V4L2_COLORSPACE_SMPTE170M)
+#define MASK_CS_SMPTE240M BIT(V4L2_COLORSPACE_SMPTE240M)
+#define MASK_CS_REC709 BIT(V4L2_COLORSPACE_REC709)
+#define MASK_CS_BT878 BIT(V4L2_COLORSPACE_BT878)
+#define MASK_CS_470_M BIT(V4L2_COLORSPACE_470_SYSTEM_M)
+#define MASK_CS_470_BG BIT(V4L2_COLORSPACE_470_SYSTEM_BG)
+#define MASK_CS_JPEG BIT(V4L2_COLORSPACE_JPEG)
+#define MASK_CS_SRGB BIT(V4L2_COLORSPACE_SRGB)
+#define MASK_CS_OPRGB BIT(V4L2_COLORSPACE_OPRGB)
+#define MASK_CS_BT2020 BIT(V4L2_COLORSPACE_BT2020)
+#define MASK_CS_RAW BIT(V4L2_COLORSPACE_RAW)
+#define MASK_CS_DCI_P3 BIT(V4L2_COLORSPACE_DCI_P3)
+
+#define MAX_COLORSPACE 32
+
/*
* struct unicam_fmt - Unicam media bus format information
* @pixelformat: V4L2 pixel format FCC identifier. 0 if n/a.
* @code: V4L2 media bus format code.
* @depth: Bits per pixel as delivered from the source.
* @csi_dt: CSI data type.
+ * @valid_colorspaces: Bitmask of valid colorspaces so that the Media Controller
+ * centric try_fmt can validate the colorspace and pass
+ * v4l2-compliance.
* @check_variants: Flag to denote that there are multiple mediabus formats
* still in the list that could match this V4L2 format.
+ * @mc_skip: Media Controller shouldn't list this format via ENUM_FMT as it is
+ * a duplicate of an earlier format.
+ * @metadata_fmt: This format only applies to the metadata pad.
*/
struct unicam_fmt {
u32 fourcc;
u32 code;
u8 depth;
u8 csi_dt;
- u8 check_variants;
+ u32 valid_colorspaces;
+ u8 check_variants:1;
+ u8 mc_skip:1;
+ u8 metadata_fmt:1;
};
static const struct unicam_fmt formats[] = {
.depth = 16,
.csi_dt = 0x1e,
.check_variants = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_UYVY,
.code = MEDIA_BUS_FMT_UYVY8_2X8,
.depth = 16,
.csi_dt = 0x1e,
.check_variants = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_YVYU,
.code = MEDIA_BUS_FMT_YVYU8_2X8,
.depth = 16,
.csi_dt = 0x1e,
.check_variants = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_VYUY,
.code = MEDIA_BUS_FMT_VYUY8_2X8,
.depth = 16,
.csi_dt = 0x1e,
.check_variants = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_YUYV,
.code = MEDIA_BUS_FMT_YUYV8_1X16,
.depth = 16,
.csi_dt = 0x1e,
+ .mc_skip = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_UYVY,
.code = MEDIA_BUS_FMT_UYVY8_1X16,
.depth = 16,
.csi_dt = 0x1e,
+ .mc_skip = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_YVYU,
.code = MEDIA_BUS_FMT_YVYU8_1X16,
.depth = 16,
.csi_dt = 0x1e,
+ .mc_skip = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
.fourcc = V4L2_PIX_FMT_VYUY,
.code = MEDIA_BUS_FMT_VYUY8_1X16,
.depth = 16,
.csi_dt = 0x1e,
+ .mc_skip = 1,
+ .valid_colorspaces = MASK_CS_SMPTE170M | MASK_CS_REC709 |
+ MASK_CS_JPEG,
}, {
/* RGB Formats */
.fourcc = V4L2_PIX_FMT_RGB565, /* gggbbbbb rrrrrggg */
.code = MEDIA_BUS_FMT_RGB565_2X8_LE,
.depth = 16,
.csi_dt = 0x22,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_RGB565X, /* rrrrrggg gggbbbbb */
.code = MEDIA_BUS_FMT_RGB565_2X8_BE,
.depth = 16,
- .csi_dt = 0x22
+ .csi_dt = 0x22,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_RGB555, /* gggbbbbb arrrrrgg */
.code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE,
.depth = 16,
.csi_dt = 0x21,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_RGB555X, /* arrrrrgg gggbbbbb */
.code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_BE,
.depth = 16,
.csi_dt = 0x21,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_RGB24, /* rgb */
.code = MEDIA_BUS_FMT_RGB888_1X24,
.depth = 24,
.csi_dt = 0x24,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_BGR24, /* bgr */
.code = MEDIA_BUS_FMT_BGR888_1X24,
.depth = 24,
.csi_dt = 0x24,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
.fourcc = V4L2_PIX_FMT_RGB32, /* argb */
.code = MEDIA_BUS_FMT_ARGB8888_1X32,
.depth = 32,
.csi_dt = 0x0,
+ .valid_colorspaces = MASK_CS_SRGB,
}, {
/* Bayer Formats */
.fourcc = V4L2_PIX_FMT_SBGGR8,
.code = MEDIA_BUS_FMT_SBGGR8_1X8,
.depth = 8,
.csi_dt = 0x2a,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGBRG8,
.code = MEDIA_BUS_FMT_SGBRG8_1X8,
.depth = 8,
.csi_dt = 0x2a,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGRBG8,
.code = MEDIA_BUS_FMT_SGRBG8_1X8,
.depth = 8,
.csi_dt = 0x2a,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SRGGB8,
.code = MEDIA_BUS_FMT_SRGGB8_1X8,
.depth = 8,
.csi_dt = 0x2a,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SBGGR10P,
.repacked_fourcc = V4L2_PIX_FMT_SBGGR10,
.code = MEDIA_BUS_FMT_SBGGR10_1X10,
.depth = 10,
.csi_dt = 0x2b,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGBRG10P,
.repacked_fourcc = V4L2_PIX_FMT_SGBRG10,
.code = MEDIA_BUS_FMT_SGBRG10_1X10,
.depth = 10,
.csi_dt = 0x2b,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGRBG10P,
.repacked_fourcc = V4L2_PIX_FMT_SGRBG10,
.code = MEDIA_BUS_FMT_SGRBG10_1X10,
.depth = 10,
.csi_dt = 0x2b,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SRGGB10P,
.repacked_fourcc = V4L2_PIX_FMT_SRGGB10,
.code = MEDIA_BUS_FMT_SRGGB10_1X10,
.depth = 10,
.csi_dt = 0x2b,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SBGGR12P,
.repacked_fourcc = V4L2_PIX_FMT_SBGGR12,
.code = MEDIA_BUS_FMT_SBGGR12_1X12,
.depth = 12,
.csi_dt = 0x2c,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGBRG12P,
.repacked_fourcc = V4L2_PIX_FMT_SGBRG12,
.code = MEDIA_BUS_FMT_SGBRG12_1X12,
.depth = 12,
.csi_dt = 0x2c,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGRBG12P,
.repacked_fourcc = V4L2_PIX_FMT_SGRBG12,
.code = MEDIA_BUS_FMT_SGRBG12_1X12,
.depth = 12,
.csi_dt = 0x2c,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SRGGB12P,
.repacked_fourcc = V4L2_PIX_FMT_SRGGB12,
.code = MEDIA_BUS_FMT_SRGGB12_1X12,
.depth = 12,
.csi_dt = 0x2c,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SBGGR14P,
.repacked_fourcc = V4L2_PIX_FMT_SBGGR14,
.code = MEDIA_BUS_FMT_SBGGR14_1X14,
.depth = 14,
.csi_dt = 0x2d,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGBRG14P,
.repacked_fourcc = V4L2_PIX_FMT_SGBRG14,
.code = MEDIA_BUS_FMT_SGBRG14_1X14,
.depth = 14,
.csi_dt = 0x2d,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SGRBG14P,
.repacked_fourcc = V4L2_PIX_FMT_SGRBG14,
.code = MEDIA_BUS_FMT_SGRBG14_1X14,
.depth = 14,
.csi_dt = 0x2d,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_SRGGB14P,
.repacked_fourcc = V4L2_PIX_FMT_SRGGB14,
.code = MEDIA_BUS_FMT_SRGGB14_1X14,
.depth = 14,
.csi_dt = 0x2d,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
/*
* 16 bit Bayer formats could be supported, but there is no CSI2
.code = MEDIA_BUS_FMT_Y8_1X8,
.depth = 8,
.csi_dt = 0x2a,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_Y10P,
.repacked_fourcc = V4L2_PIX_FMT_Y10,
.code = MEDIA_BUS_FMT_Y10_1X10,
.depth = 10,
.csi_dt = 0x2b,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_Y12P,
.repacked_fourcc = V4L2_PIX_FMT_Y12,
.code = MEDIA_BUS_FMT_Y12_1X12,
.depth = 12,
.csi_dt = 0x2c,
+ .valid_colorspaces = MASK_CS_RAW,
}, {
.fourcc = V4L2_PIX_FMT_Y14P,
.repacked_fourcc = V4L2_PIX_FMT_Y14,
.code = MEDIA_BUS_FMT_Y14_1X14,
.depth = 14,
.csi_dt = 0x2d,
+ .valid_colorspaces = MASK_CS_RAW,
},
/* Embedded data format */
{
.fourcc = V4L2_META_FMT_SENSOR_DATA,
.code = MEDIA_BUS_FMT_SENSOR_DATA,
.depth = 8,
+ .metadata_fmt = 1,
}
};
struct unicam_device *dev;
struct media_pad pad;
unsigned int embedded_lines;
+ struct media_pipeline pipe;
/*
* Dummy buffer intended to be used by unicam
* if we have no other queued buffers to swap to.
struct unicam_node node[MAX_NODES];
struct v4l2_ctrl_handler ctrl_handler;
+
+ bool mc_api;
};
static inline struct unicam_device *
return IRQ_HANDLED;
}
+/* V4L2 Common IOCTLs */
static int unicam_querycap(struct file *file, void *priv,
struct v4l2_capability *cap)
{
return 0;
}
+static int unicam_log_status(struct file *file, void *fh)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ u32 reg;
+
+ /* status for sub devices */
+ v4l2_device_call_all(&dev->v4l2_dev, 0, core, log_status);
+
+ unicam_info(dev, "-----Receiver status-----\n");
+ unicam_info(dev, "V4L2 width/height: %ux%u\n",
+ node->v_fmt.fmt.pix.width, node->v_fmt.fmt.pix.height);
+ unicam_info(dev, "Mediabus format: %08x\n", node->fmt->code);
+ unicam_info(dev, "V4L2 format: %08x\n",
+ node->v_fmt.fmt.pix.pixelformat);
+ reg = reg_read(dev, UNICAM_IPIPE);
+ unicam_info(dev, "Unpacking/packing: %u / %u\n",
+ get_field(reg, UNICAM_PUM_MASK),
+ get_field(reg, UNICAM_PPM_MASK));
+ unicam_info(dev, "----Live data----\n");
+ unicam_info(dev, "Programmed stride: %4u\n",
+ reg_read(dev, UNICAM_IBLS));
+ unicam_info(dev, "Detected resolution: %ux%u\n",
+ reg_read(dev, UNICAM_IHSTA),
+ reg_read(dev, UNICAM_IVSTA));
+ unicam_info(dev, "Write pointer: %08x\n",
+ reg_read(dev, UNICAM_IBWP));
+
+ return 0;
+}
+
+/* V4L2 Video Centric IOCTLs */
static int unicam_enum_fmt_vid_cap(struct file *file, void *priv,
struct v4l2_fmtdesc *f)
{
return 0;
}
-static int unicam_queue_setup(struct vb2_queue *vq,
- unsigned int *nbuffers,
- unsigned int *nplanes,
- unsigned int sizes[],
- struct device *alloc_devs[])
+static int unicam_enum_input(struct file *file, void *priv,
+ struct v4l2_input *inp)
{
- struct unicam_node *node = vb2_get_drv_priv(vq);
+ struct unicam_node *node = video_drvdata(file);
struct unicam_device *dev = node->dev;
- unsigned int size = node->pad_id == IMAGE_PAD ?
- node->v_fmt.fmt.pix.sizeimage :
- node->v_fmt.fmt.meta.buffersize;
+ int ret;
- if (vq->num_buffers + *nbuffers < 3)
- *nbuffers = 3 - vq->num_buffers;
+ if (inp->index != 0)
+ return -EINVAL;
- if (*nplanes) {
- if (sizes[0] < size) {
- unicam_err(dev, "sizes[0] %i < size %u\n", sizes[0],
- size);
- return -EINVAL;
- }
- size = sizes[0];
+ inp->type = V4L2_INPUT_TYPE_CAMERA;
+ if (v4l2_subdev_has_op(dev->sensor, video, s_dv_timings)) {
+ inp->capabilities = V4L2_IN_CAP_DV_TIMINGS;
+ inp->std = 0;
+ } else if (v4l2_subdev_has_op(dev->sensor, video, s_std)) {
+ inp->capabilities = V4L2_IN_CAP_STD;
+ if (v4l2_subdev_call(dev->sensor, video, g_tvnorms, &inp->std) < 0)
+ inp->std = V4L2_STD_ALL;
+ } else {
+ inp->capabilities = 0;
+ inp->std = 0;
}
- *nplanes = 1;
- sizes[0] = size;
+ if (v4l2_subdev_has_op(dev->sensor, video, g_input_status)) {
+ ret = v4l2_subdev_call(dev->sensor, video, g_input_status,
+ &inp->status);
+ if (ret < 0)
+ return ret;
+ }
+ snprintf(inp->name, sizeof(inp->name), "Camera 0");
return 0;
}
-static int unicam_buffer_prepare(struct vb2_buffer *vb)
+static int unicam_g_input(struct file *file, void *priv, unsigned int *i)
{
- struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue);
- struct unicam_device *dev = node->dev;
- struct unicam_buffer *buf = to_unicam_buffer(vb);
- unsigned long size;
+ *i = 0;
- if (WARN_ON(!node->fmt))
- return -EINVAL;
+ return 0;
+}
- size = node->pad_id == IMAGE_PAD ? node->v_fmt.fmt.pix.sizeimage :
- node->v_fmt.fmt.meta.buffersize;
- if (vb2_plane_size(vb, 0) < size) {
- unicam_err(dev, "data will not fit into plane (%lu < %lu)\n",
- vb2_plane_size(vb, 0), size);
+static int unicam_s_input(struct file *file, void *priv, unsigned int i)
+{
+ /*
+ * FIXME: Ideally we would like to be able to query the source
+ * subdevice for information over the input connectors it supports,
+ * and map that through in to a call to video_ops->s_routing.
+ * There is no infrastructure support for defining that within
+ * devicetree at present. Until that is implemented we can't
+ * map a user physical connector number to s_routing input number.
+ */
+ if (i > 0)
return -EINVAL;
- }
- vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size);
return 0;
}
-static void unicam_buffer_queue(struct vb2_buffer *vb)
+static int unicam_querystd(struct file *file, void *priv,
+ v4l2_std_id *std)
{
- struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue);
- struct unicam_buffer *buf = to_unicam_buffer(vb);
- unsigned long flags;
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- spin_lock_irqsave(&node->dma_queue_lock, flags);
- list_add_tail(&buf->list, &node->dma_queue);
- spin_unlock_irqrestore(&node->dma_queue_lock, flags);
+ return v4l2_subdev_call(dev->sensor, video, querystd, std);
}
-static void unicam_set_packing_config(struct unicam_device *dev)
+static int unicam_g_std(struct file *file, void *priv, v4l2_std_id *std)
{
- u32 pack, unpack;
- u32 val;
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- if (dev->node[IMAGE_PAD].v_fmt.fmt.pix.pixelformat ==
- dev->node[IMAGE_PAD].fmt->fourcc) {
- unpack = UNICAM_PUM_NONE;
- pack = UNICAM_PPM_NONE;
- } else {
- switch (dev->node[IMAGE_PAD].fmt->depth) {
- case 8:
- unpack = UNICAM_PUM_UNPACK8;
- break;
- case 10:
- unpack = UNICAM_PUM_UNPACK10;
- break;
- case 12:
- unpack = UNICAM_PUM_UNPACK12;
- break;
- case 14:
- unpack = UNICAM_PUM_UNPACK14;
- break;
- case 16:
- unpack = UNICAM_PUM_UNPACK16;
- break;
- default:
- unpack = UNICAM_PUM_NONE;
- break;
- }
+ return v4l2_subdev_call(dev->sensor, video, g_std, std);
+}
- /* Repacking is always to 16bpp */
- pack = UNICAM_PPM_PACK16;
- }
+static int unicam_s_std(struct file *file, void *priv, v4l2_std_id std)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ int ret;
+ v4l2_std_id current_std;
- val = 0;
- set_field(&val, unpack, UNICAM_PUM_MASK);
- set_field(&val, pack, UNICAM_PPM_MASK);
- reg_write(dev, UNICAM_IPIPE, val);
+ ret = v4l2_subdev_call(dev->sensor, video, g_std, ¤t_std);
+ if (ret)
+ return ret;
+
+ if (std == current_std)
+ return 0;
+
+ if (vb2_is_busy(&node->buffer_queue))
+ return -EBUSY;
+
+ ret = v4l2_subdev_call(dev->sensor, video, s_std, std);
+
+ /* Force recomputation of bytesperline */
+ node->v_fmt.fmt.pix.bytesperline = 0;
+
+ unicam_reset_format(node);
+
+ return ret;
}
-static void unicam_cfg_image_id(struct unicam_device *dev)
+static int unicam_s_edid(struct file *file, void *priv, struct v4l2_edid *edid)
{
- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
- /* CSI2 mode, hardcode VC 0 for now. */
- reg_write(dev, UNICAM_IDI0,
- (0 << 6) | dev->node[IMAGE_PAD].fmt->csi_dt);
- } else {
- /* CCP2 mode */
- reg_write(dev, UNICAM_IDI0,
- 0x80 | dev->node[IMAGE_PAD].fmt->csi_dt);
- }
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+
+ return v4l2_subdev_call(dev->sensor, pad, set_edid, edid);
}
-static void unicam_enable_ed(struct unicam_device *dev)
+static int unicam_g_edid(struct file *file, void *priv, struct v4l2_edid *edid)
{
- u32 val = reg_read(dev, UNICAM_DCS);
-
- set_field(&val, 2, UNICAM_EDL_MASK);
- /* Do not wrap at the end of the embedded data buffer */
- set_field(&val, 0, UNICAM_DBOB);
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- reg_write(dev, UNICAM_DCS, val);
+ return v4l2_subdev_call(dev->sensor, pad, get_edid, edid);
}
-static void unicam_start_rx(struct unicam_device *dev, dma_addr_t *addr)
+static int unicam_s_selection(struct file *file, void *priv,
+ struct v4l2_selection *sel)
{
- int line_int_freq = dev->node[IMAGE_PAD].v_fmt.fmt.pix.height >> 2;
- unsigned int size, i;
- u32 val;
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ struct v4l2_subdev_selection sdsel = {
+ .which = V4L2_SUBDEV_FORMAT_ACTIVE,
+ .target = sel->target,
+ .flags = sel->flags,
+ .r = sel->r,
+ };
- if (line_int_freq < 128)
- line_int_freq = 128;
+ if (sel->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
+ return -EINVAL;
- /* Enable lane clocks */
- val = 1;
- for (i = 0; i < dev->active_data_lanes; i++)
- val = val << 2 | 1;
- clk_write(dev, val);
+ return v4l2_subdev_call(dev->sensor, pad, set_selection, NULL, &sdsel);
+}
- /* Basic init */
- reg_write(dev, UNICAM_CTRL, UNICAM_MEM);
+static int unicam_g_selection(struct file *file, void *priv,
+ struct v4l2_selection *sel)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ struct v4l2_subdev_selection sdsel = {
+ .which = V4L2_SUBDEV_FORMAT_ACTIVE,
+ .target = sel->target,
+ };
+ int ret;
- /* Enable analogue control, and leave in reset. */
- val = UNICAM_AR;
- set_field(&val, 7, UNICAM_CTATADJ_MASK);
- set_field(&val, 7, UNICAM_PTATADJ_MASK);
- reg_write(dev, UNICAM_ANA, val);
- usleep_range(1000, 2000);
+ if (sel->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
+ return -EINVAL;
- /* Come out of reset */
- reg_write_field(dev, UNICAM_ANA, 0, UNICAM_AR);
+ ret = v4l2_subdev_call(dev->sensor, pad, get_selection, NULL, &sdsel);
+ if (!ret)
+ sel->r = sdsel.r;
- /* Peripheral reset */
- reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPR);
- reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPR);
+ return ret;
+}
- reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPE);
+static int unicam_enum_framesizes(struct file *file, void *priv,
+ struct v4l2_frmsizeenum *fsize)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ const struct unicam_fmt *fmt;
+ struct v4l2_subdev_frame_size_enum fse;
+ int ret;
- /* Enable Rx control. */
- val = reg_read(dev, UNICAM_CTRL);
- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
- set_field(&val, UNICAM_CPM_CSI2, UNICAM_CPM_MASK);
- set_field(&val, UNICAM_DCM_STROBE, UNICAM_DCM_MASK);
- } else {
- set_field(&val, UNICAM_CPM_CCP2, UNICAM_CPM_MASK);
- set_field(&val, dev->bus_flags, UNICAM_DCM_MASK);
+ /* check for valid format */
+ fmt = find_format_by_pix(dev, fsize->pixel_format);
+ if (!fmt) {
+ unicam_dbg(3, dev, "Invalid pixel code: %x\n",
+ fsize->pixel_format);
+ return -EINVAL;
}
- /* Packet framer timeout */
- set_field(&val, 0xf, UNICAM_PFT_MASK);
- set_field(&val, 128, UNICAM_OET_MASK);
- reg_write(dev, UNICAM_CTRL, val);
-
- reg_write(dev, UNICAM_IHWIN, 0);
- reg_write(dev, UNICAM_IVWIN, 0);
-
- /* AXI bus access QoS setup */
- val = reg_read(dev, UNICAM_PRI);
- set_field(&val, 0, UNICAM_BL_MASK);
- set_field(&val, 0, UNICAM_BS_MASK);
- set_field(&val, 0xe, UNICAM_PP_MASK);
- set_field(&val, 8, UNICAM_NP_MASK);
- set_field(&val, 2, UNICAM_PT_MASK);
- set_field(&val, 1, UNICAM_PE);
- reg_write(dev, UNICAM_PRI, val);
-
- reg_write_field(dev, UNICAM_ANA, 0, UNICAM_DDL);
+ fse.code = fmt->code;
- /* Always start in trigger frame capture mode (UNICAM_FCM set) */
- val = UNICAM_FSIE | UNICAM_FEIE | UNICAM_FCM | UNICAM_IBOB;
- set_field(&val, line_int_freq, UNICAM_LCIE_MASK);
- reg_write(dev, UNICAM_ICTL, val);
- reg_write(dev, UNICAM_STA, UNICAM_STA_MASK_ALL);
- reg_write(dev, UNICAM_ISTA, UNICAM_ISTA_MASK_ALL);
+ fse.which = V4L2_SUBDEV_FORMAT_ACTIVE;
+ fse.index = fsize->index;
+ fse.pad = node->src_pad_id;
- /* tclk_term_en */
- reg_write_field(dev, UNICAM_CLT, 2, UNICAM_CLT1_MASK);
- /* tclk_settle */
- reg_write_field(dev, UNICAM_CLT, 6, UNICAM_CLT2_MASK);
- /* td_term_en */
- reg_write_field(dev, UNICAM_DLT, 2, UNICAM_DLT1_MASK);
- /* ths_settle */
- reg_write_field(dev, UNICAM_DLT, 6, UNICAM_DLT2_MASK);
- /* trx_enable */
- reg_write_field(dev, UNICAM_DLT, 0, UNICAM_DLT3_MASK);
+ ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_size, NULL, &fse);
+ if (ret)
+ return ret;
- reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_SOE);
+ unicam_dbg(1, dev, "%s: index: %d code: %x W:[%d,%d] H:[%d,%d]\n",
+ __func__, fse.index, fse.code, fse.min_width, fse.max_width,
+ fse.min_height, fse.max_height);
- /* Packet compare setup - required to avoid missing frame ends */
- val = 0;
- set_field(&val, 1, UNICAM_PCE);
- set_field(&val, 1, UNICAM_GI);
- set_field(&val, 1, UNICAM_CPH);
- set_field(&val, 0, UNICAM_PCVC_MASK);
- set_field(&val, 1, UNICAM_PCDT_MASK);
- reg_write(dev, UNICAM_CMP0, val);
+ fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE;
+ fsize->discrete.width = fse.max_width;
+ fsize->discrete.height = fse.max_height;
- /* Enable clock lane and set up terminations */
- val = 0;
- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
- /* CSI2 */
- set_field(&val, 1, UNICAM_CLE);
- set_field(&val, 1, UNICAM_CLLPE);
- if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) {
- set_field(&val, 1, UNICAM_CLTRE);
- set_field(&val, 1, UNICAM_CLHSE);
- }
- } else {
- /* CCP2 */
- set_field(&val, 1, UNICAM_CLE);
- set_field(&val, 1, UNICAM_CLHSE);
- set_field(&val, 1, UNICAM_CLTRE);
- }
- reg_write(dev, UNICAM_CLK, val);
+ return 0;
+}
- /*
- * Enable required data lanes with appropriate terminations.
- * The same value needs to be written to UNICAM_DATn registers for
- * the active lanes, and 0 for inactive ones.
- */
- val = 0;
- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
- /* CSI2 */
- set_field(&val, 1, UNICAM_DLE);
- set_field(&val, 1, UNICAM_DLLPE);
- if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) {
- set_field(&val, 1, UNICAM_DLTRE);
- set_field(&val, 1, UNICAM_DLHSE);
- }
- } else {
- /* CCP2 */
- set_field(&val, 1, UNICAM_DLE);
- set_field(&val, 1, UNICAM_DLHSE);
- set_field(&val, 1, UNICAM_DLTRE);
- }
- reg_write(dev, UNICAM_DAT0, val);
+static int unicam_enum_frameintervals(struct file *file, void *priv,
+ struct v4l2_frmivalenum *fival)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ const struct unicam_fmt *fmt;
+ struct v4l2_subdev_frame_interval_enum fie = {
+ .index = fival->index,
+ .pad = node->src_pad_id,
+ .width = fival->width,
+ .height = fival->height,
+ .which = V4L2_SUBDEV_FORMAT_ACTIVE,
+ };
+ int ret;
- if (dev->active_data_lanes == 1)
- val = 0;
- reg_write(dev, UNICAM_DAT1, val);
+ fmt = find_format_by_pix(dev, fival->pixel_format);
+ if (!fmt)
+ return -EINVAL;
- if (dev->max_data_lanes > 2) {
- /*
- * Registers UNICAM_DAT2 and UNICAM_DAT3 only valid if the
- * instance supports more than 2 data lanes.
- */
- if (dev->active_data_lanes == 2)
- val = 0;
- reg_write(dev, UNICAM_DAT2, val);
+ fie.code = fmt->code;
+ ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_interval,
+ NULL, &fie);
+ if (ret)
+ return ret;
- if (dev->active_data_lanes == 3)
- val = 0;
- reg_write(dev, UNICAM_DAT3, val);
- }
+ fival->type = V4L2_FRMIVAL_TYPE_DISCRETE;
+ fival->discrete = fie.interval;
- reg_write(dev, UNICAM_IBLS,
- dev->node[IMAGE_PAD].v_fmt.fmt.pix.bytesperline);
- size = dev->node[IMAGE_PAD].v_fmt.fmt.pix.sizeimage;
- unicam_wr_dma_addr(dev, addr[IMAGE_PAD], size, IMAGE_PAD);
- unicam_set_packing_config(dev);
- unicam_cfg_image_id(dev);
+ return 0;
+}
- val = reg_read(dev, UNICAM_MISC);
- set_field(&val, 1, UNICAM_FL0);
- set_field(&val, 1, UNICAM_FL1);
- reg_write(dev, UNICAM_MISC, val);
+static int unicam_g_parm(struct file *file, void *fh, struct v4l2_streamparm *a)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data) {
- size = dev->node[METADATA_PAD].v_fmt.fmt.meta.buffersize;
- unicam_enable_ed(dev);
- unicam_wr_dma_addr(dev, addr[METADATA_PAD], size, METADATA_PAD);
- }
+ return v4l2_g_parm_cap(video_devdata(file), dev->sensor, a);
+}
- /* Enable peripheral */
- reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPE);
+static int unicam_s_parm(struct file *file, void *fh, struct v4l2_streamparm *a)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- /* Load image pointers */
- reg_write_field(dev, UNICAM_ICTL, 1, UNICAM_LIP_MASK);
+ return v4l2_s_parm_cap(video_devdata(file), dev->sensor, a);
+}
- /* Load embedded data buffer pointers if needed */
- if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data)
- reg_write_field(dev, UNICAM_DCS, 1, UNICAM_LDP);
+static int unicam_g_dv_timings(struct file *file, void *priv,
+ struct v4l2_dv_timings *timings)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- /*
- * Enable trigger only for the first frame to
- * sync correctly to the FS from the source.
- */
- reg_write_field(dev, UNICAM_ICTL, 1, UNICAM_TFC);
+ return v4l2_subdev_call(dev->sensor, video, g_dv_timings, timings);
}
-static void unicam_disable(struct unicam_device *dev)
+static int unicam_s_dv_timings(struct file *file, void *priv,
+ struct v4l2_dv_timings *timings)
{
- /* Analogue lane control disable */
- reg_write_field(dev, UNICAM_ANA, 1, UNICAM_DDL);
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ struct v4l2_dv_timings current_timings;
+ int ret;
- /* Stop the output engine */
- reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_SOE);
+ ret = v4l2_subdev_call(dev->sensor, video, g_dv_timings,
+ ¤t_timings);
- /* Disable the data lanes. */
- reg_write(dev, UNICAM_DAT0, 0);
- reg_write(dev, UNICAM_DAT1, 0);
+ if (ret < 0)
+ return ret;
- if (dev->max_data_lanes > 2) {
- reg_write(dev, UNICAM_DAT2, 0);
- reg_write(dev, UNICAM_DAT3, 0);
- }
+ if (v4l2_match_dv_timings(timings, ¤t_timings, 0, false))
+ return 0;
- /* Peripheral reset */
- reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPR);
- usleep_range(50, 100);
- reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPR);
+ if (vb2_is_busy(&node->buffer_queue))
+ return -EBUSY;
- /* Disable peripheral */
- reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPE);
+ ret = v4l2_subdev_call(dev->sensor, video, s_dv_timings, timings);
- /* Clear ED setup */
- reg_write(dev, UNICAM_DCS, 0);
+ /* Force recomputation of bytesperline */
+ node->v_fmt.fmt.pix.bytesperline = 0;
- /* Disable all lane clocks */
- clk_write(dev, 0);
-}
+ unicam_reset_format(node);
-static void unicam_return_buffers(struct unicam_node *node,
- enum vb2_buffer_state state)
-{
- struct unicam_buffer *buf, *tmp;
- unsigned long flags;
-
- spin_lock_irqsave(&node->dma_queue_lock, flags);
- list_for_each_entry_safe(buf, tmp, &node->dma_queue, list) {
- list_del(&buf->list);
- vb2_buffer_done(&buf->vb.vb2_buf, state);
- }
+ return ret;
+}
- if (node->cur_frm)
- vb2_buffer_done(&node->cur_frm->vb.vb2_buf,
- state);
- if (node->next_frm && node->cur_frm != node->next_frm)
- vb2_buffer_done(&node->next_frm->vb.vb2_buf,
- state);
+static int unicam_query_dv_timings(struct file *file, void *priv,
+ struct v4l2_dv_timings *timings)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
- node->cur_frm = NULL;
- node->next_frm = NULL;
- spin_unlock_irqrestore(&node->dma_queue_lock, flags);
+ return v4l2_subdev_call(dev->sensor, video, query_dv_timings, timings);
}
-static int unicam_start_streaming(struct vb2_queue *vq, unsigned int count)
+static int unicam_enum_dv_timings(struct file *file, void *priv,
+ struct v4l2_enum_dv_timings *timings)
{
- struct unicam_node *node = vb2_get_drv_priv(vq);
+ struct unicam_node *node = video_drvdata(file);
struct unicam_device *dev = node->dev;
- dma_addr_t buffer_addr[MAX_NODES] = { 0 };
- unsigned long flags;
- unsigned int i;
int ret;
- node->streaming = true;
- if (!(dev->node[IMAGE_PAD].open && dev->node[IMAGE_PAD].streaming &&
- (!dev->node[METADATA_PAD].open ||
- dev->node[METADATA_PAD].streaming))) {
- /*
- * Metadata pad must be enabled before image pad if it is
- * wanted.
- */
- unicam_dbg(3, dev, "Not all nodes are streaming yet.");
- return 0;
- }
+ timings->pad = node->src_pad_id;
+ ret = v4l2_subdev_call(dev->sensor, pad, enum_dv_timings, timings);
+ timings->pad = node->pad_id;
- dev->sequence = 0;
- ret = unicam_runtime_get(dev);
- if (ret < 0) {
- unicam_dbg(3, dev, "unicam_runtime_get failed\n");
- goto err_streaming;
- }
+ return ret;
+}
- dev->active_data_lanes = dev->max_data_lanes;
+static int unicam_dv_timings_cap(struct file *file, void *priv,
+ struct v4l2_dv_timings_cap *cap)
+{
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ int ret;
- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
- struct v4l2_mbus_config mbus_config = { 0 };
+ cap->pad = node->src_pad_id;
+ ret = v4l2_subdev_call(dev->sensor, pad, dv_timings_cap, cap);
+ cap->pad = node->pad_id;
- ret = v4l2_subdev_call(dev->sensor, pad, get_mbus_config,
- 0, &mbus_config);
- if (ret < 0 && ret != -ENOIOCTLCMD) {
- unicam_dbg(3, dev, "g_mbus_config failed\n");
- goto err_pm_put;
- }
+ return ret;
+}
- dev->active_data_lanes =
- (mbus_config.flags & V4L2_MBUS_CSI2_LANE_MASK) >>
- __ffs(V4L2_MBUS_CSI2_LANE_MASK);
- if (!dev->active_data_lanes)
- dev->active_data_lanes = dev->max_data_lanes;
- if (dev->active_data_lanes > dev->max_data_lanes) {
- unicam_err(dev, "Device has requested %u data lanes, which is >%u configured in DT\n",
- dev->active_data_lanes,
- dev->max_data_lanes);
- ret = -EINVAL;
- goto err_pm_put;
- }
+static int unicam_subscribe_event(struct v4l2_fh *fh,
+ const struct v4l2_event_subscription *sub)
+{
+ switch (sub->type) {
+ case V4L2_EVENT_FRAME_SYNC:
+ return v4l2_event_subscribe(fh, sub, 2, NULL);
+ case V4L2_EVENT_SOURCE_CHANGE:
+ return v4l2_event_subscribe(fh, sub, 4, NULL);
}
- unicam_dbg(1, dev, "Running with %u data lanes\n",
- dev->active_data_lanes);
-
- dev->vpu_req = clk_request_start(dev->vpu_clock, MIN_VPU_CLOCK_RATE);
- if (!dev->vpu_req) {
- unicam_err(dev, "failed to set up VPU clock\n");
- goto err_pm_put;
- }
+ return v4l2_ctrl_subscribe_event(fh, sub);
+}
- ret = clk_prepare_enable(dev->vpu_clock);
- if (ret) {
- unicam_err(dev, "Failed to enable VPU clock: %d\n", ret);
- goto err_pm_put;
- }
+static void unicam_notify(struct v4l2_subdev *sd,
+ unsigned int notification, void *arg)
+{
+ struct unicam_device *dev = to_unicam_device(sd->v4l2_dev);
- ret = clk_set_rate(dev->clock, 100 * 1000 * 1000);
- if (ret) {
- unicam_err(dev, "failed to set up CSI clock\n");
- goto err_vpu_clock;
+ switch (notification) {
+ case V4L2_DEVICE_NOTIFY_EVENT:
+ v4l2_event_queue(&dev->node[IMAGE_PAD].video_dev, arg);
+ break;
+ default:
+ break;
}
+}
- ret = clk_prepare_enable(dev->clock);
- if (ret) {
- unicam_err(dev, "Failed to enable CSI clock: %d\n", ret);
- goto err_vpu_clock;
- }
+/* unicam capture ioctl operations */
+static const struct v4l2_ioctl_ops unicam_ioctl_ops = {
+ .vidioc_querycap = unicam_querycap,
+ .vidioc_enum_fmt_vid_cap = unicam_enum_fmt_vid_cap,
+ .vidioc_g_fmt_vid_cap = unicam_g_fmt_vid_cap,
+ .vidioc_s_fmt_vid_cap = unicam_s_fmt_vid_cap,
+ .vidioc_try_fmt_vid_cap = unicam_try_fmt_vid_cap,
- for (i = 0; i < ARRAY_SIZE(dev->node); i++) {
- struct unicam_buffer *buf;
+ .vidioc_enum_fmt_meta_cap = unicam_enum_fmt_meta_cap,
+ .vidioc_g_fmt_meta_cap = unicam_g_fmt_meta_cap,
+ .vidioc_s_fmt_meta_cap = unicam_g_fmt_meta_cap,
+ .vidioc_try_fmt_meta_cap = unicam_g_fmt_meta_cap,
- if (!dev->node[i].streaming)
- continue;
+ .vidioc_enum_input = unicam_enum_input,
+ .vidioc_g_input = unicam_g_input,
+ .vidioc_s_input = unicam_s_input,
- spin_lock_irqsave(&dev->node[i].dma_queue_lock, flags);
- buf = list_first_entry(&dev->node[i].dma_queue,
- struct unicam_buffer, list);
- dev->node[i].cur_frm = buf;
- dev->node[i].next_frm = buf;
- list_del(&buf->list);
- spin_unlock_irqrestore(&dev->node[i].dma_queue_lock, flags);
+ .vidioc_querystd = unicam_querystd,
+ .vidioc_s_std = unicam_s_std,
+ .vidioc_g_std = unicam_g_std,
- buffer_addr[i] =
- vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0);
- }
+ .vidioc_g_edid = unicam_g_edid,
+ .vidioc_s_edid = unicam_s_edid,
- unicam_start_rx(dev, buffer_addr);
+ .vidioc_enum_framesizes = unicam_enum_framesizes,
+ .vidioc_enum_frameintervals = unicam_enum_frameintervals,
- ret = v4l2_subdev_call(dev->sensor, video, s_stream, 1);
- if (ret < 0) {
- unicam_err(dev, "stream on failed in subdev\n");
- goto err_disable_unicam;
- }
+ .vidioc_g_selection = unicam_g_selection,
+ .vidioc_s_selection = unicam_s_selection,
- dev->clocks_enabled = true;
- return 0;
+ .vidioc_g_parm = unicam_g_parm,
+ .vidioc_s_parm = unicam_s_parm,
-err_disable_unicam:
- unicam_disable(dev);
- clk_disable_unprepare(dev->clock);
-err_vpu_clock:
- clk_request_done(dev->vpu_req);
- clk_disable_unprepare(dev->vpu_clock);
-err_pm_put:
- unicam_runtime_put(dev);
-err_streaming:
- unicam_return_buffers(node, VB2_BUF_STATE_QUEUED);
- node->streaming = false;
+ .vidioc_s_dv_timings = unicam_s_dv_timings,
+ .vidioc_g_dv_timings = unicam_g_dv_timings,
+ .vidioc_query_dv_timings = unicam_query_dv_timings,
+ .vidioc_enum_dv_timings = unicam_enum_dv_timings,
+ .vidioc_dv_timings_cap = unicam_dv_timings_cap,
- return ret;
-}
+ .vidioc_reqbufs = vb2_ioctl_reqbufs,
+ .vidioc_create_bufs = vb2_ioctl_create_bufs,
+ .vidioc_prepare_buf = vb2_ioctl_prepare_buf,
+ .vidioc_querybuf = vb2_ioctl_querybuf,
+ .vidioc_qbuf = vb2_ioctl_qbuf,
+ .vidioc_dqbuf = vb2_ioctl_dqbuf,
+ .vidioc_expbuf = vb2_ioctl_expbuf,
+ .vidioc_streamon = vb2_ioctl_streamon,
+ .vidioc_streamoff = vb2_ioctl_streamoff,
-static void unicam_stop_streaming(struct vb2_queue *vq)
-{
- struct unicam_node *node = vb2_get_drv_priv(vq);
- struct unicam_device *dev = node->dev;
+ .vidioc_log_status = unicam_log_status,
+ .vidioc_subscribe_event = unicam_subscribe_event,
+ .vidioc_unsubscribe_event = v4l2_event_unsubscribe,
+};
- node->streaming = false;
+/* V4L2 Media Controller Centric IOCTLs */
- if (node->pad_id == IMAGE_PAD) {
- /*
- * Stop streaming the sensor and disable the peripheral.
- * We cannot continue streaming embedded data with the
- * image pad disabled.
- */
- if (v4l2_subdev_call(dev->sensor, video, s_stream, 0) < 0)
- unicam_err(dev, "stream off failed in subdev\n");
+static int unicam_mc_enum_fmt_vid_cap(struct file *file, void *priv,
+ struct v4l2_fmtdesc *f)
+{
+ int i, j;
- unicam_disable(dev);
+ for (i = 0, j = 0; i < ARRAY_SIZE(formats); i++) {
+ if (f->mbus_code && formats[i].code != f->mbus_code)
+ continue;
+ if (formats[i].mc_skip || formats[i].metadata_fmt)
+ continue;
- if (dev->clocks_enabled) {
- clk_request_done(dev->vpu_req);
- clk_disable_unprepare(dev->vpu_clock);
- clk_disable_unprepare(dev->clock);
- dev->clocks_enabled = false;
+ if (formats[i].fourcc) {
+ if (j == f->index) {
+ f->pixelformat = formats[i].fourcc;
+ f->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
+ return 0;
+ }
+ j++;
+ }
+ if (formats[i].repacked_fourcc) {
+ if (j == f->index) {
+ f->pixelformat = formats[i].repacked_fourcc;
+ f->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
+ return 0;
+ }
+ j++;
}
- unicam_runtime_put(dev);
-
- } else if (node->pad_id == METADATA_PAD) {
- /*
- * Allow the hardware to spin in the dummy buffer.
- * This is only really needed if the embedded data pad is
- * disabled before the image pad.
- */
- unicam_wr_dma_addr(dev, node->dummy_buf_dma_addr,
- DUMMY_BUF_SIZE, METADATA_PAD);
}
- /* Clear all queued buffers for the node */
- unicam_return_buffers(node, VB2_BUF_STATE_ERROR);
+ return -EINVAL;
}
-static int unicam_enum_input(struct file *file, void *priv,
- struct v4l2_input *inp)
+static int unicam_mc_g_fmt_vid_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
{
struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- int ret;
- if (inp->index != 0)
+ if (node->pad_id != IMAGE_PAD)
return -EINVAL;
- inp->type = V4L2_INPUT_TYPE_CAMERA;
- if (v4l2_subdev_has_op(dev->sensor, video, s_dv_timings)) {
- inp->capabilities = V4L2_IN_CAP_DV_TIMINGS;
- inp->std = 0;
- } else if (v4l2_subdev_has_op(dev->sensor, video, s_std)) {
- inp->capabilities = V4L2_IN_CAP_STD;
- if (v4l2_subdev_call(dev->sensor, video, g_tvnorms, &inp->std) < 0)
- inp->std = V4L2_STD_ALL;
- } else {
- inp->capabilities = 0;
- inp->std = 0;
+ *f = node->v_fmt;
+
+ return 0;
+}
+
+static void unicam_mc_try_fmt(struct unicam_node *node, struct v4l2_format *f,
+ const struct unicam_fmt **ret_fmt)
+{
+ struct v4l2_pix_format *v4l2_format = &f->fmt.pix;
+ struct unicam_device *dev = node->dev;
+ const struct unicam_fmt *fmt;
+ int is_rgb;
+
+ /*
+ * Default to the first format if the requested pixel format code isn't
+ * supported.
+ */
+ fmt = find_format_by_pix(dev, v4l2_format->pixelformat);
+ if (!fmt) {
+ fmt = &formats[0];
+ v4l2_format->pixelformat = fmt->fourcc;
}
- if (v4l2_subdev_has_op(dev->sensor, video, g_input_status)) {
- ret = v4l2_subdev_call(dev->sensor, video, g_input_status,
- &inp->status);
- if (ret < 0)
- return ret;
+ unicam_calc_format_size_bpl(dev, fmt, f);
+
+ if (v4l2_format->field == V4L2_FIELD_ANY)
+ v4l2_format->field = V4L2_FIELD_NONE;
+
+ if (ret_fmt)
+ *ret_fmt = fmt;
+
+ if (v4l2_format->colorspace >= MAX_COLORSPACE ||
+ !(fmt->valid_colorspaces & (1 << v4l2_format->colorspace))) {
+ v4l2_format->colorspace = __ffs(fmt->valid_colorspaces);
+
+ v4l2_format->xfer_func =
+ V4L2_MAP_XFER_FUNC_DEFAULT(v4l2_format->colorspace);
+ v4l2_format->ycbcr_enc =
+ V4L2_MAP_YCBCR_ENC_DEFAULT(v4l2_format->colorspace);
+ is_rgb = v4l2_format->colorspace == V4L2_COLORSPACE_SRGB;
+ v4l2_format->quantization =
+ V4L2_MAP_QUANTIZATION_DEFAULT(is_rgb,
+ v4l2_format->colorspace,
+ v4l2_format->ycbcr_enc);
}
- snprintf(inp->name, sizeof(inp->name), "Camera 0");
+ unicam_dbg(3, dev, "%s: %08x %ux%u (bytesperline %u sizeimage %u)\n",
+ __func__, v4l2_format->pixelformat,
+ v4l2_format->width, v4l2_format->height,
+ v4l2_format->bytesperline, v4l2_format->sizeimage);
+}
+
+static int unicam_mc_try_fmt_vid_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
+{
+ struct unicam_node *node = video_drvdata(file);
+
+ unicam_mc_try_fmt(node, f, NULL);
return 0;
}
-static int unicam_g_input(struct file *file, void *priv, unsigned int *i)
+static int unicam_mc_s_fmt_vid_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
{
- *i = 0;
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+ const struct unicam_fmt *fmt;
+
+ if (vb2_is_busy(&node->buffer_queue)) {
+ unicam_dbg(3, dev, "%s device busy\n", __func__);
+ return -EBUSY;
+ }
+
+ unicam_mc_try_fmt(node, f, &fmt);
+
+ node->v_fmt = *f;
+ node->fmt = fmt;
return 0;
}
-static int unicam_s_input(struct file *file, void *priv, unsigned int i)
+static int unicam_mc_enum_framesizes(struct file *file, void *fh,
+ struct v4l2_frmsizeenum *fsize)
{
- /*
- * FIXME: Ideally we would like to be able to query the source
- * subdevice for information over the input connectors it supports,
- * and map that through in to a call to video_ops->s_routing.
- * There is no infrastructure support for defining that within
- * devicetree at present. Until that is implemented we can't
- * map a user physical connector number to s_routing input number.
- */
- if (i > 0)
+ struct unicam_node *node = video_drvdata(file);
+ struct unicam_device *dev = node->dev;
+
+ if (fsize->index > 0)
+ return -EINVAL;
+
+ if (!find_format_by_pix(dev, fsize->pixel_format)) {
+ unicam_dbg(3, dev, "Invalid pixel format 0x%08x\n",
+ fsize->pixel_format);
return -EINVAL;
+ }
+
+ fsize->type = V4L2_FRMSIZE_TYPE_STEPWISE;
+ fsize->stepwise.min_width = MIN_WIDTH;
+ fsize->stepwise.max_width = MAX_WIDTH;
+ fsize->stepwise.step_width = 1;
+ fsize->stepwise.min_height = MIN_HEIGHT;
+ fsize->stepwise.max_height = MAX_HEIGHT;
+ fsize->stepwise.step_height = 1;
return 0;
}
-static int unicam_querystd(struct file *file, void *priv,
- v4l2_std_id *std)
+static int unicam_mc_enum_fmt_meta_cap(struct file *file, void *priv,
+ struct v4l2_fmtdesc *f)
+{
+ int i, j;
+
+ for (i = 0, j = 0; i < ARRAY_SIZE(formats); i++) {
+ if (f->mbus_code && formats[i].code != f->mbus_code)
+ continue;
+ if (!formats[i].metadata_fmt)
+ continue;
+
+ if (formats[i].fourcc) {
+ if (j == f->index) {
+ f->pixelformat = formats[i].fourcc;
+ f->type = V4L2_BUF_TYPE_META_CAPTURE;
+ return 0;
+ }
+ j++;
+ }
+ }
+
+ return -EINVAL;
+}
+
+static int unicam_mc_g_fmt_meta_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
{
struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- return v4l2_subdev_call(dev->sensor, video, querystd, std);
+ if (node->pad_id != METADATA_PAD)
+ return -EINVAL;
+
+ *f = node->v_fmt;
+
+ return 0;
}
-static int unicam_g_std(struct file *file, void *priv, v4l2_std_id *std)
+static int unicam_mc_try_fmt_meta_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
{
struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- return v4l2_subdev_call(dev->sensor, video, g_std, std);
+ if (node->pad_id != METADATA_PAD)
+ return -EINVAL;
+
+ f->fmt.meta.dataformat = V4L2_META_FMT_SENSOR_DATA;
+
+ return 0;
}
-static int unicam_s_std(struct file *file, void *priv, v4l2_std_id std)
+static int unicam_mc_s_fmt_meta_cap(struct file *file, void *priv,
+ struct v4l2_format *f)
{
struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+
+ if (node->pad_id != METADATA_PAD)
+ return -EINVAL;
+
+ unicam_mc_try_fmt_meta_cap(file, priv, f);
+
+ node->v_fmt = *f;
+
+ return 0;
+}
+
+static const struct v4l2_ioctl_ops unicam_mc_ioctl_ops = {
+ .vidioc_querycap = unicam_querycap,
+ .vidioc_enum_fmt_vid_cap = unicam_mc_enum_fmt_vid_cap,
+ .vidioc_g_fmt_vid_cap = unicam_mc_g_fmt_vid_cap,
+ .vidioc_try_fmt_vid_cap = unicam_mc_try_fmt_vid_cap,
+ .vidioc_s_fmt_vid_cap = unicam_mc_s_fmt_vid_cap,
+
+ .vidioc_enum_fmt_meta_cap = unicam_mc_enum_fmt_meta_cap,
+ .vidioc_g_fmt_meta_cap = unicam_mc_g_fmt_meta_cap,
+ .vidioc_try_fmt_meta_cap = unicam_mc_try_fmt_meta_cap,
+ .vidioc_s_fmt_meta_cap = unicam_mc_s_fmt_meta_cap,
+
+ .vidioc_enum_framesizes = unicam_mc_enum_framesizes,
+ .vidioc_reqbufs = vb2_ioctl_reqbufs,
+ .vidioc_create_bufs = vb2_ioctl_create_bufs,
+ .vidioc_prepare_buf = vb2_ioctl_prepare_buf,
+ .vidioc_querybuf = vb2_ioctl_querybuf,
+ .vidioc_qbuf = vb2_ioctl_qbuf,
+ .vidioc_dqbuf = vb2_ioctl_dqbuf,
+ .vidioc_expbuf = vb2_ioctl_expbuf,
+ .vidioc_streamon = vb2_ioctl_streamon,
+ .vidioc_streamoff = vb2_ioctl_streamoff,
+
+ .vidioc_log_status = unicam_log_status,
+ .vidioc_subscribe_event = unicam_subscribe_event,
+ .vidioc_unsubscribe_event = v4l2_event_unsubscribe,
+};
+
+static int
+unicam_mc_subdev_link_validate_get_format(struct media_pad *pad,
+ struct v4l2_subdev_format *fmt)
+{
+ if (is_media_entity_v4l2_subdev(pad->entity)) {
+ struct v4l2_subdev *sd =
+ media_entity_to_v4l2_subdev(pad->entity);
+
+ fmt->which = V4L2_SUBDEV_FORMAT_ACTIVE;
+ fmt->pad = pad->index;
+ return v4l2_subdev_call(sd, pad, get_fmt, NULL, fmt);
+ }
+
+ return -EINVAL;
+}
+
+static int unicam_mc_video_link_validate(struct media_link *link)
+{
+ struct video_device *vd = container_of(link->sink->entity,
+ struct video_device, entity);
+ struct unicam_node *node = container_of(vd, struct unicam_node,
+ video_dev);
+ struct unicam_device *unicam = node->dev;
+ struct v4l2_subdev_format source_fmt;
int ret;
- v4l2_std_id current_std;
- ret = v4l2_subdev_call(dev->sensor, video, g_std, ¤t_std);
- if (ret)
- return ret;
+ if (!media_entity_remote_pad(link->sink->entity->pads)) {
+ unicam_dbg(1, unicam,
+ "video node %s pad not connected\n", vd->name);
+ return -ENOTCONN;
+ }
- if (std == current_std)
+ ret = unicam_mc_subdev_link_validate_get_format(link->source,
+ &source_fmt);
+ if (ret < 0)
return 0;
- if (vb2_is_busy(&node->buffer_queue))
- return -EBUSY;
+ if (node->pad_id == IMAGE_PAD) {
+ struct v4l2_pix_format *pix_fmt = &node->v_fmt.fmt.pix;
+ const struct unicam_fmt *fmt;
- ret = v4l2_subdev_call(dev->sensor, video, s_std, std);
+ if (source_fmt.format.width != pix_fmt->width ||
+ source_fmt.format.height != pix_fmt->height) {
+ unicam_err(unicam,
+ "Wrong width or height %ux%u (remote pad set to %ux%u)\n",
+ pix_fmt->width, pix_fmt->height,
+ source_fmt.format.width,
+ source_fmt.format.height);
+ return -EINVAL;
+ }
+
+ fmt = find_format_by_code(source_fmt.format.code);
+
+ if (!fmt || (fmt->fourcc != pix_fmt->pixelformat &&
+ fmt->repacked_fourcc != pix_fmt->pixelformat))
+ return -EINVAL;
+ } else {
+ struct v4l2_meta_format *meta_fmt = &node->v_fmt.fmt.meta;
+
+ if (source_fmt.format.width != meta_fmt->buffersize ||
+ source_fmt.format.height != 1 ||
+ source_fmt.format.code != MEDIA_BUS_FMT_SENSOR_DATA) {
+ unicam_err(unicam,
+ "Wrong metadata width/height/code %ux%u %08x (remote pad set to %ux%u %08x)\n",
+ meta_fmt->buffersize, 1,
+ MEDIA_BUS_FMT_SENSOR_DATA,
+ source_fmt.format.width,
+ source_fmt.format.height,
+ source_fmt.format.code);
+ return -EINVAL;
+ }
+ }
+
+ return 0;
+}
+
+static const struct media_entity_operations unicam_mc_entity_ops = {
+ .link_validate = unicam_mc_video_link_validate,
+};
+
+/* videobuf2 Operations */
+
+static int unicam_queue_setup(struct vb2_queue *vq,
+ unsigned int *nbuffers,
+ unsigned int *nplanes,
+ unsigned int sizes[],
+ struct device *alloc_devs[])
+{
+ struct unicam_node *node = vb2_get_drv_priv(vq);
+ struct unicam_device *dev = node->dev;
+ unsigned int size = node->pad_id == IMAGE_PAD ?
+ node->v_fmt.fmt.pix.sizeimage :
+ node->v_fmt.fmt.meta.buffersize;
+
+ if (vq->num_buffers + *nbuffers < 3)
+ *nbuffers = 3 - vq->num_buffers;
+
+ if (*nplanes) {
+ if (sizes[0] < size) {
+ unicam_err(dev, "sizes[0] %i < size %u\n", sizes[0],
+ size);
+ return -EINVAL;
+ }
+ size = sizes[0];
+ }
+
+ *nplanes = 1;
+ sizes[0] = size;
+
+ return 0;
+}
+
+static int unicam_buffer_prepare(struct vb2_buffer *vb)
+{
+ struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue);
+ struct unicam_device *dev = node->dev;
+ struct unicam_buffer *buf = to_unicam_buffer(vb);
+ unsigned long size;
+
+ if (WARN_ON(!node->fmt))
+ return -EINVAL;
+
+ size = node->pad_id == IMAGE_PAD ? node->v_fmt.fmt.pix.sizeimage :
+ node->v_fmt.fmt.meta.buffersize;
+ if (vb2_plane_size(vb, 0) < size) {
+ unicam_err(dev, "data will not fit into plane (%lu < %lu)\n",
+ vb2_plane_size(vb, 0), size);
+ return -EINVAL;
+ }
+
+ vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size);
+ return 0;
+}
+
+static void unicam_buffer_queue(struct vb2_buffer *vb)
+{
+ struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue);
+ struct unicam_buffer *buf = to_unicam_buffer(vb);
+ unsigned long flags;
+
+ spin_lock_irqsave(&node->dma_queue_lock, flags);
+ list_add_tail(&buf->list, &node->dma_queue);
+ spin_unlock_irqrestore(&node->dma_queue_lock, flags);
+}
+
+static void unicam_set_packing_config(struct unicam_device *dev)
+{
+ u32 pack, unpack;
+ u32 val;
+
+ if (dev->node[IMAGE_PAD].v_fmt.fmt.pix.pixelformat ==
+ dev->node[IMAGE_PAD].fmt->fourcc) {
+ unpack = UNICAM_PUM_NONE;
+ pack = UNICAM_PPM_NONE;
+ } else {
+ switch (dev->node[IMAGE_PAD].fmt->depth) {
+ case 8:
+ unpack = UNICAM_PUM_UNPACK8;
+ break;
+ case 10:
+ unpack = UNICAM_PUM_UNPACK10;
+ break;
+ case 12:
+ unpack = UNICAM_PUM_UNPACK12;
+ break;
+ case 14:
+ unpack = UNICAM_PUM_UNPACK14;
+ break;
+ case 16:
+ unpack = UNICAM_PUM_UNPACK16;
+ break;
+ default:
+ unpack = UNICAM_PUM_NONE;
+ break;
+ }
+
+ /* Repacking is always to 16bpp */
+ pack = UNICAM_PPM_PACK16;
+ }
+
+ val = 0;
+ set_field(&val, unpack, UNICAM_PUM_MASK);
+ set_field(&val, pack, UNICAM_PPM_MASK);
+ reg_write(dev, UNICAM_IPIPE, val);
+}
+
+static void unicam_cfg_image_id(struct unicam_device *dev)
+{
+ if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
+ /* CSI2 mode, hardcode VC 0 for now. */
+ reg_write(dev, UNICAM_IDI0,
+ (0 << 6) | dev->node[IMAGE_PAD].fmt->csi_dt);
+ } else {
+ /* CCP2 mode */
+ reg_write(dev, UNICAM_IDI0,
+ 0x80 | dev->node[IMAGE_PAD].fmt->csi_dt);
+ }
+}
+
+static void unicam_enable_ed(struct unicam_device *dev)
+{
+ u32 val = reg_read(dev, UNICAM_DCS);
+
+ set_field(&val, 2, UNICAM_EDL_MASK);
+ /* Do not wrap at the end of the embedded data buffer */
+ set_field(&val, 0, UNICAM_DBOB);
+
+ reg_write(dev, UNICAM_DCS, val);
+}
+
+static void unicam_start_rx(struct unicam_device *dev, dma_addr_t *addr)
+{
+ int line_int_freq = dev->node[IMAGE_PAD].v_fmt.fmt.pix.height >> 2;
+ unsigned int size, i;
+ u32 val;
+
+ if (line_int_freq < 128)
+ line_int_freq = 128;
+
+ /* Enable lane clocks */
+ val = 1;
+ for (i = 0; i < dev->active_data_lanes; i++)
+ val = val << 2 | 1;
+ clk_write(dev, val);
+
+ /* Basic init */
+ reg_write(dev, UNICAM_CTRL, UNICAM_MEM);
+
+ /* Enable analogue control, and leave in reset. */
+ val = UNICAM_AR;
+ set_field(&val, 7, UNICAM_CTATADJ_MASK);
+ set_field(&val, 7, UNICAM_PTATADJ_MASK);
+ reg_write(dev, UNICAM_ANA, val);
+ usleep_range(1000, 2000);
+
+ /* Come out of reset */
+ reg_write_field(dev, UNICAM_ANA, 0, UNICAM_AR);
+
+ /* Peripheral reset */
+ reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPR);
+ reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPR);
+
+ reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPE);
+
+ /* Enable Rx control. */
+ val = reg_read(dev, UNICAM_CTRL);
+ if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
+ set_field(&val, UNICAM_CPM_CSI2, UNICAM_CPM_MASK);
+ set_field(&val, UNICAM_DCM_STROBE, UNICAM_DCM_MASK);
+ } else {
+ set_field(&val, UNICAM_CPM_CCP2, UNICAM_CPM_MASK);
+ set_field(&val, dev->bus_flags, UNICAM_DCM_MASK);
+ }
+ /* Packet framer timeout */
+ set_field(&val, 0xf, UNICAM_PFT_MASK);
+ set_field(&val, 128, UNICAM_OET_MASK);
+ reg_write(dev, UNICAM_CTRL, val);
+
+ reg_write(dev, UNICAM_IHWIN, 0);
+ reg_write(dev, UNICAM_IVWIN, 0);
+
+ /* AXI bus access QoS setup */
+ val = reg_read(dev, UNICAM_PRI);
+ set_field(&val, 0, UNICAM_BL_MASK);
+ set_field(&val, 0, UNICAM_BS_MASK);
+ set_field(&val, 0xe, UNICAM_PP_MASK);
+ set_field(&val, 8, UNICAM_NP_MASK);
+ set_field(&val, 2, UNICAM_PT_MASK);
+ set_field(&val, 1, UNICAM_PE);
+ reg_write(dev, UNICAM_PRI, val);
+
+ reg_write_field(dev, UNICAM_ANA, 0, UNICAM_DDL);
+
+ /* Always start in trigger frame capture mode (UNICAM_FCM set) */
+ val = UNICAM_FSIE | UNICAM_FEIE | UNICAM_FCM | UNICAM_IBOB;
+ set_field(&val, line_int_freq, UNICAM_LCIE_MASK);
+ reg_write(dev, UNICAM_ICTL, val);
+ reg_write(dev, UNICAM_STA, UNICAM_STA_MASK_ALL);
+ reg_write(dev, UNICAM_ISTA, UNICAM_ISTA_MASK_ALL);
+
+ /* tclk_term_en */
+ reg_write_field(dev, UNICAM_CLT, 2, UNICAM_CLT1_MASK);
+ /* tclk_settle */
+ reg_write_field(dev, UNICAM_CLT, 6, UNICAM_CLT2_MASK);
+ /* td_term_en */
+ reg_write_field(dev, UNICAM_DLT, 2, UNICAM_DLT1_MASK);
+ /* ths_settle */
+ reg_write_field(dev, UNICAM_DLT, 6, UNICAM_DLT2_MASK);
+ /* trx_enable */
+ reg_write_field(dev, UNICAM_DLT, 0, UNICAM_DLT3_MASK);
+
+ reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_SOE);
+
+ /* Packet compare setup - required to avoid missing frame ends */
+ val = 0;
+ set_field(&val, 1, UNICAM_PCE);
+ set_field(&val, 1, UNICAM_GI);
+ set_field(&val, 1, UNICAM_CPH);
+ set_field(&val, 0, UNICAM_PCVC_MASK);
+ set_field(&val, 1, UNICAM_PCDT_MASK);
+ reg_write(dev, UNICAM_CMP0, val);
+
+ /* Enable clock lane and set up terminations */
+ val = 0;
+ if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
+ /* CSI2 */
+ set_field(&val, 1, UNICAM_CLE);
+ set_field(&val, 1, UNICAM_CLLPE);
+ if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) {
+ set_field(&val, 1, UNICAM_CLTRE);
+ set_field(&val, 1, UNICAM_CLHSE);
+ }
+ } else {
+ /* CCP2 */
+ set_field(&val, 1, UNICAM_CLE);
+ set_field(&val, 1, UNICAM_CLHSE);
+ set_field(&val, 1, UNICAM_CLTRE);
+ }
+ reg_write(dev, UNICAM_CLK, val);
+
+ /*
+ * Enable required data lanes with appropriate terminations.
+ * The same value needs to be written to UNICAM_DATn registers for
+ * the active lanes, and 0 for inactive ones.
+ */
+ val = 0;
+ if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
+ /* CSI2 */
+ set_field(&val, 1, UNICAM_DLE);
+ set_field(&val, 1, UNICAM_DLLPE);
+ if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) {
+ set_field(&val, 1, UNICAM_DLTRE);
+ set_field(&val, 1, UNICAM_DLHSE);
+ }
+ } else {
+ /* CCP2 */
+ set_field(&val, 1, UNICAM_DLE);
+ set_field(&val, 1, UNICAM_DLHSE);
+ set_field(&val, 1, UNICAM_DLTRE);
+ }
+ reg_write(dev, UNICAM_DAT0, val);
+
+ if (dev->active_data_lanes == 1)
+ val = 0;
+ reg_write(dev, UNICAM_DAT1, val);
+
+ if (dev->max_data_lanes > 2) {
+ /*
+ * Registers UNICAM_DAT2 and UNICAM_DAT3 only valid if the
+ * instance supports more than 2 data lanes.
+ */
+ if (dev->active_data_lanes == 2)
+ val = 0;
+ reg_write(dev, UNICAM_DAT2, val);
+
+ if (dev->active_data_lanes == 3)
+ val = 0;
+ reg_write(dev, UNICAM_DAT3, val);
+ }
+
+ reg_write(dev, UNICAM_IBLS,
+ dev->node[IMAGE_PAD].v_fmt.fmt.pix.bytesperline);
+ size = dev->node[IMAGE_PAD].v_fmt.fmt.pix.sizeimage;
+ unicam_wr_dma_addr(dev, addr[IMAGE_PAD], size, IMAGE_PAD);
+ unicam_set_packing_config(dev);
+ unicam_cfg_image_id(dev);
+
+ val = reg_read(dev, UNICAM_MISC);
+ set_field(&val, 1, UNICAM_FL0);
+ set_field(&val, 1, UNICAM_FL1);
+ reg_write(dev, UNICAM_MISC, val);
- /* Force recomputation of bytesperline */
- node->v_fmt.fmt.pix.bytesperline = 0;
+ if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data) {
+ size = dev->node[METADATA_PAD].v_fmt.fmt.meta.buffersize;
+ unicam_enable_ed(dev);
+ unicam_wr_dma_addr(dev, addr[METADATA_PAD], size, METADATA_PAD);
+ }
- unicam_reset_format(node);
+ /* Enable peripheral */
+ reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPE);
- return ret;
-}
+ /* Load image pointers */
+ reg_write_field(dev, UNICAM_ICTL, 1, UNICAM_LIP_MASK);
-static int unicam_s_edid(struct file *file, void *priv, struct v4l2_edid *edid)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+ /* Load embedded data buffer pointers if needed */
+ if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data)
+ reg_write_field(dev, UNICAM_DCS, 1, UNICAM_LDP);
- return v4l2_subdev_call(dev->sensor, pad, set_edid, edid);
+ /*
+ * Enable trigger only for the first frame to
+ * sync correctly to the FS from the source.
+ */
+ reg_write_field(dev, UNICAM_ICTL, 1, UNICAM_TFC);
}
-static int unicam_g_edid(struct file *file, void *priv, struct v4l2_edid *edid)
+static void unicam_disable(struct unicam_device *dev)
{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
-
- return v4l2_subdev_call(dev->sensor, pad, get_edid, edid);
-}
+ /* Analogue lane control disable */
+ reg_write_field(dev, UNICAM_ANA, 1, UNICAM_DDL);
-static int unicam_s_selection(struct file *file, void *priv,
- struct v4l2_selection *sel)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- struct v4l2_subdev_selection sdsel = {
- .which = V4L2_SUBDEV_FORMAT_ACTIVE,
- .target = sel->target,
- .flags = sel->flags,
- .r = sel->r,
- };
+ /* Stop the output engine */
+ reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_SOE);
- if (sel->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
- return -EINVAL;
+ /* Disable the data lanes. */
+ reg_write(dev, UNICAM_DAT0, 0);
+ reg_write(dev, UNICAM_DAT1, 0);
- return v4l2_subdev_call(dev->sensor, pad, set_selection, NULL, &sdsel);
-}
+ if (dev->max_data_lanes > 2) {
+ reg_write(dev, UNICAM_DAT2, 0);
+ reg_write(dev, UNICAM_DAT3, 0);
+ }
-static int unicam_g_selection(struct file *file, void *priv,
- struct v4l2_selection *sel)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- struct v4l2_subdev_selection sdsel = {
- .which = V4L2_SUBDEV_FORMAT_ACTIVE,
- .target = sel->target,
- };
- int ret;
+ /* Peripheral reset */
+ reg_write_field(dev, UNICAM_CTRL, 1, UNICAM_CPR);
+ usleep_range(50, 100);
+ reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPR);
- if (sel->type != V4L2_BUF_TYPE_VIDEO_CAPTURE)
- return -EINVAL;
+ /* Disable peripheral */
+ reg_write_field(dev, UNICAM_CTRL, 0, UNICAM_CPE);
- ret = v4l2_subdev_call(dev->sensor, pad, get_selection, NULL, &sdsel);
- if (!ret)
- sel->r = sdsel.r;
+ /* Clear ED setup */
+ reg_write(dev, UNICAM_DCS, 0);
- return ret;
+ /* Disable all lane clocks */
+ clk_write(dev, 0);
}
-static int unicam_enum_framesizes(struct file *file, void *priv,
- struct v4l2_frmsizeenum *fsize)
+static void unicam_return_buffers(struct unicam_node *node,
+ enum vb2_buffer_state state)
{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- const struct unicam_fmt *fmt;
- struct v4l2_subdev_frame_size_enum fse;
- int ret;
+ struct unicam_buffer *buf, *tmp;
+ unsigned long flags;
- /* check for valid format */
- fmt = find_format_by_pix(dev, fsize->pixel_format);
- if (!fmt) {
- unicam_dbg(3, dev, "Invalid pixel code: %x\n",
- fsize->pixel_format);
- return -EINVAL;
+ spin_lock_irqsave(&node->dma_queue_lock, flags);
+ list_for_each_entry_safe(buf, tmp, &node->dma_queue, list) {
+ list_del(&buf->list);
+ vb2_buffer_done(&buf->vb.vb2_buf, state);
}
- fse.code = fmt->code;
-
- fse.which = V4L2_SUBDEV_FORMAT_ACTIVE;
- fse.index = fsize->index;
- fse.pad = node->src_pad_id;
-
- ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_size, NULL, &fse);
- if (ret)
- return ret;
-
- unicam_dbg(1, dev, "%s: index: %d code: %x W:[%d,%d] H:[%d,%d]\n",
- __func__, fse.index, fse.code, fse.min_width, fse.max_width,
- fse.min_height, fse.max_height);
- fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE;
- fsize->discrete.width = fse.max_width;
- fsize->discrete.height = fse.max_height;
+ if (node->cur_frm)
+ vb2_buffer_done(&node->cur_frm->vb.vb2_buf,
+ state);
+ if (node->next_frm && node->cur_frm != node->next_frm)
+ vb2_buffer_done(&node->next_frm->vb.vb2_buf,
+ state);
- return 0;
+ node->cur_frm = NULL;
+ node->next_frm = NULL;
+ spin_unlock_irqrestore(&node->dma_queue_lock, flags);
}
-static int unicam_enum_frameintervals(struct file *file, void *priv,
- struct v4l2_frmivalenum *fival)
+static int unicam_start_streaming(struct vb2_queue *vq, unsigned int count)
{
- struct unicam_node *node = video_drvdata(file);
+ struct unicam_node *node = vb2_get_drv_priv(vq);
struct unicam_device *dev = node->dev;
- const struct unicam_fmt *fmt;
- struct v4l2_subdev_frame_interval_enum fie = {
- .index = fival->index,
- .pad = node->src_pad_id,
- .width = fival->width,
- .height = fival->height,
- .which = V4L2_SUBDEV_FORMAT_ACTIVE,
- };
+ dma_addr_t buffer_addr[MAX_NODES] = { 0 };
+ unsigned long flags;
+ unsigned int i;
int ret;
- fmt = find_format_by_pix(dev, fival->pixel_format);
- if (!fmt)
- return -EINVAL;
-
- fie.code = fmt->code;
- ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_interval,
- NULL, &fie);
- if (ret)
- return ret;
-
- fival->type = V4L2_FRMIVAL_TYPE_DISCRETE;
- fival->discrete = fie.interval;
-
- return 0;
-}
+ node->streaming = true;
+ if (!(dev->node[IMAGE_PAD].open && dev->node[IMAGE_PAD].streaming &&
+ (!dev->node[METADATA_PAD].open ||
+ dev->node[METADATA_PAD].streaming))) {
+ /*
+ * Metadata pad must be enabled before image pad if it is
+ * wanted.
+ */
+ unicam_dbg(3, dev, "Not all nodes are streaming yet.");
+ return 0;
+ }
-static int unicam_g_parm(struct file *file, void *fh, struct v4l2_streamparm *a)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+ dev->sequence = 0;
+ ret = unicam_runtime_get(dev);
+ if (ret < 0) {
+ unicam_dbg(3, dev, "unicam_runtime_get failed\n");
+ goto err_streaming;
+ }
- return v4l2_g_parm_cap(video_devdata(file), dev->sensor, a);
-}
+ ret = media_pipeline_start(&node->video_dev.entity, &node->pipe);
+ if (ret < 0) {
+ unicam_err(dev, "Failed to start media pipeline: %d\n", ret);
+ goto err_pm_put;
+ }
-static int unicam_s_parm(struct file *file, void *fh, struct v4l2_streamparm *a)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+ dev->active_data_lanes = dev->max_data_lanes;
- return v4l2_s_parm_cap(video_devdata(file), dev->sensor, a);
-}
+ if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) {
+ struct v4l2_mbus_config mbus_config = { 0 };
-static int unicam_g_dv_timings(struct file *file, void *priv,
- struct v4l2_dv_timings *timings)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+ ret = v4l2_subdev_call(dev->sensor, pad, get_mbus_config,
+ 0, &mbus_config);
+ if (ret < 0 && ret != -ENOIOCTLCMD) {
+ unicam_dbg(3, dev, "g_mbus_config failed\n");
+ goto error_pipeline;
+ }
- return v4l2_subdev_call(dev->sensor, video, g_dv_timings, timings);
-}
+ dev->active_data_lanes =
+ (mbus_config.flags & V4L2_MBUS_CSI2_LANE_MASK) >>
+ __ffs(V4L2_MBUS_CSI2_LANE_MASK);
+ if (!dev->active_data_lanes)
+ dev->active_data_lanes = dev->max_data_lanes;
+ if (dev->active_data_lanes > dev->max_data_lanes) {
+ unicam_err(dev, "Device has requested %u data lanes, which is >%u configured in DT\n",
+ dev->active_data_lanes,
+ dev->max_data_lanes);
+ ret = -EINVAL;
+ goto error_pipeline;
+ }
+ }
-static int unicam_s_dv_timings(struct file *file, void *priv,
- struct v4l2_dv_timings *timings)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- struct v4l2_dv_timings current_timings;
- int ret;
+ unicam_dbg(1, dev, "Running with %u data lanes\n",
+ dev->active_data_lanes);
- ret = v4l2_subdev_call(dev->sensor, video, g_dv_timings,
- ¤t_timings);
+ dev->vpu_req = clk_request_start(dev->vpu_clock, MIN_VPU_CLOCK_RATE);
+ if (!dev->vpu_req) {
+ unicam_err(dev, "failed to set up VPU clock\n");
+ goto error_pipeline;
+ }
- if (ret < 0)
- return ret;
+ ret = clk_prepare_enable(dev->vpu_clock);
+ if (ret) {
+ unicam_err(dev, "Failed to enable VPU clock: %d\n", ret);
+ goto error_pipeline;
+ }
- if (v4l2_match_dv_timings(timings, ¤t_timings, 0, false))
- return 0;
+ ret = clk_set_rate(dev->clock, 100 * 1000 * 1000);
+ if (ret) {
+ unicam_err(dev, "failed to set up CSI clock\n");
+ goto err_vpu_clock;
+ }
- if (vb2_is_busy(&node->buffer_queue))
- return -EBUSY;
+ ret = clk_prepare_enable(dev->clock);
+ if (ret) {
+ unicam_err(dev, "Failed to enable CSI clock: %d\n", ret);
+ goto err_vpu_clock;
+ }
- ret = v4l2_subdev_call(dev->sensor, video, s_dv_timings, timings);
+ for (i = 0; i < ARRAY_SIZE(dev->node); i++) {
+ struct unicam_buffer *buf;
- /* Force recomputation of bytesperline */
- node->v_fmt.fmt.pix.bytesperline = 0;
+ if (!dev->node[i].streaming)
+ continue;
- unicam_reset_format(node);
+ spin_lock_irqsave(&dev->node[i].dma_queue_lock, flags);
+ buf = list_first_entry(&dev->node[i].dma_queue,
+ struct unicam_buffer, list);
+ dev->node[i].cur_frm = buf;
+ dev->node[i].next_frm = buf;
+ list_del(&buf->list);
+ spin_unlock_irqrestore(&dev->node[i].dma_queue_lock, flags);
- return ret;
-}
+ buffer_addr[i] =
+ vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0);
+ }
-static int unicam_query_dv_timings(struct file *file, void *priv,
- struct v4l2_dv_timings *timings)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
+ unicam_start_rx(dev, buffer_addr);
- return v4l2_subdev_call(dev->sensor, video, query_dv_timings, timings);
-}
+ ret = v4l2_subdev_call(dev->sensor, video, s_stream, 1);
+ if (ret < 0) {
+ unicam_err(dev, "stream on failed in subdev\n");
+ goto err_disable_unicam;
+ }
-static int unicam_enum_dv_timings(struct file *file, void *priv,
- struct v4l2_enum_dv_timings *timings)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- int ret;
+ dev->clocks_enabled = true;
+ return 0;
- timings->pad = node->src_pad_id;
- ret = v4l2_subdev_call(dev->sensor, pad, enum_dv_timings, timings);
- timings->pad = node->pad_id;
+err_disable_unicam:
+ unicam_disable(dev);
+ clk_disable_unprepare(dev->clock);
+err_vpu_clock:
+ clk_request_done(dev->vpu_req);
+ clk_disable_unprepare(dev->vpu_clock);
+error_pipeline:
+ media_pipeline_stop(&node->video_dev.entity);
+err_pm_put:
+ unicam_runtime_put(dev);
+err_streaming:
+ unicam_return_buffers(node, VB2_BUF_STATE_QUEUED);
+ node->streaming = false;
return ret;
}
-static int unicam_dv_timings_cap(struct file *file, void *priv,
- struct v4l2_dv_timings_cap *cap)
+static void unicam_stop_streaming(struct vb2_queue *vq)
{
- struct unicam_node *node = video_drvdata(file);
+ struct unicam_node *node = vb2_get_drv_priv(vq);
struct unicam_device *dev = node->dev;
- int ret;
- cap->pad = node->src_pad_id;
- ret = v4l2_subdev_call(dev->sensor, pad, dv_timings_cap, cap);
- cap->pad = node->pad_id;
-
- return ret;
-}
+ node->streaming = false;
-static int unicam_subscribe_event(struct v4l2_fh *fh,
- const struct v4l2_event_subscription *sub)
-{
- switch (sub->type) {
- case V4L2_EVENT_FRAME_SYNC:
- return v4l2_event_subscribe(fh, sub, 2, NULL);
- case V4L2_EVENT_SOURCE_CHANGE:
- return v4l2_event_subscribe(fh, sub, 4, NULL);
- }
+ if (node->pad_id == IMAGE_PAD) {
+ /*
+ * Stop streaming the sensor and disable the peripheral.
+ * We cannot continue streaming embedded data with the
+ * image pad disabled.
+ */
+ if (v4l2_subdev_call(dev->sensor, video, s_stream, 0) < 0)
+ unicam_err(dev, "stream off failed in subdev\n");
- return v4l2_ctrl_subscribe_event(fh, sub);
-}
+ unicam_disable(dev);
-static int unicam_log_status(struct file *file, void *fh)
-{
- struct unicam_node *node = video_drvdata(file);
- struct unicam_device *dev = node->dev;
- u32 reg;
+ media_pipeline_stop(&node->video_dev.entity);
- /* status for sub devices */
- v4l2_device_call_all(&dev->v4l2_dev, 0, core, log_status);
+ if (dev->clocks_enabled) {
+ clk_request_done(dev->vpu_req);
+ clk_disable_unprepare(dev->vpu_clock);
+ clk_disable_unprepare(dev->clock);
+ dev->clocks_enabled = false;
+ }
+ unicam_runtime_put(dev);
- unicam_info(dev, "-----Receiver status-----\n");
- unicam_info(dev, "V4L2 width/height: %ux%u\n",
- node->v_fmt.fmt.pix.width, node->v_fmt.fmt.pix.height);
- unicam_info(dev, "Mediabus format: %08x\n", node->fmt->code);
- unicam_info(dev, "V4L2 format: %08x\n",
- node->v_fmt.fmt.pix.pixelformat);
- reg = reg_read(dev, UNICAM_IPIPE);
- unicam_info(dev, "Unpacking/packing: %u / %u\n",
- get_field(reg, UNICAM_PUM_MASK),
- get_field(reg, UNICAM_PPM_MASK));
- unicam_info(dev, "----Live data----\n");
- unicam_info(dev, "Programmed stride: %4u\n",
- reg_read(dev, UNICAM_IBLS));
- unicam_info(dev, "Detected resolution: %ux%u\n",
- reg_read(dev, UNICAM_IHSTA),
- reg_read(dev, UNICAM_IVSTA));
- unicam_info(dev, "Write pointer: %08x\n",
- reg_read(dev, UNICAM_IBWP));
+ } else if (node->pad_id == METADATA_PAD) {
+ /*
+ * Allow the hardware to spin in the dummy buffer.
+ * This is only really needed if the embedded data pad is
+ * disabled before the image pad.
+ */
+ unicam_wr_dma_addr(dev, node->dummy_buf_dma_addr,
+ DUMMY_BUF_SIZE, METADATA_PAD);
+ }
- return 0;
+ /* Clear all queued buffers for the node */
+ unicam_return_buffers(node, VB2_BUF_STATE_ERROR);
}
-static void unicam_notify(struct v4l2_subdev *sd,
- unsigned int notification, void *arg)
-{
- struct unicam_device *dev = to_unicam_device(sd->v4l2_dev);
-
- switch (notification) {
- case V4L2_DEVICE_NOTIFY_EVENT:
- v4l2_event_queue(&dev->node[IMAGE_PAD].video_dev, arg);
- break;
- default:
- break;
- }
-}
static const struct vb2_ops unicam_video_qops = {
.wait_prepare = vb2_ops_wait_prepare,
.mmap = vb2_fop_mmap,
};
-/* unicam capture ioctl operations */
-static const struct v4l2_ioctl_ops unicam_ioctl_ops = {
- .vidioc_querycap = unicam_querycap,
- .vidioc_enum_fmt_vid_cap = unicam_enum_fmt_vid_cap,
- .vidioc_g_fmt_vid_cap = unicam_g_fmt_vid_cap,
- .vidioc_s_fmt_vid_cap = unicam_s_fmt_vid_cap,
- .vidioc_try_fmt_vid_cap = unicam_try_fmt_vid_cap,
-
- .vidioc_enum_fmt_meta_cap = unicam_enum_fmt_meta_cap,
- .vidioc_g_fmt_meta_cap = unicam_g_fmt_meta_cap,
- .vidioc_s_fmt_meta_cap = unicam_g_fmt_meta_cap,
- .vidioc_try_fmt_meta_cap = unicam_g_fmt_meta_cap,
-
- .vidioc_enum_input = unicam_enum_input,
- .vidioc_g_input = unicam_g_input,
- .vidioc_s_input = unicam_s_input,
-
- .vidioc_querystd = unicam_querystd,
- .vidioc_s_std = unicam_s_std,
- .vidioc_g_std = unicam_g_std,
-
- .vidioc_g_edid = unicam_g_edid,
- .vidioc_s_edid = unicam_s_edid,
-
- .vidioc_enum_framesizes = unicam_enum_framesizes,
- .vidioc_enum_frameintervals = unicam_enum_frameintervals,
-
- .vidioc_g_selection = unicam_g_selection,
- .vidioc_s_selection = unicam_s_selection,
-
- .vidioc_g_parm = unicam_g_parm,
- .vidioc_s_parm = unicam_s_parm,
-
- .vidioc_s_dv_timings = unicam_s_dv_timings,
- .vidioc_g_dv_timings = unicam_g_dv_timings,
- .vidioc_query_dv_timings = unicam_query_dv_timings,
- .vidioc_enum_dv_timings = unicam_enum_dv_timings,
- .vidioc_dv_timings_cap = unicam_dv_timings_cap,
-
- .vidioc_reqbufs = vb2_ioctl_reqbufs,
- .vidioc_create_bufs = vb2_ioctl_create_bufs,
- .vidioc_prepare_buf = vb2_ioctl_prepare_buf,
- .vidioc_querybuf = vb2_ioctl_querybuf,
- .vidioc_qbuf = vb2_ioctl_qbuf,
- .vidioc_dqbuf = vb2_ioctl_dqbuf,
- .vidioc_expbuf = vb2_ioctl_expbuf,
- .vidioc_streamon = vb2_ioctl_streamon,
- .vidioc_streamoff = vb2_ioctl_streamoff,
-
- .vidioc_log_status = unicam_log_status,
- .vidioc_subscribe_event = unicam_subscribe_event,
- .vidioc_unsubscribe_event = v4l2_event_unsubscribe,
-};
-
static int
unicam_async_bound(struct v4l2_async_notifier *notifier,
struct v4l2_subdev *subdev,
unicam_put(node->dev);
}
-static int register_node(struct unicam_device *unicam, struct unicam_node *node,
- enum v4l2_buf_type type, int pad_id)
+static int unicam_set_default_format(struct unicam_device *unicam,
+ struct unicam_node *node,
+ int pad_id,
+ const struct unicam_fmt **ret_fmt)
{
- struct video_device *vdev;
- struct vb2_queue *q;
struct v4l2_mbus_framefmt mbus_fmt = {0};
const struct unicam_fmt *fmt;
int ret;
node->v_fmt.fmt.meta.dataformat = fmt->fourcc;
}
+ *ret_fmt = fmt;
+
+ return 0;
+}
+
+static void unicam_mc_set_default_format(struct unicam_node *node, int pad_id)
+{
+ if (pad_id == IMAGE_PAD) {
+ struct v4l2_pix_format *pix_fmt = &node->v_fmt.fmt.pix;
+
+ pix_fmt->width = 640;
+ pix_fmt->height = 480;
+ pix_fmt->field = V4L2_FIELD_NONE;
+ pix_fmt->colorspace = V4L2_COLORSPACE_SRGB;
+ pix_fmt->ycbcr_enc = V4L2_YCBCR_ENC_601;
+ pix_fmt->quantization = V4L2_QUANTIZATION_LIM_RANGE;
+ pix_fmt->xfer_func = V4L2_XFER_FUNC_SRGB;
+ pix_fmt->pixelformat = formats[0].fourcc;
+ unicam_calc_format_size_bpl(node->dev, &formats[0],
+ &node->v_fmt);
+ node->v_fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
+
+ node->fmt = &formats[0];
+ } else {
+ const struct unicam_fmt *fmt;
+
+ /* Fix this node format as embedded data. */
+ fmt = find_format_by_code(MEDIA_BUS_FMT_SENSOR_DATA);
+ node->v_fmt.fmt.meta.dataformat = fmt->fourcc;
+ node->fmt = fmt;
+
+ node->v_fmt.fmt.meta.buffersize = UNICAM_EMBEDDED_SIZE;
+ node->embedded_lines = 1;
+ node->v_fmt.type = V4L2_BUF_TYPE_META_CAPTURE;
+ }
+}
+
+static int register_node(struct unicam_device *unicam, struct unicam_node *node,
+ enum v4l2_buf_type type, int pad_id)
+{
+ struct video_device *vdev;
+ struct vb2_queue *q;
+ int ret;
+
node->dev = unicam;
node->pad_id = pad_id;
- node->fmt = fmt;
- /* Read current subdev format */
- if (fmt)
- unicam_reset_format(node);
+ if (!unicam->mc_api) {
+ const struct unicam_fmt *fmt;
+
+ ret = unicam_set_default_format(unicam, node, pad_id, &fmt);
+ if (ret)
+ return ret;
+ node->fmt = fmt;
+ /* Read current subdev format */
+ if (fmt)
+ unicam_reset_format(node);
+ } else {
+ unicam_mc_set_default_format(node, pad_id);
+ }
- if (v4l2_subdev_has_op(unicam->sensor, video, s_std)) {
+ if (!unicam->mc_api &&
+ v4l2_subdev_has_op(unicam->sensor, video, s_std)) {
v4l2_std_id tvnorms;
if (WARN_ON(!v4l2_subdev_has_op(unicam->sensor, video,
vdev = &node->video_dev;
if (pad_id == IMAGE_PAD) {
- /* Add controls from the subdevice */
- ret = v4l2_ctrl_add_handler(&unicam->ctrl_handler,
- unicam->sensor->ctrl_handler, NULL,
- true);
- if (ret < 0)
- return ret;
+ if (!unicam->mc_api) {
+ /* Add controls from the subdevice */
+ ret = v4l2_ctrl_add_handler(&unicam->ctrl_handler,
+ unicam->sensor->ctrl_handler,
+ NULL,
+ true);
+ if (ret < 0)
+ return ret;
+ }
/*
* If the sensor subdevice has any controls, associate the node
vdev->release = unicam_node_release;
vdev->fops = &unicam_fops;
- vdev->ioctl_ops = &unicam_ioctl_ops;
+ vdev->ioctl_ops = unicam->mc_api ? &unicam_mc_ioctl_ops :
+ &unicam_ioctl_ops;
vdev->v4l2_dev = &unicam->v4l2_dev;
vdev->vfl_dir = VFL_DIR_RX;
vdev->queue = q;
vdev->device_caps = (pad_id == IMAGE_PAD) ?
V4L2_CAP_VIDEO_CAPTURE : V4L2_CAP_META_CAPTURE;
vdev->device_caps |= V4L2_CAP_READWRITE | V4L2_CAP_STREAMING;
+ if (unicam->mc_api) {
+ vdev->device_caps |= V4L2_CAP_IO_MC;
+ vdev->entity.ops = &unicam_mc_entity_ops;
+ }
/* Define the device names */
snprintf(vdev->name, sizeof(vdev->name), "%s-%s", UNICAM_MODULE_NAME,
unicam_err(unicam, "Unable to allocate dummy buffer.\n");
return -ENOMEM;
}
-
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, video, s_std)) {
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_STD);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_STD);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUMSTD);
- }
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, video, querystd))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_QUERYSTD);
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, video, s_dv_timings)) {
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_EDID);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_EDID);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_DV_TIMINGS_CAP);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_DV_TIMINGS);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_DV_TIMINGS);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUM_DV_TIMINGS);
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_QUERY_DV_TIMINGS);
- }
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, pad, enum_frame_interval))
- v4l2_disable_ioctl(&node->video_dev,
- VIDIOC_ENUM_FRAMEINTERVALS);
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, video, g_frame_interval))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_PARM);
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, video, s_frame_interval))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_PARM);
-
- if (pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, pad, enum_frame_size))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUM_FRAMESIZES);
-
- if (node->pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, pad, set_selection))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_SELECTION);
-
- if (node->pad_id == METADATA_PAD ||
- !v4l2_subdev_has_op(unicam->sensor, pad, get_selection))
- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_SELECTION);
+ if (!unicam->mc_api) {
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, video, s_std)) {
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_STD);
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_STD);
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUMSTD);
+ }
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, video, querystd))
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_QUERYSTD);
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, video, s_dv_timings)) {
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_EDID);
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_EDID);
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_DV_TIMINGS_CAP);
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_G_DV_TIMINGS);
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_S_DV_TIMINGS);
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_ENUM_DV_TIMINGS);
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_QUERY_DV_TIMINGS);
+ }
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, pad,
+ enum_frame_interval))
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_ENUM_FRAMEINTERVALS);
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, video,
+ g_frame_interval))
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_PARM);
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, video,
+ s_frame_interval))
+ v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_PARM);
+
+ if (pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, pad,
+ enum_frame_size))
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_ENUM_FRAMESIZES);
+
+ if (node->pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, pad, set_selection))
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_S_SELECTION);
+
+ if (node->pad_id == METADATA_PAD ||
+ !v4l2_subdev_has_op(unicam->sensor, pad, get_selection))
+ v4l2_disable_ioctl(&node->video_dev,
+ VIDIOC_G_SELECTION);
+ }
ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1);
if (ret) {
if (unicam->sensor->entity.pads[i].flags & MEDIA_PAD_FL_SOURCE) {
if (source_pads < MAX_NODES) {
unicam->node[source_pads].src_pad_id = i;
- unicam_err(unicam, "source pad %u is index %u\n",
+ unicam_dbg(3, unicam, "source pad %u is index %u\n",
source_pads, i);
}
source_pads++;
}
}
- ret = v4l2_device_register_ro_subdev_nodes(&unicam->v4l2_dev);
+ if (unicam->mc_api)
+ ret = v4l2_device_register_subdev_nodes(&unicam->v4l2_dev);
+ else
+ ret = v4l2_device_register_ro_subdev_nodes(&unicam->v4l2_dev);
if (ret) {
unicam_err(unicam, "Unable to register subdev nodes.\n");
goto unregister;
kref_init(&unicam->kref);
unicam->pdev = pdev;
+ /*
+ * Adopt the current setting of the module parameter, and check if
+ * device tree requests it.
+ */
+ unicam->mc_api = media_controller;
+ if (of_property_read_bool(pdev->dev.of_node, "brcm,media-controller"))
+ unicam->mc_api = true;
+
unicam->base = devm_platform_ioremap_resource(pdev, 0);
if (IS_ERR(unicam->base)) {
unicam_err(unicam, "Failed to get main io block\n");