diff options
Diffstat (limited to 'target/linux/bcm27xx/patches-5.15/950-0245-drivers-media-Remove-the-downstream-version-of-bcm28.patch')
-rw-r--r-- | target/linux/bcm27xx/patches-5.15/950-0245-drivers-media-Remove-the-downstream-version-of-bcm28.patch | 3175 |
1 files changed, 3175 insertions, 0 deletions
diff --git a/target/linux/bcm27xx/patches-5.15/950-0245-drivers-media-Remove-the-downstream-version-of-bcm28.patch b/target/linux/bcm27xx/patches-5.15/950-0245-drivers-media-Remove-the-downstream-version-of-bcm28.patch new file mode 100644 index 0000000000..36893a99ea --- /dev/null +++ b/target/linux/bcm27xx/patches-5.15/950-0245-drivers-media-Remove-the-downstream-version-of-bcm28.patch @@ -0,0 +1,3175 @@ +From 606a02e0617f3adbcdd7aa82913771403d137269 Mon Sep 17 00:00:00 2001 +From: Dave Stevenson <dave.stevenson@raspberrypi.com> +Date: Tue, 23 Jun 2020 10:05:57 +0100 +Subject: [PATCH] drivers: media: Remove the downstream version of + bcm2835-unicam + +About to be replaced by the upstream version. + +Signed-off-by: Dave Stevenson <dave.stevenson@raspberrypi.com> +--- + drivers/media/platform/bcm2835/Kconfig | 14 - + drivers/media/platform/bcm2835/Makefile | 3 - + .../media/platform/bcm2835/bcm2835-unicam.c | 2873 ----------------- + .../media/platform/bcm2835/vc4-regs-unicam.h | 253 -- + 4 files changed, 3143 deletions(-) + delete mode 100644 drivers/media/platform/bcm2835/Kconfig + delete mode 100644 drivers/media/platform/bcm2835/Makefile + delete mode 100644 drivers/media/platform/bcm2835/bcm2835-unicam.c + delete mode 100644 drivers/media/platform/bcm2835/vc4-regs-unicam.h + +--- a/drivers/media/platform/bcm2835/Kconfig ++++ /dev/null +@@ -1,14 +0,0 @@ +-# Broadcom VideoCore4 V4L2 camera support +- +-config VIDEO_BCM2835_UNICAM +- tristate "Broadcom BCM2835 Unicam video capture driver" +- depends on VIDEO_V4L2 && VIDEO_V4L2_SUBDEV_API && MEDIA_CONTROLLER +- depends on ARCH_BCM2835 || COMPILE_TEST +- select VIDEOBUF2_DMA_CONTIG +- select V4L2_FWNODE +- help +- Say Y here to enable V4L2 subdevice for CSI2 receiver. +- This is a V4L2 subdevice that interfaces directly to the VC4 peripheral. +- +- To compile this driver as a module, choose M here. The module +- will be called bcm2835-unicam. +--- a/drivers/media/platform/bcm2835/Makefile ++++ /dev/null +@@ -1,3 +0,0 @@ +-# Makefile for BCM2835 Unicam driver +- +-obj-$(CONFIG_VIDEO_BCM2835_UNICAM) += bcm2835-unicam.o +--- a/drivers/media/platform/bcm2835/bcm2835-unicam.c ++++ /dev/null +@@ -1,2873 +0,0 @@ +-// SPDX-License-Identifier: GPL-2.0-only +-/* +- * BCM2835 Unicam Capture Driver +- * +- * Copyright (C) 2017-2020 - Raspberry Pi (Trading) Ltd. +- * +- * Dave Stevenson <dave.stevenson@raspberrypi.com> +- * +- * Based on TI am437x driver by +- * Benoit Parrot <bparrot@ti.com> +- * Lad, Prabhakar <prabhakar.csengg@gmail.com> +- * +- * and TI CAL camera interface driver by +- * Benoit Parrot <bparrot@ti.com> +- * +- * +- * There are two camera drivers in the kernel for BCM283x - this one +- * and bcm2835-camera (currently in staging). +- * +- * This driver directly controls the Unicam peripheral - there is no +- * involvement with the VideoCore firmware. Unicam receives CSI-2 or +- * CCP2 data and writes it into SDRAM. +- * The only potential processing options are to repack Bayer data into an +- * alternate format, and applying windowing. +- * The repacking does not shift the data, so can repack V4L2_PIX_FMT_Sxxxx10P +- * to V4L2_PIX_FMT_Sxxxx10, or V4L2_PIX_FMT_Sxxxx12P to V4L2_PIX_FMT_Sxxxx12, +- * but not generically up to V4L2_PIX_FMT_Sxxxx16. The driver will add both +- * formats where the relevant formats are defined, and will automatically +- * configure the repacking as required. +- * Support for windowing may be added later. +- * +- * It should be possible to connect this driver to any sensor with a +- * suitable output interface and V4L2 subdevice driver. +- * +- * bcm2835-camera uses the VideoCore firmware to control the sensor, +- * Unicam, ISP, and all tuner control loops. Fully processed frames are +- * delivered to the driver by the firmware. It only has sensor drivers +- * for Omnivision OV5647, and Sony IMX219 sensors. +- * +- * The two drivers are mutually exclusive for the same Unicam instance. +- * The VideoCore firmware checks the device tree configuration during boot. +- * If it finds device tree nodes called csi0 or csi1 it will block the +- * firmware from accessing the peripheral, and bcm2835-camera will +- * not be able to stream data. +- */ +- +-#include <linux/clk.h> +-#include <linux/delay.h> +-#include <linux/device.h> +-#include <linux/dma-mapping.h> +-#include <linux/err.h> +-#include <linux/init.h> +-#include <linux/interrupt.h> +-#include <linux/io.h> +-#include <linux/module.h> +-#include <linux/of_device.h> +-#include <linux/of_graph.h> +-#include <linux/pinctrl/consumer.h> +-#include <linux/platform_device.h> +-#include <linux/pm_runtime.h> +-#include <linux/slab.h> +-#include <linux/uaccess.h> +-#include <linux/videodev2.h> +- +-#include <media/v4l2-common.h> +-#include <media/v4l2-ctrls.h> +-#include <media/v4l2-dev.h> +-#include <media/v4l2-device.h> +-#include <media/v4l2-dv-timings.h> +-#include <media/v4l2-event.h> +-#include <media/v4l2-ioctl.h> +-#include <media/v4l2-fwnode.h> +-#include <media/videobuf2-dma-contig.h> +- +-#include "vc4-regs-unicam.h" +- +-#define UNICAM_MODULE_NAME "unicam" +-#define UNICAM_VERSION "0.1.0" +- +-static int debug; +-module_param(debug, int, 0644); +-MODULE_PARM_DESC(debug, "Debug level 0-3"); +- +-#define unicam_dbg(level, dev, fmt, arg...) \ +- v4l2_dbg(level, debug, &(dev)->v4l2_dev, fmt, ##arg) +-#define unicam_info(dev, fmt, arg...) \ +- v4l2_info(&(dev)->v4l2_dev, fmt, ##arg) +-#define unicam_err(dev, fmt, arg...) \ +- v4l2_err(&(dev)->v4l2_dev, fmt, ##arg) +- +-/* To protect against a dodgy sensor driver never returning an error from +- * enum_mbus_code, set a maximum index value to be used. +- */ +-#define MAX_ENUM_MBUS_CODE 128 +- +-/* +- * Stride is a 16 bit register, but also has to be a multiple of 32. +- */ +-#define BPL_ALIGNMENT 32 +-#define MAX_BYTESPERLINE ((1 << 16) - BPL_ALIGNMENT) +-/* +- * Max width is therefore determined by the max stride divided by +- * the number of bits per pixel. Take 32bpp as a +- * worst case. +- * No imposed limit on the height, so adopt a square image for want +- * of anything better. +- */ +-#define MAX_WIDTH (MAX_BYTESPERLINE / 4) +-#define MAX_HEIGHT MAX_WIDTH +-/* Define a nominal minimum image size */ +-#define MIN_WIDTH 16 +-#define MIN_HEIGHT 16 +-/* Default size of the embedded buffer */ +-#define UNICAM_EMBEDDED_SIZE 8192 +- +-/* +- * Size of the dummy buffer. Can be any size really, but the DMA +- * allocation works in units of page sizes. +- */ +-#define DUMMY_BUF_SIZE (PAGE_SIZE) +- +-enum pad_types { +- IMAGE_PAD, +- METADATA_PAD, +- MAX_NODES +-}; +- +-/* +- * struct unicam_fmt - Unicam media bus format information +- * @pixelformat: V4L2 pixel format FCC identifier. 0 if n/a. +- * @repacked_fourcc: V4L2 pixel format FCC identifier if the data is expanded +- * out to 16bpp. 0 if n/a. +- * @code: V4L2 media bus format code. +- * @depth: Bits per pixel as delivered from the source. +- * @csi_dt: CSI data type. +- * @check_variants: Flag to denote that there are multiple mediabus formats +- * still in the list that could match this V4L2 format. +- */ +-struct unicam_fmt { +- u32 fourcc; +- u32 repacked_fourcc; +- u32 code; +- u8 depth; +- u8 csi_dt; +- u8 check_variants; +-}; +- +-static const struct unicam_fmt formats[] = { +- /* YUV Formats */ +- { +- .fourcc = V4L2_PIX_FMT_YUYV, +- .code = MEDIA_BUS_FMT_YUYV8_2X8, +- .depth = 16, +- .csi_dt = 0x1e, +- .check_variants = 1, +- }, { +- .fourcc = V4L2_PIX_FMT_UYVY, +- .code = MEDIA_BUS_FMT_UYVY8_2X8, +- .depth = 16, +- .csi_dt = 0x1e, +- .check_variants = 1, +- }, { +- .fourcc = V4L2_PIX_FMT_YVYU, +- .code = MEDIA_BUS_FMT_YVYU8_2X8, +- .depth = 16, +- .csi_dt = 0x1e, +- .check_variants = 1, +- }, { +- .fourcc = V4L2_PIX_FMT_VYUY, +- .code = MEDIA_BUS_FMT_VYUY8_2X8, +- .depth = 16, +- .csi_dt = 0x1e, +- .check_variants = 1, +- }, { +- .fourcc = V4L2_PIX_FMT_YUYV, +- .code = MEDIA_BUS_FMT_YUYV8_1X16, +- .depth = 16, +- .csi_dt = 0x1e, +- }, { +- .fourcc = V4L2_PIX_FMT_UYVY, +- .code = MEDIA_BUS_FMT_UYVY8_1X16, +- .depth = 16, +- .csi_dt = 0x1e, +- }, { +- .fourcc = V4L2_PIX_FMT_YVYU, +- .code = MEDIA_BUS_FMT_YVYU8_1X16, +- .depth = 16, +- .csi_dt = 0x1e, +- }, { +- .fourcc = V4L2_PIX_FMT_VYUY, +- .code = MEDIA_BUS_FMT_VYUY8_1X16, +- .depth = 16, +- .csi_dt = 0x1e, +- }, { +- /* RGB Formats */ +- .fourcc = V4L2_PIX_FMT_RGB565, /* gggbbbbb rrrrrggg */ +- .code = MEDIA_BUS_FMT_RGB565_2X8_LE, +- .depth = 16, +- .csi_dt = 0x22, +- }, { +- .fourcc = V4L2_PIX_FMT_RGB565X, /* rrrrrggg gggbbbbb */ +- .code = MEDIA_BUS_FMT_RGB565_2X8_BE, +- .depth = 16, +- .csi_dt = 0x22 +- }, { +- .fourcc = V4L2_PIX_FMT_RGB555, /* gggbbbbb arrrrrgg */ +- .code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_LE, +- .depth = 16, +- .csi_dt = 0x21, +- }, { +- .fourcc = V4L2_PIX_FMT_RGB555X, /* arrrrrgg gggbbbbb */ +- .code = MEDIA_BUS_FMT_RGB555_2X8_PADHI_BE, +- .depth = 16, +- .csi_dt = 0x21, +- }, { +- .fourcc = V4L2_PIX_FMT_RGB24, /* rgb */ +- .code = MEDIA_BUS_FMT_RGB888_1X24, +- .depth = 24, +- .csi_dt = 0x24, +- }, { +- .fourcc = V4L2_PIX_FMT_BGR24, /* bgr */ +- .code = MEDIA_BUS_FMT_BGR888_1X24, +- .depth = 24, +- .csi_dt = 0x24, +- }, { +- .fourcc = V4L2_PIX_FMT_RGB32, /* argb */ +- .code = MEDIA_BUS_FMT_ARGB8888_1X32, +- .depth = 32, +- .csi_dt = 0x0, +- }, { +- /* Bayer Formats */ +- .fourcc = V4L2_PIX_FMT_SBGGR8, +- .code = MEDIA_BUS_FMT_SBGGR8_1X8, +- .depth = 8, +- .csi_dt = 0x2a, +- }, { +- .fourcc = V4L2_PIX_FMT_SGBRG8, +- .code = MEDIA_BUS_FMT_SGBRG8_1X8, +- .depth = 8, +- .csi_dt = 0x2a, +- }, { +- .fourcc = V4L2_PIX_FMT_SGRBG8, +- .code = MEDIA_BUS_FMT_SGRBG8_1X8, +- .depth = 8, +- .csi_dt = 0x2a, +- }, { +- .fourcc = V4L2_PIX_FMT_SRGGB8, +- .code = MEDIA_BUS_FMT_SRGGB8_1X8, +- .depth = 8, +- .csi_dt = 0x2a, +- }, { +- .fourcc = V4L2_PIX_FMT_SBGGR10P, +- .repacked_fourcc = V4L2_PIX_FMT_SBGGR10, +- .code = MEDIA_BUS_FMT_SBGGR10_1X10, +- .depth = 10, +- .csi_dt = 0x2b, +- }, { +- .fourcc = V4L2_PIX_FMT_SGBRG10P, +- .repacked_fourcc = V4L2_PIX_FMT_SGBRG10, +- .code = MEDIA_BUS_FMT_SGBRG10_1X10, +- .depth = 10, +- .csi_dt = 0x2b, +- }, { +- .fourcc = V4L2_PIX_FMT_SGRBG10P, +- .repacked_fourcc = V4L2_PIX_FMT_SGRBG10, +- .code = MEDIA_BUS_FMT_SGRBG10_1X10, +- .depth = 10, +- .csi_dt = 0x2b, +- }, { +- .fourcc = V4L2_PIX_FMT_SRGGB10P, +- .repacked_fourcc = V4L2_PIX_FMT_SRGGB10, +- .code = MEDIA_BUS_FMT_SRGGB10_1X10, +- .depth = 10, +- .csi_dt = 0x2b, +- }, { +- .fourcc = V4L2_PIX_FMT_SBGGR12P, +- .repacked_fourcc = V4L2_PIX_FMT_SBGGR12, +- .code = MEDIA_BUS_FMT_SBGGR12_1X12, +- .depth = 12, +- .csi_dt = 0x2c, +- }, { +- .fourcc = V4L2_PIX_FMT_SGBRG12P, +- .repacked_fourcc = V4L2_PIX_FMT_SGBRG12, +- .code = MEDIA_BUS_FMT_SGBRG12_1X12, +- .depth = 12, +- .csi_dt = 0x2c, +- }, { +- .fourcc = V4L2_PIX_FMT_SGRBG12P, +- .repacked_fourcc = V4L2_PIX_FMT_SGRBG12, +- .code = MEDIA_BUS_FMT_SGRBG12_1X12, +- .depth = 12, +- .csi_dt = 0x2c, +- }, { +- .fourcc = V4L2_PIX_FMT_SRGGB12P, +- .repacked_fourcc = V4L2_PIX_FMT_SRGGB12, +- .code = MEDIA_BUS_FMT_SRGGB12_1X12, +- .depth = 12, +- .csi_dt = 0x2c, +- }, { +- .fourcc = V4L2_PIX_FMT_SBGGR14P, +- .code = MEDIA_BUS_FMT_SBGGR14_1X14, +- .depth = 14, +- .csi_dt = 0x2d, +- }, { +- .fourcc = V4L2_PIX_FMT_SGBRG14P, +- .code = MEDIA_BUS_FMT_SGBRG14_1X14, +- .depth = 14, +- .csi_dt = 0x2d, +- }, { +- .fourcc = V4L2_PIX_FMT_SGRBG14P, +- .code = MEDIA_BUS_FMT_SGRBG14_1X14, +- .depth = 14, +- .csi_dt = 0x2d, +- }, { +- .fourcc = V4L2_PIX_FMT_SRGGB14P, +- .code = MEDIA_BUS_FMT_SRGGB14_1X14, +- .depth = 14, +- .csi_dt = 0x2d, +- }, { +- /* +- * 16 bit Bayer formats could be supported, but there is no CSI2 +- * data_type defined for raw 16, and no sensors that produce it at +- * present. +- */ +- +- /* Greyscale formats */ +- .fourcc = V4L2_PIX_FMT_GREY, +- .code = MEDIA_BUS_FMT_Y8_1X8, +- .depth = 8, +- .csi_dt = 0x2a, +- }, { +- .fourcc = V4L2_PIX_FMT_Y10P, +- .repacked_fourcc = V4L2_PIX_FMT_Y10, +- .code = MEDIA_BUS_FMT_Y10_1X10, +- .depth = 10, +- .csi_dt = 0x2b, +- }, { +- /* NB There is no packed V4L2 fourcc for this format. */ +- .repacked_fourcc = V4L2_PIX_FMT_Y12, +- .code = MEDIA_BUS_FMT_Y12_1X12, +- .depth = 12, +- .csi_dt = 0x2c, +- }, +- /* Embedded data format */ +- { +- .fourcc = V4L2_META_FMT_SENSOR_DATA, +- .code = MEDIA_BUS_FMT_SENSOR_DATA, +- .depth = 8, +- } +-}; +- +-struct unicam_dmaqueue { +- struct list_head active; +-}; +- +-struct unicam_buffer { +- struct vb2_v4l2_buffer vb; +- struct list_head list; +-}; +- +-struct unicam_cfg { +- /* peripheral base address */ +- void __iomem *base; +- /* clock gating base address */ +- void __iomem *clk_gate_base; +-}; +- +-#define MAX_POSSIBLE_PIX_FMTS (ARRAY_SIZE(formats)) +- +-struct unicam_node { +- int registered; +- int open; +- int streaming; +- unsigned int pad_id; +- /* Pointer pointing to current v4l2_buffer */ +- struct unicam_buffer *cur_frm; +- /* Pointer pointing to next v4l2_buffer */ +- struct unicam_buffer *next_frm; +- /* video capture */ +- const struct unicam_fmt *fmt; +- /* Used to store current pixel format */ +- struct v4l2_format v_fmt; +- /* Used to store current mbus frame format */ +- struct v4l2_mbus_framefmt m_fmt; +- /* Buffer queue used in video-buf */ +- struct vb2_queue buffer_queue; +- /* Queue of filled frames */ +- struct unicam_dmaqueue dma_queue; +- /* IRQ lock for DMA queue */ +- spinlock_t dma_queue_lock; +- /* lock used to access this structure */ +- struct mutex lock; +- /* Identifies video device for this channel */ +- struct video_device video_dev; +- /* Pointer to the parent handle */ +- struct unicam_device *dev; +- struct media_pad pad; +- struct v4l2_ctrl_handler ctrl_handler; +- unsigned int embedded_lines; +- /* +- * Dummy buffer intended to be used by unicam +- * if we have no other queued buffers to swap to. +- */ +- void *dummy_buf_cpu_addr; +- dma_addr_t dummy_buf_dma_addr; +-}; +- +-struct unicam_device { +- /* V4l2 specific parameters */ +- +- struct v4l2_fwnode_endpoint endpoint; +- +- struct v4l2_async_subdev asd; +- +- /* unicam cfg */ +- struct unicam_cfg cfg; +- /* clock handle */ +- struct clk *clock; +- /* V4l2 device */ +- struct v4l2_device v4l2_dev; +- struct media_device mdev; +- +- /* parent device */ +- struct platform_device *pdev; +- /* subdevice async Notifier */ +- struct v4l2_async_notifier notifier; +- unsigned int sequence; +- +- /* ptr to sub device */ +- struct v4l2_subdev *sensor; +- /* Pad config for the sensor */ +- struct v4l2_subdev_pad_config *sensor_config; +- +- unsigned int virtual_channel; +- enum v4l2_mbus_type bus_type; +- /* +- * Stores bus.mipi_csi2.flags for CSI2 sensors, or +- * bus.mipi_csi1.strobe for CCP2. +- */ +- unsigned int bus_flags; +- unsigned int max_data_lanes; +- unsigned int active_data_lanes; +- bool sensor_embedded_data; +- +- struct unicam_node node[MAX_NODES]; +-}; +- +-/* Hardware access */ +-#define clk_write(dev, val) writel((val) | 0x5a000000, (dev)->clk_gate_base) +-#define clk_read(dev) readl((dev)->clk_gate_base) +- +-#define reg_read(dev, offset) readl((dev)->base + (offset)) +-#define reg_write(dev, offset, val) writel(val, (dev)->base + (offset)) +- +-#define reg_read_field(dev, offset, mask) get_field(reg_read((dev), (offset), \ +- mask)) +- +-static inline int get_field(u32 value, u32 mask) +-{ +- return (value & mask) >> __ffs(mask); +-} +- +-static inline void set_field(u32 *valp, u32 field, u32 mask) +-{ +- u32 val = *valp; +- +- val &= ~mask; +- val |= (field << __ffs(mask)) & mask; +- *valp = val; +-} +- +-static inline void reg_write_field(struct unicam_cfg *dev, u32 offset, +- u32 field, u32 mask) +-{ +- u32 val = reg_read((dev), (offset)); +- +- set_field(&val, field, mask); +- reg_write((dev), (offset), val); +-} +- +-/* Power management functions */ +-static inline int unicam_runtime_get(struct unicam_device *dev) +-{ +- return pm_runtime_get_sync(&dev->pdev->dev); +-} +- +-static inline void unicam_runtime_put(struct unicam_device *dev) +-{ +- pm_runtime_put_sync(&dev->pdev->dev); +-} +- +-/* Format setup functions */ +-static const struct unicam_fmt *find_format_by_code(u32 code) +-{ +- unsigned int i; +- +- for (i = 0; i < ARRAY_SIZE(formats); i++) { +- if (formats[i].code == code) +- return &formats[i]; +- } +- +- return NULL; +-} +- +-static int check_mbus_format(struct unicam_device *dev, +- const struct unicam_fmt *format) +-{ +- struct v4l2_subdev_mbus_code_enum mbus_code; +- int ret = 0; +- int i; +- +- for (i = 0; !ret && i < MAX_ENUM_MBUS_CODE; i++) { +- memset(&mbus_code, 0, sizeof(mbus_code)); +- mbus_code.index = i; +- mbus_code.pad = IMAGE_PAD; +- mbus_code.which = V4L2_SUBDEV_FORMAT_ACTIVE; +- +- ret = v4l2_subdev_call(dev->sensor, pad, enum_mbus_code, +- NULL, &mbus_code); +- +- if (!ret && mbus_code.code == format->code) +- return 1; +- } +- +- return 0; +-} +- +-static const struct unicam_fmt *find_format_by_pix(struct unicam_device *dev, +- u32 pixelformat) +-{ +- unsigned int i; +- +- for (i = 0; i < ARRAY_SIZE(formats); i++) { +- if (formats[i].fourcc == pixelformat || +- formats[i].repacked_fourcc == pixelformat) { +- if (formats[i].check_variants && +- !check_mbus_format(dev, &formats[i])) +- continue; +- return &formats[i]; +- } +- } +- +- return NULL; +-} +- +-static inline unsigned int bytes_per_line(u32 width, +- const struct unicam_fmt *fmt, +- u32 v4l2_fourcc) +-{ +- if (v4l2_fourcc == fmt->repacked_fourcc) +- /* Repacking always goes to 16bpp */ +- return ALIGN(width << 1, BPL_ALIGNMENT); +- else +- return ALIGN((width * fmt->depth) >> 3, BPL_ALIGNMENT); +-} +- +-static int __subdev_get_format(struct unicam_device *dev, +- struct v4l2_mbus_framefmt *fmt, int pad_id) +-{ +- struct v4l2_subdev_format sd_fmt = { +- .which = V4L2_SUBDEV_FORMAT_ACTIVE, +- .pad = pad_id +- }; +- int ret; +- +- ret = v4l2_subdev_call(dev->sensor, pad, get_fmt, dev->sensor_config, +- &sd_fmt); +- if (ret < 0) +- return ret; +- +- *fmt = sd_fmt.format; +- +- unicam_dbg(1, dev, "%s %dx%d code:%04x\n", __func__, +- fmt->width, fmt->height, fmt->code); +- +- return 0; +-} +- +-static int __subdev_set_format(struct unicam_device *dev, +- struct v4l2_mbus_framefmt *fmt, int pad_id) +-{ +- struct v4l2_subdev_format sd_fmt = { +- .which = V4L2_SUBDEV_FORMAT_ACTIVE, +- .pad = pad_id +- }; +- int ret; +- +- sd_fmt.format = *fmt; +- +- ret = v4l2_subdev_call(dev->sensor, pad, set_fmt, dev->sensor_config, +- &sd_fmt); +- if (ret < 0) +- return ret; +- +- if (pad_id == IMAGE_PAD) +- unicam_dbg(1, dev, "%s %dx%d code:%04x\n", __func__, fmt->width, +- fmt->height, fmt->code); +- else +- unicam_dbg(1, dev, "%s Embedded data code:%04x\n", __func__, +- sd_fmt.format.code); +- +- return 0; +-} +- +-static int unicam_calc_format_size_bpl(struct unicam_device *dev, +- const struct unicam_fmt *fmt, +- struct v4l2_format *f) +-{ +- unsigned int min_bytesperline; +- +- v4l_bound_align_image(&f->fmt.pix.width, MIN_WIDTH, MAX_WIDTH, 2, +- &f->fmt.pix.height, MIN_HEIGHT, MAX_HEIGHT, 0, +- 0); +- +- min_bytesperline = bytes_per_line(f->fmt.pix.width, fmt, +- f->fmt.pix.pixelformat); +- +- if (f->fmt.pix.bytesperline > min_bytesperline && +- f->fmt.pix.bytesperline <= MAX_BYTESPERLINE) +- f->fmt.pix.bytesperline = ALIGN(f->fmt.pix.bytesperline, +- BPL_ALIGNMENT); +- else +- f->fmt.pix.bytesperline = min_bytesperline; +- +- f->fmt.pix.sizeimage = f->fmt.pix.height * f->fmt.pix.bytesperline; +- +- unicam_dbg(3, dev, "%s: fourcc: %08X size: %dx%d bpl:%d img_size:%d\n", +- __func__, +- f->fmt.pix.pixelformat, +- f->fmt.pix.width, f->fmt.pix.height, +- f->fmt.pix.bytesperline, f->fmt.pix.sizeimage); +- +- return 0; +-} +- +-static int unicam_reset_format(struct unicam_node *node) +-{ +- struct unicam_device *dev = node->dev; +- struct v4l2_mbus_framefmt mbus_fmt; +- int ret; +- +- if (dev->sensor_embedded_data || node->pad_id != METADATA_PAD) { +- ret = __subdev_get_format(dev, &mbus_fmt, node->pad_id); +- if (ret) { +- unicam_err(dev, "Failed to get_format - ret %d\n", ret); +- return ret; +- } +- +- if (mbus_fmt.code != node->fmt->code) { +- unicam_err(dev, "code mismatch - fmt->code %08x, mbus_fmt.code %08x\n", +- node->fmt->code, mbus_fmt.code); +- return ret; +- } +- } +- +- if (node->pad_id == IMAGE_PAD) { +- v4l2_fill_pix_format(&node->v_fmt.fmt.pix, &mbus_fmt); +- node->v_fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; +- unicam_calc_format_size_bpl(dev, node->fmt, &node->v_fmt); +- } else { +- node->v_fmt.type = V4L2_BUF_TYPE_META_CAPTURE; +- node->v_fmt.fmt.meta.dataformat = V4L2_META_FMT_SENSOR_DATA; +- if (dev->sensor_embedded_data) { +- node->v_fmt.fmt.meta.buffersize = +- mbus_fmt.width * mbus_fmt.height; +- node->embedded_lines = mbus_fmt.height; +- } else { +- node->v_fmt.fmt.meta.buffersize = UNICAM_EMBEDDED_SIZE; +- node->embedded_lines = 1; +- } +- } +- +- node->m_fmt = mbus_fmt; +- return 0; +-} +- +-static void unicam_wr_dma_addr(struct unicam_cfg *cfg, dma_addr_t dmaaddr, +- unsigned int buffer_size, int pad_id) +-{ +- dma_addr_t endaddr = dmaaddr + buffer_size; +- +- /* +- * dmaaddr and endaddr should be a 32-bit address with the top two bits +- * set to 0x3 to signify uncached access through the Videocore memory +- * controller. +- */ +- BUG_ON((dmaaddr >> 30) != 0x3 && (endaddr >> 30) != 0x3); +- +- if (pad_id == IMAGE_PAD) { +- reg_write(cfg, UNICAM_IBSA0, dmaaddr); +- reg_write(cfg, UNICAM_IBEA0, endaddr); +- } else { +- reg_write(cfg, UNICAM_DBSA0, dmaaddr); +- reg_write(cfg, UNICAM_DBEA0, endaddr); +- } +-} +- +-static inline unsigned int unicam_get_lines_done(struct unicam_device *dev) +-{ +- dma_addr_t start_addr, cur_addr; +- unsigned int stride = dev->node[IMAGE_PAD].v_fmt.fmt.pix.bytesperline; +- struct unicam_buffer *frm = dev->node[IMAGE_PAD].cur_frm; +- +- if (!frm) +- return 0; +- +- start_addr = vb2_dma_contig_plane_dma_addr(&frm->vb.vb2_buf, 0); +- cur_addr = reg_read(&dev->cfg, UNICAM_IBWP); +- return (unsigned int)(cur_addr - start_addr) / stride; +-} +- +-static inline void unicam_schedule_next_buffer(struct unicam_node *node) +-{ +- struct unicam_device *dev = node->dev; +- struct unicam_dmaqueue *dma_q = &node->dma_queue; +- struct unicam_buffer *buf; +- unsigned int size; +- dma_addr_t addr; +- +- buf = list_entry(dma_q->active.next, struct unicam_buffer, list); +- node->next_frm = buf; +- list_del(&buf->list); +- +- addr = vb2_dma_contig_plane_dma_addr(&buf->vb.vb2_buf, 0); +- size = (node->pad_id == IMAGE_PAD) ? +- dev->node[IMAGE_PAD].v_fmt.fmt.pix.sizeimage : +- dev->node[METADATA_PAD].v_fmt.fmt.meta.buffersize; +- +- unicam_wr_dma_addr(&dev->cfg, addr, size, node->pad_id); +-} +- +-static inline void unicam_schedule_dummy_buffer(struct unicam_node *node) +-{ +- struct unicam_device *dev = node->dev; +- dma_addr_t addr = node->dummy_buf_dma_addr; +- +- unicam_dbg(3, dev, "Scheduling dummy buffer for node %d\n", +- node->pad_id); +- +- unicam_wr_dma_addr(&dev->cfg, addr, DUMMY_BUF_SIZE, node->pad_id); +- node->next_frm = NULL; +-} +- +-static inline void unicam_process_buffer_complete(struct unicam_node *node, +- unsigned int sequence) +-{ +- node->cur_frm->vb.field = node->m_fmt.field; +- node->cur_frm->vb.sequence = sequence; +- +- vb2_buffer_done(&node->cur_frm->vb.vb2_buf, VB2_BUF_STATE_DONE); +-} +- +-static int unicam_num_nodes_streaming(struct unicam_device *dev) +-{ +- return dev->node[IMAGE_PAD].streaming + +- dev->node[METADATA_PAD].streaming; +-} +- +-static int unicam_all_nodes_streaming(struct unicam_device *dev) +-{ +- int ret; +- +- ret = dev->node[IMAGE_PAD].open && dev->node[IMAGE_PAD].streaming; +- ret &= !dev->node[METADATA_PAD].open || +- dev->node[METADATA_PAD].streaming; +- return ret; +-} +- +-static void unicam_queue_event_sof(struct unicam_device *unicam) +-{ +- struct v4l2_event event = { +- .type = V4L2_EVENT_FRAME_SYNC, +- .u.frame_sync.frame_sequence = unicam->sequence, +- }; +- +- v4l2_event_queue(&unicam->node[IMAGE_PAD].video_dev, &event); +-} +- +-/* +- * unicam_isr : ISR handler for unicam capture +- * @irq: irq number +- * @dev_id: dev_id ptr +- * +- * It changes status of the captured buffer, takes next buffer from the queue +- * and sets its address in unicam registers +- */ +-static irqreturn_t unicam_isr(int irq, void *dev) +-{ +- struct unicam_device *unicam = (struct unicam_device *)dev; +- struct unicam_cfg *cfg = &unicam->cfg; +- unsigned int lines_done = unicam_get_lines_done(dev); +- unsigned int sequence = unicam->sequence; +- int num_nodes_streaming = unicam_num_nodes_streaming(dev); +- int ista, sta; +- u64 ts; +- int i; +- +- sta = reg_read(cfg, UNICAM_STA); +- /* Write value back to clear the interrupts */ +- reg_write(cfg, UNICAM_STA, sta); +- +- ista = reg_read(cfg, UNICAM_ISTA); +- /* Write value back to clear the interrupts */ +- reg_write(cfg, UNICAM_ISTA, ista); +- +- unicam_dbg(3, unicam, "ISR: ISTA: 0x%X, STA: 0x%X, sequence %d, lines done %d", +- ista, sta, sequence, lines_done); +- +- if (!(sta && (UNICAM_IS | UNICAM_PI0))) +- return IRQ_HANDLED; +- +- /* +- * We must run the frame end handler first. If we have a valid next_frm +- * and we get a simultaneout FE + FS interrupt, running the FS handler +- * first would null out the next_frm ptr and we would have lost the +- * buffer forever. +- */ +- if (ista & UNICAM_FEI || sta & UNICAM_PI0) { +- /* +- * Ensure we have swapped buffers already as we can't +- * stop the peripheral. If no buffer is available, use a +- * dummy buffer to dump out frames until we get a new buffer +- * to use. +- */ +- for (i = 0; i < num_nodes_streaming; i++) { +- if (unicam->node[i].cur_frm) +- unicam_process_buffer_complete(&unicam->node[i], +- sequence); +- unicam->node[i].cur_frm = unicam->node[i].next_frm; +- } +- unicam->sequence++; +- } +- +- if (ista & UNICAM_FSI) { +- /* +- * Timestamp is to be when the first data byte was captured, +- * aka frame start. +- */ +- ts = ktime_get_ns(); +- for (i = 0; i < num_nodes_streaming; i++) { +- if (unicam->node[i].cur_frm) +- unicam->node[i].cur_frm->vb.vb2_buf.timestamp = +- ts; +- /* +- * Set the next frame output to go to a dummy frame +- * if we have not managed to obtain another frame +- * from the queue. +- */ +- unicam_schedule_dummy_buffer(&unicam->node[i]); +- } +- +- unicam_queue_event_sof(unicam); +- } +- /* +- * Cannot swap buffer at frame end, there may be a race condition +- * where the HW does not actually swap it if the new frame has +- * already started. +- */ +- if (ista & (UNICAM_FSI | UNICAM_LCI) && !(ista & UNICAM_FEI)) { +- for (i = 0; i < num_nodes_streaming; i++) { +- spin_lock(&unicam->node[i].dma_queue_lock); +- if (!list_empty(&unicam->node[i].dma_queue.active) && +- !unicam->node[i].next_frm) +- unicam_schedule_next_buffer(&unicam->node[i]); +- spin_unlock(&unicam->node[i].dma_queue_lock); +- } +- } +- +- if (reg_read(&unicam->cfg, UNICAM_ICTL) & UNICAM_FCM) { +- /* Switch out of trigger mode if selected */ +- reg_write_field(&unicam->cfg, UNICAM_ICTL, 1, UNICAM_TFC); +- reg_write_field(&unicam->cfg, UNICAM_ICTL, 0, UNICAM_FCM); +- } +- return IRQ_HANDLED; +-} +- +-static int unicam_querycap(struct file *file, void *priv, +- struct v4l2_capability *cap) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- strlcpy(cap->driver, UNICAM_MODULE_NAME, sizeof(cap->driver)); +- strlcpy(cap->card, UNICAM_MODULE_NAME, sizeof(cap->card)); +- +- snprintf(cap->bus_info, sizeof(cap->bus_info), +- "platform:%s", dev->v4l2_dev.name); +- +- cap->capabilities = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING | +- V4L2_CAP_READWRITE | V4L2_CAP_DEVICE_CAPS | +- V4L2_CAP_META_CAPTURE; +- +- if (node->pad_id == IMAGE_PAD) +- cap->device_caps = V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING; +- else +- cap->device_caps = V4L2_CAP_META_CAPTURE | V4L2_CAP_STREAMING; +- +- return 0; +-} +- +-static int unicam_enum_fmt_vid_cap(struct file *file, void *priv, +- struct v4l2_fmtdesc *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev_mbus_code_enum mbus_code; +- const struct unicam_fmt *fmt = NULL; +- int index = 0; +- int ret = 0; +- int i; +- +- if (node->pad_id == METADATA_PAD) +- return -EINVAL; +- +- for (i = 0; !ret && i < MAX_ENUM_MBUS_CODE; i++) { +- memset(&mbus_code, 0, sizeof(mbus_code)); +- mbus_code.index = i; +- mbus_code.pad = IMAGE_PAD; +- mbus_code.which = V4L2_SUBDEV_FORMAT_ACTIVE; +- +- ret = v4l2_subdev_call(dev->sensor, pad, enum_mbus_code, +- NULL, &mbus_code); +- if (ret < 0) { +- unicam_dbg(2, dev, +- "subdev->enum_mbus_code idx %d returned %d - index invalid\n", +- i, ret); +- return -EINVAL; +- } +- +- fmt = find_format_by_code(mbus_code.code); +- if (fmt) { +- if (fmt->fourcc) { +- if (index == f->index) { +- f->pixelformat = fmt->fourcc; +- break; +- } +- index++; +- } +- if (fmt->repacked_fourcc) { +- if (index == f->index) { +- f->pixelformat = fmt->repacked_fourcc; +- break; +- } +- index++; +- } +- } +- } +- +- return 0; +-} +- +-static int unicam_g_fmt_vid_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct v4l2_mbus_framefmt mbus_fmt = {0}; +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- const struct unicam_fmt *fmt = NULL; +- int ret; +- +- if (node->pad_id != IMAGE_PAD) +- return -EINVAL; +- +- /* +- * If a flip has occurred in the sensor, the fmt code might have +- * changed. So we will need to re-fetch the format from the subdevice. +- */ +- ret = __subdev_get_format(dev, &mbus_fmt, node->pad_id); +- if (ret) +- return -EINVAL; +- +- /* Find the V4L2 format from mbus code. We must match a known format. */ +- fmt = find_format_by_code(mbus_fmt.code); +- if (!fmt) +- return -EINVAL; +- +- if (node->fmt != fmt) { +- /* +- * The sensor format has changed so the pixelformat needs to +- * be updated. Try and retain the packed/unpacked choice if +- * at all possible. +- */ +- if (node->fmt->repacked_fourcc == +- node->v_fmt.fmt.pix.pixelformat) +- /* Using the repacked format */ +- node->v_fmt.fmt.pix.pixelformat = fmt->repacked_fourcc; +- else +- /* Using the native format */ +- node->v_fmt.fmt.pix.pixelformat = fmt->fourcc; +- +- node->fmt = fmt; +- } +- +- *f = node->v_fmt; +- +- return 0; +-} +- +-static +-const struct unicam_fmt *get_first_supported_format(struct unicam_device *dev) +-{ +- struct v4l2_subdev_mbus_code_enum mbus_code; +- const struct unicam_fmt *fmt = NULL; +- int ret = 0; +- int j; +- +- for (j = 0; ret != -EINVAL && ret != -ENOIOCTLCMD; ++j) { +- memset(&mbus_code, 0, sizeof(mbus_code)); +- mbus_code.index = j; +- mbus_code.pad = IMAGE_PAD; +- mbus_code.which = V4L2_SUBDEV_FORMAT_ACTIVE; +- +- ret = v4l2_subdev_call(dev->sensor, pad, enum_mbus_code, NULL, +- &mbus_code); +- if (ret < 0) { +- unicam_dbg(2, dev, +- "subdev->enum_mbus_code idx %d returned %d - continue\n", +- j, ret); +- continue; +- } +- +- unicam_dbg(2, dev, "subdev %s: code: 0x%08x idx: %d\n", +- dev->sensor->name, mbus_code.code, j); +- +- fmt = find_format_by_code(mbus_code.code); +- unicam_dbg(2, dev, "fmt 0x%08x returned as %p, V4L2 FOURCC 0x%08x, csi_dt 0x%02x\n", +- mbus_code.code, fmt, fmt ? fmt->fourcc : 0, +- fmt ? fmt->csi_dt : 0); +- if (fmt) +- return fmt; +- } +- +- return NULL; +-} +- +-static int unicam_try_fmt_vid_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev_format sd_fmt = { +- .which = V4L2_SUBDEV_FORMAT_TRY, +- .pad = IMAGE_PAD +- }; +- struct v4l2_mbus_framefmt *mbus_fmt = &sd_fmt.format; +- const struct unicam_fmt *fmt; +- int ret; +- +- if (node->pad_id == METADATA_PAD) +- return -EINVAL; +- +- fmt = find_format_by_pix(dev, f->fmt.pix.pixelformat); +- if (!fmt) { +- /* Pixel format not supported by unicam. Choose the first +- * supported format, and let the sensor choose something else. +- */ +- unicam_dbg(3, dev, "Fourcc format (0x%08x) not found. Use first format.\n", +- f->fmt.pix.pixelformat); +- +- fmt = &formats[0]; +- f->fmt.pix.pixelformat = fmt->fourcc; +- } +- +- v4l2_fill_mbus_format(mbus_fmt, &f->fmt.pix, fmt->code); +- /* +- * No support for receiving interlaced video, so never +- * request it from the sensor subdev. +- */ +- mbus_fmt->field = V4L2_FIELD_NONE; +- +- ret = v4l2_subdev_call(dev->sensor, pad, set_fmt, dev->sensor_config, +- &sd_fmt); +- if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV) +- return ret; +- +- if (mbus_fmt->field != V4L2_FIELD_NONE) +- unicam_info(dev, "Sensor trying to send interlaced video - results may be unpredictable\n"); +- +- v4l2_fill_pix_format(&f->fmt.pix, &sd_fmt.format); +- if (mbus_fmt->code != fmt->code) { +- /* Sensor has returned an alternate format */ +- fmt = find_format_by_code(mbus_fmt->code); +- if (!fmt) { +- /* The alternate format is one unicam can't support. +- * Find the first format that is supported by both, and +- * then set that. +- */ +- fmt = get_first_supported_format(dev); +- mbus_fmt->code = fmt->code; +- +- ret = v4l2_subdev_call(dev->sensor, pad, set_fmt, +- dev->sensor_config, &sd_fmt); +- if (ret && ret != -ENOIOCTLCMD && ret != -ENODEV) +- return ret; +- +- if (mbus_fmt->field != V4L2_FIELD_NONE) +- unicam_info(dev, "Sensor trying to send interlaced video - results may be unpredictable\n"); +- +- v4l2_fill_pix_format(&f->fmt.pix, &sd_fmt.format); +- +- if (mbus_fmt->code != fmt->code) { +- /* We've set a format that the sensor reports +- * as being supported, but it refuses to set it. +- * Not much else we can do. +- * Assume that the sensor driver may accept the +- * format when it is set (rather than tried). +- */ +- unicam_err(dev, "Sensor won't accept default format, and Unicam can't support sensor default\n"); +- } +- } +- +- if (fmt->fourcc) +- f->fmt.pix.pixelformat = fmt->fourcc; +- else +- f->fmt.pix.pixelformat = fmt->repacked_fourcc; +- } +- +- return unicam_calc_format_size_bpl(dev, fmt, f); +-} +- +-static int unicam_s_fmt_vid_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct vb2_queue *q = &node->buffer_queue; +- struct v4l2_mbus_framefmt mbus_fmt = {0}; +- const struct unicam_fmt *fmt; +- int ret; +- +- if (vb2_is_busy(q)) +- return -EBUSY; +- +- ret = unicam_try_fmt_vid_cap(file, priv, f); +- if (ret < 0) +- return ret; +- +- fmt = find_format_by_pix(dev, f->fmt.pix.pixelformat); +- if (!fmt) { +- /* Unknown pixel format - adopt a default. +- * This shouldn't happen as try_fmt should have resolved any +- * issues first. +- */ +- fmt = get_first_supported_format(dev); +- if (!fmt) +- /* It shouldn't be possible to get here with no +- * supported formats +- */ +- return -EINVAL; +- f->fmt.pix.pixelformat = fmt->fourcc; +- return -EINVAL; +- } +- +- v4l2_fill_mbus_format(&mbus_fmt, &f->fmt.pix, fmt->code); +- +- ret = __subdev_set_format(dev, &mbus_fmt, node->pad_id); +- if (ret) { +- unicam_dbg(3, dev, "%s __subdev_set_format failed %d\n", +- __func__, ret); +- return ret; +- } +- +- /* Just double check nothing has gone wrong */ +- if (mbus_fmt.code != fmt->code) { +- unicam_dbg(3, dev, +- "%s subdev changed format on us, this should not happen\n", +- __func__); +- return -EINVAL; +- } +- +- node->fmt = fmt; +- node->v_fmt.fmt.pix.pixelformat = f->fmt.pix.pixelformat; +- node->v_fmt.fmt.pix.bytesperline = f->fmt.pix.bytesperline; +- unicam_reset_format(node); +- +- unicam_dbg(3, dev, +- "%s %dx%d, mbus_fmt 0x%08X, V4L2 pix 0x%08X.\n", +- __func__, node->v_fmt.fmt.pix.width, +- node->v_fmt.fmt.pix.height, mbus_fmt.code, +- node->v_fmt.fmt.pix.pixelformat); +- +- *f = node->v_fmt; +- +- return 0; +-} +- +-static int unicam_enum_fmt_meta_cap(struct file *file, void *priv, +- struct v4l2_fmtdesc *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev_mbus_code_enum mbus_code; +- const struct unicam_fmt *fmt = NULL; +- int ret = 0; +- +- if (node->pad_id != METADATA_PAD || f->index != 0) +- return -EINVAL; +- +- if (dev->sensor_embedded_data) { +- memset(&mbus_code, 0, sizeof(mbus_code)); +- mbus_code.index = f->index; +- mbus_code.which = V4L2_SUBDEV_FORMAT_ACTIVE; +- mbus_code.pad = METADATA_PAD; +- +- ret = v4l2_subdev_call(dev->sensor, pad, enum_mbus_code, NULL, +- &mbus_code); +- if (ret < 0) { +- unicam_dbg(2, dev, +- "subdev->enum_mbus_code idx 0 returned %d - index invalid\n", +- ret); +- return -EINVAL; +- } +- } else { +- mbus_code.code = MEDIA_BUS_FMT_SENSOR_DATA; +- } +- +- fmt = find_format_by_code(mbus_code.code); +- if (fmt) +- f->pixelformat = fmt->fourcc; +- +- return 0; +-} +- +-static int unicam_g_fmt_meta_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- +- if (node->pad_id != METADATA_PAD) +- return -EINVAL; +- +- *f = node->v_fmt; +- +- return 0; +-} +- +-static int unicam_try_fmt_meta_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- +- if (node->pad_id != METADATA_PAD) +- return -EINVAL; +- +- *f = node->v_fmt; +- +- return 0; +-} +- +-static int unicam_s_fmt_meta_cap(struct file *file, void *priv, +- struct v4l2_format *f) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_mbus_framefmt mbus_fmt = { 0 }; +- const struct unicam_fmt *fmt; +- int ret; +- +- if (node->pad_id == IMAGE_PAD) +- return -EINVAL; +- +- if (dev->sensor_embedded_data) { +- fmt = find_format_by_pix(dev, f->fmt.meta.dataformat); +- if (!fmt) { +- unicam_err(dev, "unknown format: V4L2 pix 0x%08x\n", +- f->fmt.meta.dataformat); +- return -EINVAL; +- } +- mbus_fmt.code = fmt->code; +- ret = __subdev_set_format(dev, &mbus_fmt, node->pad_id); +- if (ret) { +- unicam_dbg(3, dev, "%s __subdev_set_format failed %d\n", +- __func__, ret); +- return ret; +- } +- } +- +- *f = node->v_fmt; +- +- unicam_dbg(3, dev, "%s size %d, V4L2 pix 0x%08x\n", +- __func__, node->v_fmt.fmt.meta.buffersize, +- node->v_fmt.fmt.meta.dataformat); +- +- return 0; +-} +- +-static int unicam_queue_setup(struct vb2_queue *vq, +- unsigned int *nbuffers, +- unsigned int *nplanes, +- unsigned int sizes[], +- struct device *alloc_devs[]) +-{ +- struct unicam_node *node = vb2_get_drv_priv(vq); +- struct unicam_device *dev = node->dev; +- unsigned int size = node->pad_id == IMAGE_PAD ? +- node->v_fmt.fmt.pix.sizeimage : +- node->v_fmt.fmt.meta.buffersize; +- +- if (vq->num_buffers + *nbuffers < 3) +- *nbuffers = 3 - vq->num_buffers; +- +- if (*nplanes) { +- if (sizes[0] < size) { +- unicam_err(dev, "sizes[0] %i < size %u\n", sizes[0], +- size); +- return -EINVAL; +- } +- size = sizes[0]; +- } +- +- *nplanes = 1; +- sizes[0] = size; +- +- return 0; +-} +- +-static int unicam_buffer_prepare(struct vb2_buffer *vb) +-{ +- struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue); +- struct unicam_device *dev = node->dev; +- struct unicam_buffer *buf = container_of(vb, struct unicam_buffer, +- vb.vb2_buf); +- unsigned long size; +- +- if (WARN_ON(!node->fmt)) +- return -EINVAL; +- +- size = node->pad_id == IMAGE_PAD ? node->v_fmt.fmt.pix.sizeimage : +- node->v_fmt.fmt.meta.buffersize; +- if (vb2_plane_size(vb, 0) < size) { +- unicam_err(dev, "data will not fit into plane (%lu < %lu)\n", +- vb2_plane_size(vb, 0), size); +- return -EINVAL; +- } +- +- vb2_set_plane_payload(&buf->vb.vb2_buf, 0, size); +- return 0; +-} +- +-static void unicam_buffer_queue(struct vb2_buffer *vb) +-{ +- struct unicam_node *node = vb2_get_drv_priv(vb->vb2_queue); +- struct unicam_buffer *buf = container_of(vb, struct unicam_buffer, +- vb.vb2_buf); +- struct unicam_dmaqueue *dma_queue = &node->dma_queue; +- unsigned long flags = 0; +- +- spin_lock_irqsave(&node->dma_queue_lock, flags); +- list_add_tail(&buf->list, &dma_queue->active); +- spin_unlock_irqrestore(&node->dma_queue_lock, flags); +-} +- +-static void unicam_set_packing_config(struct unicam_device *dev) +-{ +- int pack, unpack; +- u32 val; +- +- if (dev->node[IMAGE_PAD].v_fmt.fmt.pix.pixelformat == +- dev->node[IMAGE_PAD].fmt->fourcc) { +- unpack = UNICAM_PUM_NONE; +- pack = UNICAM_PPM_NONE; +- } else { +- switch (dev->node[IMAGE_PAD].fmt->depth) { +- case 8: +- unpack = UNICAM_PUM_UNPACK8; +- break; +- case 10: +- unpack = UNICAM_PUM_UNPACK10; +- break; +- case 12: +- unpack = UNICAM_PUM_UNPACK12; +- break; +- case 14: +- unpack = UNICAM_PUM_UNPACK14; +- break; +- case 16: +- unpack = UNICAM_PUM_UNPACK16; +- break; +- default: +- unpack = UNICAM_PUM_NONE; +- break; +- } +- +- /* Repacking is always to 16bpp */ +- pack = UNICAM_PPM_PACK16; +- } +- +- val = 0; +- set_field(&val, unpack, UNICAM_PUM_MASK); +- set_field(&val, pack, UNICAM_PPM_MASK); +- reg_write(&dev->cfg, UNICAM_IPIPE, val); +-} +- +-static void unicam_cfg_image_id(struct unicam_device *dev) +-{ +- struct unicam_cfg *cfg = &dev->cfg; +- +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) { +- /* CSI2 mode */ +- reg_write(cfg, UNICAM_IDI0, +- (dev->virtual_channel << 6) | +- dev->node[IMAGE_PAD].fmt->csi_dt); +- } else { +- /* CCP2 mode */ +- reg_write(cfg, UNICAM_IDI0, +- 0x80 | dev->node[IMAGE_PAD].fmt->csi_dt); +- } +-} +- +-static void unicam_enable_ed(struct unicam_device *dev) +-{ +- struct unicam_cfg *cfg = &dev->cfg; +- u32 val = reg_read(cfg, UNICAM_DCS); +- +- set_field(&val, 2, UNICAM_EDL_MASK); +- /* Do not wrap at the end of the embedded data buffer */ +- set_field(&val, 0, UNICAM_DBOB); +- +- reg_write(cfg, UNICAM_DCS, val); +-} +- +-static void unicam_start_rx(struct unicam_device *dev, dma_addr_t *addr) +-{ +- struct unicam_cfg *cfg = &dev->cfg; +- int line_int_freq = dev->node[IMAGE_PAD].v_fmt.fmt.pix.height >> 2; +- unsigned int size, i; +- u32 val; +- +- if (line_int_freq < 128) +- line_int_freq = 128; +- +- /* Enable lane clocks */ +- val = 1; +- for (i = 0; i < dev->active_data_lanes; i++) +- val = val << 2 | 1; +- clk_write(cfg, val); +- +- /* Basic init */ +- reg_write(cfg, UNICAM_CTRL, UNICAM_MEM); +- +- /* Enable analogue control, and leave in reset. */ +- val = UNICAM_AR; +- set_field(&val, 7, UNICAM_CTATADJ_MASK); +- set_field(&val, 7, UNICAM_PTATADJ_MASK); +- reg_write(cfg, UNICAM_ANA, val); +- usleep_range(1000, 2000); +- +- /* Come out of reset */ +- reg_write_field(cfg, UNICAM_ANA, 0, UNICAM_AR); +- +- /* Peripheral reset */ +- reg_write_field(cfg, UNICAM_CTRL, 1, UNICAM_CPR); +- reg_write_field(cfg, UNICAM_CTRL, 0, UNICAM_CPR); +- +- reg_write_field(cfg, UNICAM_CTRL, 0, UNICAM_CPE); +- +- /* Enable Rx control. */ +- val = reg_read(cfg, UNICAM_CTRL); +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) { +- set_field(&val, UNICAM_CPM_CSI2, UNICAM_CPM_MASK); +- set_field(&val, UNICAM_DCM_STROBE, UNICAM_DCM_MASK); +- } else { +- set_field(&val, UNICAM_CPM_CCP2, UNICAM_CPM_MASK); +- set_field(&val, dev->bus_flags, UNICAM_DCM_MASK); +- } +- /* Packet framer timeout */ +- set_field(&val, 0xf, UNICAM_PFT_MASK); +- set_field(&val, 128, UNICAM_OET_MASK); +- reg_write(cfg, UNICAM_CTRL, val); +- +- reg_write(cfg, UNICAM_IHWIN, 0); +- reg_write(cfg, UNICAM_IVWIN, 0); +- +- /* AXI bus access QoS setup */ +- val = reg_read(&dev->cfg, UNICAM_PRI); +- set_field(&val, 0, UNICAM_BL_MASK); +- set_field(&val, 0, UNICAM_BS_MASK); +- set_field(&val, 0xe, UNICAM_PP_MASK); +- set_field(&val, 8, UNICAM_NP_MASK); +- set_field(&val, 2, UNICAM_PT_MASK); +- set_field(&val, 1, UNICAM_PE); +- reg_write(cfg, UNICAM_PRI, val); +- +- reg_write_field(cfg, UNICAM_ANA, 0, UNICAM_DDL); +- +- /* Always start in trigger frame capture mode (UNICAM_FCM set) */ +- val = UNICAM_FSIE | UNICAM_FEIE | UNICAM_FCM | UNICAM_IBOB; +- set_field(&val, line_int_freq, UNICAM_LCIE_MASK); +- reg_write(cfg, UNICAM_ICTL, val); +- reg_write(cfg, UNICAM_STA, UNICAM_STA_MASK_ALL); +- reg_write(cfg, UNICAM_ISTA, UNICAM_ISTA_MASK_ALL); +- +- /* tclk_term_en */ +- reg_write_field(cfg, UNICAM_CLT, 2, UNICAM_CLT1_MASK); +- /* tclk_settle */ +- reg_write_field(cfg, UNICAM_CLT, 6, UNICAM_CLT2_MASK); +- /* td_term_en */ +- reg_write_field(cfg, UNICAM_DLT, 2, UNICAM_DLT1_MASK); +- /* ths_settle */ +- reg_write_field(cfg, UNICAM_DLT, 6, UNICAM_DLT2_MASK); +- /* trx_enable */ +- reg_write_field(cfg, UNICAM_DLT, 0, UNICAM_DLT3_MASK); +- +- reg_write_field(cfg, UNICAM_CTRL, 0, UNICAM_SOE); +- +- /* Packet compare setup - required to avoid missing frame ends */ +- val = 0; +- set_field(&val, 1, UNICAM_PCE); +- set_field(&val, 1, UNICAM_GI); +- set_field(&val, 1, UNICAM_CPH); +- set_field(&val, 0, UNICAM_PCVC_MASK); +- set_field(&val, 1, UNICAM_PCDT_MASK); +- reg_write(cfg, UNICAM_CMP0, val); +- +- /* Enable clock lane and set up terminations */ +- val = 0; +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) { +- /* CSI2 */ +- set_field(&val, 1, UNICAM_CLE); +- set_field(&val, 1, UNICAM_CLLPE); +- if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) { +- set_field(&val, 1, UNICAM_CLTRE); +- set_field(&val, 1, UNICAM_CLHSE); +- } +- } else { +- /* CCP2 */ +- set_field(&val, 1, UNICAM_CLE); +- set_field(&val, 1, UNICAM_CLHSE); +- set_field(&val, 1, UNICAM_CLTRE); +- } +- reg_write(cfg, UNICAM_CLK, val); +- +- /* +- * Enable required data lanes with appropriate terminations. +- * The same value needs to be written to UNICAM_DATn registers for +- * the active lanes, and 0 for inactive ones. +- */ +- val = 0; +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) { +- /* CSI2 */ +- set_field(&val, 1, UNICAM_DLE); +- set_field(&val, 1, UNICAM_DLLPE); +- if (dev->bus_flags & V4L2_MBUS_CSI2_CONTINUOUS_CLOCK) { +- set_field(&val, 1, UNICAM_DLTRE); +- set_field(&val, 1, UNICAM_DLHSE); +- } +- } else { +- /* CCP2 */ +- set_field(&val, 1, UNICAM_DLE); +- set_field(&val, 1, UNICAM_DLHSE); +- set_field(&val, 1, UNICAM_DLTRE); +- } +- reg_write(cfg, UNICAM_DAT0, val); +- +- if (dev->active_data_lanes == 1) +- val = 0; +- reg_write(cfg, UNICAM_DAT1, val); +- +- if (dev->max_data_lanes > 2) { +- /* +- * Registers UNICAM_DAT2 and UNICAM_DAT3 only valid if the +- * instance supports more than 2 data lanes. +- */ +- if (dev->active_data_lanes == 2) +- val = 0; +- reg_write(cfg, UNICAM_DAT2, val); +- +- if (dev->active_data_lanes == 3) +- val = 0; +- reg_write(cfg, UNICAM_DAT3, val); +- } +- +- reg_write(&dev->cfg, UNICAM_IBLS, +- dev->node[IMAGE_PAD].v_fmt.fmt.pix.bytesperline); +- size = dev->node[IMAGE_PAD].v_fmt.fmt.pix.sizeimage; +- unicam_wr_dma_addr(&dev->cfg, addr[IMAGE_PAD], size, IMAGE_PAD); +- unicam_set_packing_config(dev); +- unicam_cfg_image_id(dev); +- +- val = reg_read(cfg, UNICAM_MISC); +- set_field(&val, 1, UNICAM_FL0); +- set_field(&val, 1, UNICAM_FL1); +- reg_write(cfg, UNICAM_MISC, val); +- +- if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data) { +- size = dev->node[METADATA_PAD].v_fmt.fmt.meta.buffersize; +- unicam_enable_ed(dev); +- unicam_wr_dma_addr(&dev->cfg, addr[METADATA_PAD], size, +- METADATA_PAD); +- } +- +- /* Enable peripheral */ +- reg_write_field(cfg, UNICAM_CTRL, 1, UNICAM_CPE); +- +- /* Load image pointers */ +- reg_write_field(cfg, UNICAM_ICTL, 1, UNICAM_LIP_MASK); +- +- /* Load embedded data buffer pointers if needed */ +- if (dev->node[METADATA_PAD].streaming && dev->sensor_embedded_data) +- reg_write_field(cfg, UNICAM_DCS, 1, UNICAM_LDP); +- +- /* +- * Enable trigger only for the first frame to +- * sync correctly to the FS from the source. +- */ +- reg_write_field(cfg, UNICAM_ICTL, 1, UNICAM_TFC); +-} +- +-static void unicam_disable(struct unicam_device *dev) +-{ +- struct unicam_cfg *cfg = &dev->cfg; +- +- /* Analogue lane control disable */ +- reg_write_field(cfg, UNICAM_ANA, 1, UNICAM_DDL); +- +- /* Stop the output engine */ +- reg_write_field(cfg, UNICAM_CTRL, 1, UNICAM_SOE); +- +- /* Disable the data lanes. */ +- reg_write(cfg, UNICAM_DAT0, 0); +- reg_write(cfg, UNICAM_DAT1, 0); +- +- if (dev->max_data_lanes > 2) { +- reg_write(cfg, UNICAM_DAT2, 0); +- reg_write(cfg, UNICAM_DAT3, 0); +- } +- +- /* Peripheral reset */ +- reg_write_field(cfg, UNICAM_CTRL, 1, UNICAM_CPR); +- usleep_range(50, 100); +- reg_write_field(cfg, UNICAM_CTRL, 0, UNICAM_CPR); +- +- /* Disable peripheral */ +- reg_write_field(cfg, UNICAM_CTRL, 0, UNICAM_CPE); +- +- /* Clear ED setup */ +- reg_write(cfg, UNICAM_DCS, 0); +- +- /* Disable all lane clocks */ +- clk_write(cfg, 0); +-} +- +-static int unicam_start_streaming(struct vb2_queue *vq, unsigned int count) +-{ +- struct unicam_node *node = vb2_get_drv_priv(vq); +- struct unicam_device *dev = node->dev; +- struct unicam_buffer *buf; +- dma_addr_t buffer_addr[MAX_NODES] = { 0 }; +- int num_nodes_streaming; +- unsigned long flags; +- int ret, i; +- +- node->streaming = 1; +- if (!unicam_all_nodes_streaming(dev)) { +- unicam_dbg(3, dev, "Not all nodes are streaming yet."); +- return 0; +- } +- +- dev->sequence = 0; +- ret = unicam_runtime_get(dev); +- if (ret < 0) { +- unicam_dbg(3, dev, "unicam_runtime_get failed\n"); +- return ret; +- } +- +- dev->active_data_lanes = dev->max_data_lanes; +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY && +- v4l2_subdev_has_op(dev->sensor, video, g_mbus_config)) { +- struct v4l2_mbus_config mbus_config; +- +- ret = v4l2_subdev_call(dev->sensor, video, g_mbus_config, +- &mbus_config); +- if (ret < 0) { +- unicam_dbg(3, dev, "g_mbus_config failed\n"); +- goto err_pm_put; +- } +- +- dev->active_data_lanes = +- (mbus_config.flags & V4L2_MBUS_CSI2_LANE_MASK) >> +- __ffs(V4L2_MBUS_CSI2_LANE_MASK); +- if (!dev->active_data_lanes) +- dev->active_data_lanes = dev->max_data_lanes; +- } +- if (dev->active_data_lanes > dev->max_data_lanes) { +- unicam_err(dev, "Device has requested %u data lanes, which is >%u configured in DT\n", +- dev->active_data_lanes, dev->max_data_lanes); +- ret = -EINVAL; +- goto err_pm_put; +- } +- +- unicam_dbg(1, dev, "Running with %u data lanes\n", +- dev->active_data_lanes); +- +- ret = clk_set_rate(dev->clock, 100 * 1000 * 1000); +- if (ret) { +- unicam_err(dev, "failed to set up clock\n"); +- goto err_pm_put; +- } +- +- ret = clk_prepare_enable(dev->clock); +- if (ret) { +- unicam_err(dev, "Failed to enable CSI clock: %d\n", ret); +- goto err_pm_put; +- } +- +- num_nodes_streaming = unicam_num_nodes_streaming(dev); +- for (i = 0; i < num_nodes_streaming; i++) { +- spin_lock_irqsave(&dev->node[i].dma_queue_lock, flags); +- buf = list_entry(dev->node[i].dma_queue.active.next, +- struct unicam_buffer, list); +- dev->node[i].cur_frm = buf; +- dev->node[i].next_frm = buf; +- list_del(&buf->list); +- spin_unlock_irqrestore(&dev->node[i].dma_queue_lock, flags); +- buffer_addr[i] = +- vb2_dma_contig_plane_dma_addr(&dev->node[i].cur_frm->vb.vb2_buf, +- 0); +- } +- +- unicam_start_rx(dev, buffer_addr); +- +- ret = v4l2_subdev_call(dev->sensor, video, s_stream, 1); +- if (ret < 0) { +- unicam_err(dev, "stream on failed in subdev\n"); +- goto err_disable_unicam; +- } +- +- return 0; +- +-err_disable_unicam: +- node->streaming = 0; +- unicam_disable(dev); +- clk_disable_unprepare(dev->clock); +-err_pm_put: +- unicam_runtime_put(dev); +- +- return ret; +-} +- +-static void unicam_stop_streaming(struct vb2_queue *vq) +-{ +- struct unicam_node *node = vb2_get_drv_priv(vq); +- struct unicam_device *dev = node->dev; +- struct unicam_dmaqueue *dma_q = &node->dma_queue; +- struct unicam_buffer *buf, *tmp; +- unsigned long flags; +- +- node->streaming = 0; +- +- if (node->pad_id == IMAGE_PAD) { +- /* Stop streaming the sensor and disable the peripheral. +- * We cannot continue streaming embedded data with the +- * image pad disabled. +- */ +- if (v4l2_subdev_call(dev->sensor, video, s_stream, 0) < 0) +- unicam_err(dev, "stream off failed in subdev\n"); +- +- unicam_disable(dev); +- clk_disable_unprepare(dev->clock); +- unicam_runtime_put(dev); +- +- } else if (node->pad_id == METADATA_PAD) { +- /* Allow the hardware to spin in the dummy buffer. +- * This is only really needed if the embedded data pad is +- * disabled before the image pad. The 0x3 in the top two bits +- * signifies uncached accesses through the Videocore memory +- * controller. +- */ +- unicam_wr_dma_addr(&dev->cfg, node->dummy_buf_dma_addr, +- DUMMY_BUF_SIZE, METADATA_PAD); +- } +- +- /* Clear all queued buffers for the node */ +- spin_lock_irqsave(&node->dma_queue_lock, flags); +- list_for_each_entry_safe(buf, tmp, &dma_q->active, list) { +- list_del(&buf->list); +- vb2_buffer_done(&buf->vb.vb2_buf, VB2_BUF_STATE_ERROR); +- } +- +- if (node->cur_frm) +- vb2_buffer_done(&node->cur_frm->vb.vb2_buf, +- VB2_BUF_STATE_ERROR); +- if (node->next_frm && node->cur_frm != node->next_frm) +- vb2_buffer_done(&node->next_frm->vb.vb2_buf, +- VB2_BUF_STATE_ERROR); +- +- node->cur_frm = NULL; +- node->next_frm = NULL; +- spin_unlock_irqrestore(&node->dma_queue_lock, flags); +-} +- +-static int unicam_enum_input(struct file *file, void *priv, +- struct v4l2_input *inp) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- if (inp->index != 0) +- return -EINVAL; +- +- inp->type = V4L2_INPUT_TYPE_CAMERA; +- if (v4l2_subdev_has_op(dev->sensor, video, s_dv_timings)) { +- inp->capabilities = V4L2_IN_CAP_DV_TIMINGS; +- inp->std = 0; +- } else if (v4l2_subdev_has_op(dev->sensor, video, s_std)) { +- inp->capabilities = V4L2_IN_CAP_STD; +- if (v4l2_subdev_call(dev->sensor, video, g_tvnorms, &inp->std) +- < 0) +- inp->std = V4L2_STD_ALL; +- } else { +- inp->capabilities = 0; +- inp->std = 0; +- } +- sprintf(inp->name, "Camera 0"); +- return 0; +-} +- +-static int unicam_g_input(struct file *file, void *priv, unsigned int *i) +-{ +- *i = 0; +- +- return 0; +-} +- +-static int unicam_s_input(struct file *file, void *priv, unsigned int i) +-{ +- /* +- * FIXME: Ideally we would like to be able to query the source +- * subdevice for information over the input connectors it supports, +- * and map that through in to a call to video_ops->s_routing. +- * There is no infrastructure support for defining that within +- * devicetree at present. Until that is implemented we can't +- * map a user physical connector number to s_routing input number. +- */ +- if (i > 0) +- return -EINVAL; +- +- return 0; +-} +- +-static int unicam_querystd(struct file *file, void *priv, +- v4l2_std_id *std) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, video, querystd, std); +-} +- +-static int unicam_g_std(struct file *file, void *priv, v4l2_std_id *std) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, video, g_std, std); +-} +- +-static int unicam_s_std(struct file *file, void *priv, v4l2_std_id std) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- int ret; +- v4l2_std_id current_std; +- +- ret = v4l2_subdev_call(dev->sensor, video, g_std, ¤t_std); +- if (ret) +- return ret; +- +- if (std == current_std) +- return 0; +- +- if (vb2_is_busy(&node->buffer_queue)) +- return -EBUSY; +- +- ret = v4l2_subdev_call(dev->sensor, video, s_std, std); +- +- /* Force recomputation of bytesperline */ +- node->v_fmt.fmt.pix.bytesperline = 0; +- +- unicam_reset_format(node); +- +- return ret; +-} +- +-static int unicam_s_edid(struct file *file, void *priv, struct v4l2_edid *edid) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, pad, set_edid, edid); +-} +- +-static int unicam_g_edid(struct file *file, void *priv, struct v4l2_edid *edid) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, pad, get_edid, edid); +-} +- +-static int unicam_s_selection(struct file *file, void *priv, +- struct v4l2_selection *sel) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev_selection sdsel = { +- .which = V4L2_SUBDEV_FORMAT_ACTIVE, +- .target = sel->target, +- .flags = sel->flags, +- .r = sel->r, +- }; +- +- return v4l2_subdev_call(dev->sensor, pad, set_selection, NULL, &sdsel); +-} +- +-static int unicam_g_selection(struct file *file, void *priv, +- struct v4l2_selection *sel) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev_selection sdsel = { +- .which = V4L2_SUBDEV_FORMAT_ACTIVE, +- .target = sel->target, +- }; +- int ret; +- +- ret = v4l2_subdev_call(dev->sensor, pad, get_selection, NULL, &sdsel); +- if (!ret) +- sel->r = sdsel.r; +- +- return ret; +-} +- +-static int unicam_enum_framesizes(struct file *file, void *priv, +- struct v4l2_frmsizeenum *fsize) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- const struct unicam_fmt *fmt; +- struct v4l2_subdev_frame_size_enum fse; +- int ret; +- +- if (node->pad_id == IMAGE_PAD) { +- /* check for valid format */ +- fmt = find_format_by_pix(dev, fsize->pixel_format); +- if (!fmt) { +- unicam_dbg(3, dev, "Invalid pixel code: %x\n", +- fsize->pixel_format); +- return -EINVAL; +- } +- fse.code = fmt->code; +- } else { +- /* This pad is for embedded data, so just set the format */ +- fse.code = MEDIA_BUS_FMT_SENSOR_DATA; +- } +- +- fse.which = V4L2_SUBDEV_FORMAT_ACTIVE; +- fse.index = fsize->index; +- fse.pad = node->pad_id; +- +- ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_size, NULL, &fse); +- if (ret) +- return ret; +- +- unicam_dbg(1, dev, "%s: index: %d code: %x W:[%d,%d] H:[%d,%d]\n", +- __func__, fse.index, fse.code, fse.min_width, fse.max_width, +- fse.min_height, fse.max_height); +- +- fsize->type = V4L2_FRMSIZE_TYPE_DISCRETE; +- fsize->discrete.width = fse.max_width; +- fsize->discrete.height = fse.max_height; +- +- return 0; +-} +- +-static int unicam_enum_frameintervals(struct file *file, void *priv, +- struct v4l2_frmivalenum *fival) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- const struct unicam_fmt *fmt; +- struct v4l2_subdev_frame_interval_enum fie = { +- .index = fival->index, +- .width = fival->width, +- .height = fival->height, +- .which = V4L2_SUBDEV_FORMAT_ACTIVE, +- }; +- int ret; +- +- fmt = find_format_by_pix(dev, fival->pixel_format); +- if (!fmt) +- return -EINVAL; +- +- fie.code = fmt->code; +- ret = v4l2_subdev_call(dev->sensor, pad, enum_frame_interval, +- NULL, &fie); +- if (ret) +- return ret; +- +- fival->type = V4L2_FRMIVAL_TYPE_DISCRETE; +- fival->discrete = fie.interval; +- +- return 0; +-} +- +-static int unicam_g_parm(struct file *file, void *fh, struct v4l2_streamparm *a) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_g_parm_cap(video_devdata(file), dev->sensor, a); +-} +- +-static int unicam_s_parm(struct file *file, void *fh, struct v4l2_streamparm *a) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_s_parm_cap(video_devdata(file), dev->sensor, a); +-} +- +-static int unicam_g_dv_timings(struct file *file, void *priv, +- struct v4l2_dv_timings *timings) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, video, g_dv_timings, timings); +-} +- +-static int unicam_s_dv_timings(struct file *file, void *priv, +- struct v4l2_dv_timings *timings) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_dv_timings current_timings; +- int ret; +- +- ret = v4l2_subdev_call(dev->sensor, video, g_dv_timings, +- ¤t_timings); +- +- if (v4l2_match_dv_timings(timings, ¤t_timings, 0, false)) +- return 0; +- +- if (vb2_is_busy(&node->buffer_queue)) +- return -EBUSY; +- +- ret = v4l2_subdev_call(dev->sensor, video, s_dv_timings, timings); +- +- /* Force recomputation of bytesperline */ +- node->v_fmt.fmt.pix.bytesperline = 0; +- +- unicam_reset_format(node); +- +- return ret; +-} +- +-static int unicam_query_dv_timings(struct file *file, void *priv, +- struct v4l2_dv_timings *timings) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, video, query_dv_timings, timings); +-} +- +-static int unicam_enum_dv_timings(struct file *file, void *priv, +- struct v4l2_enum_dv_timings *timings) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, pad, enum_dv_timings, timings); +-} +- +-static int unicam_dv_timings_cap(struct file *file, void *priv, +- struct v4l2_dv_timings_cap *cap) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- +- return v4l2_subdev_call(dev->sensor, pad, dv_timings_cap, cap); +-} +- +-static int unicam_subscribe_event(struct v4l2_fh *fh, +- const struct v4l2_event_subscription *sub) +-{ +- switch (sub->type) { +- case V4L2_EVENT_FRAME_SYNC: +- return v4l2_event_subscribe(fh, sub, 2, NULL); +- case V4L2_EVENT_SOURCE_CHANGE: +- return v4l2_event_subscribe(fh, sub, 4, NULL); +- } +- +- return v4l2_ctrl_subscribe_event(fh, sub); +-} +- +-static int unicam_log_status(struct file *file, void *fh) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct unicam_cfg *cfg = &dev->cfg; +- u32 reg; +- +- /* status for sub devices */ +- v4l2_device_call_all(&dev->v4l2_dev, 0, core, log_status); +- +- unicam_info(dev, "-----Receiver status-----\n"); +- unicam_info(dev, "V4L2 width/height: %ux%u\n", +- node->v_fmt.fmt.pix.width, node->v_fmt.fmt.pix.height); +- unicam_info(dev, "Mediabus format: %08x\n", node->fmt->code); +- unicam_info(dev, "V4L2 format: %08x\n", +- node->v_fmt.fmt.pix.pixelformat); +- reg = reg_read(&dev->cfg, UNICAM_IPIPE); +- unicam_info(dev, "Unpacking/packing: %u / %u\n", +- get_field(reg, UNICAM_PUM_MASK), +- get_field(reg, UNICAM_PPM_MASK)); +- unicam_info(dev, "----Live data----\n"); +- unicam_info(dev, "Programmed stride: %4u\n", +- reg_read(cfg, UNICAM_IBLS)); +- unicam_info(dev, "Detected resolution: %ux%u\n", +- reg_read(cfg, UNICAM_IHSTA), +- reg_read(cfg, UNICAM_IVSTA)); +- unicam_info(dev, "Write pointer: %08x\n", +- reg_read(cfg, UNICAM_IBWP)); +- +- return 0; +-} +- +-static void unicam_notify(struct v4l2_subdev *sd, +- unsigned int notification, void *arg) +-{ +- struct unicam_device *dev = +- container_of(sd->v4l2_dev, struct unicam_device, v4l2_dev); +- +- switch (notification) { +- case V4L2_DEVICE_NOTIFY_EVENT: +- v4l2_event_queue(&dev->node[IMAGE_PAD].video_dev, arg); +- break; +- default: +- break; +- } +-} +- +-static const struct vb2_ops unicam_video_qops = { +- .wait_prepare = vb2_ops_wait_prepare, +- .wait_finish = vb2_ops_wait_finish, +- .queue_setup = unicam_queue_setup, +- .buf_prepare = unicam_buffer_prepare, +- .buf_queue = unicam_buffer_queue, +- .start_streaming = unicam_start_streaming, +- .stop_streaming = unicam_stop_streaming, +-}; +- +-/* +- * unicam_open : This function is based on the v4l2_fh_open helper function. +- * It has been augmented to handle sensor subdevice power management, +- */ +-static int unicam_open(struct file *file) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- int ret; +- +- mutex_lock(&node->lock); +- +- ret = v4l2_fh_open(file); +- if (ret) { +- unicam_err(dev, "v4l2_fh_open failed\n"); +- goto unlock; +- } +- +- node->open++; +- +- if (!v4l2_fh_is_singular_file(file)) +- goto unlock; +- +- ret = v4l2_subdev_call(dev->sensor, core, s_power, 1); +- if (ret < 0 && ret != -ENOIOCTLCMD) { +- v4l2_fh_release(file); +- node->open--; +- goto unlock; +- } +- +- ret = 0; +- +-unlock: +- mutex_unlock(&node->lock); +- return ret; +-} +- +-static int unicam_release(struct file *file) +-{ +- struct unicam_node *node = video_drvdata(file); +- struct unicam_device *dev = node->dev; +- struct v4l2_subdev *sd = dev->sensor; +- bool fh_singular; +- int ret; +- +- mutex_lock(&node->lock); +- +- fh_singular = v4l2_fh_is_singular_file(file); +- +- ret = _vb2_fop_release(file, NULL); +- +- if (fh_singular) +- v4l2_subdev_call(sd, core, s_power, 0); +- +- node->open--; +- mutex_unlock(&node->lock); +- +- return ret; +-} +- +-/* unicam capture driver file operations */ +-static const struct v4l2_file_operations unicam_fops = { +- .owner = THIS_MODULE, +- .open = unicam_open, +- .release = unicam_release, +- .read = vb2_fop_read, +- .poll = vb2_fop_poll, +- .unlocked_ioctl = video_ioctl2, +- .mmap = vb2_fop_mmap, +-}; +- +-/* unicam capture ioctl operations */ +-static const struct v4l2_ioctl_ops unicam_ioctl_ops = { +- .vidioc_querycap = unicam_querycap, +- .vidioc_enum_fmt_vid_cap = unicam_enum_fmt_vid_cap, +- .vidioc_g_fmt_vid_cap = unicam_g_fmt_vid_cap, +- .vidioc_s_fmt_vid_cap = unicam_s_fmt_vid_cap, +- .vidioc_try_fmt_vid_cap = unicam_try_fmt_vid_cap, +- +- .vidioc_enum_fmt_meta_cap = unicam_enum_fmt_meta_cap, +- .vidioc_g_fmt_meta_cap = unicam_g_fmt_meta_cap, +- .vidioc_s_fmt_meta_cap = unicam_s_fmt_meta_cap, +- .vidioc_try_fmt_meta_cap = unicam_try_fmt_meta_cap, +- +- .vidioc_enum_input = unicam_enum_input, +- .vidioc_g_input = unicam_g_input, +- .vidioc_s_input = unicam_s_input, +- +- .vidioc_querystd = unicam_querystd, +- .vidioc_s_std = unicam_s_std, +- .vidioc_g_std = unicam_g_std, +- +- .vidioc_g_edid = unicam_g_edid, +- .vidioc_s_edid = unicam_s_edid, +- +- .vidioc_enum_framesizes = unicam_enum_framesizes, +- .vidioc_enum_frameintervals = unicam_enum_frameintervals, +- +- .vidioc_g_selection = unicam_g_selection, +- .vidioc_s_selection = unicam_s_selection, +- +- .vidioc_g_parm = unicam_g_parm, +- .vidioc_s_parm = unicam_s_parm, +- +- .vidioc_s_dv_timings = unicam_s_dv_timings, +- .vidioc_g_dv_timings = unicam_g_dv_timings, +- .vidioc_query_dv_timings = unicam_query_dv_timings, +- .vidioc_enum_dv_timings = unicam_enum_dv_timings, +- .vidioc_dv_timings_cap = unicam_dv_timings_cap, +- +- .vidioc_reqbufs = vb2_ioctl_reqbufs, +- .vidioc_create_bufs = vb2_ioctl_create_bufs, +- .vidioc_prepare_buf = vb2_ioctl_prepare_buf, +- .vidioc_querybuf = vb2_ioctl_querybuf, +- .vidioc_qbuf = vb2_ioctl_qbuf, +- .vidioc_dqbuf = vb2_ioctl_dqbuf, +- .vidioc_expbuf = vb2_ioctl_expbuf, +- .vidioc_streamon = vb2_ioctl_streamon, +- .vidioc_streamoff = vb2_ioctl_streamoff, +- +- .vidioc_log_status = unicam_log_status, +- .vidioc_subscribe_event = unicam_subscribe_event, +- .vidioc_unsubscribe_event = v4l2_event_unsubscribe, +-}; +- +-static int +-unicam_async_bound(struct v4l2_async_notifier *notifier, +- struct v4l2_subdev *subdev, +- struct v4l2_async_subdev *asd) +-{ +- struct unicam_device *unicam = container_of(notifier->v4l2_dev, +- struct unicam_device, v4l2_dev); +- +- if (unicam->sensor) { +- unicam_info(unicam, "Rejecting subdev %s (Already set!!)", +- subdev->name); +- return 0; +- } +- +- unicam->sensor = subdev; +- unicam_dbg(1, unicam, "Using sensor %s for capture\n", subdev->name); +- +- return 0; +-} +- +-static int register_node(struct unicam_device *unicam, struct unicam_node *node, +- enum v4l2_buf_type type, int pad_id) +-{ +- struct video_device *vdev; +- struct vb2_queue *q; +- struct v4l2_mbus_framefmt mbus_fmt = {0}; +- const struct unicam_fmt *fmt; +- int ret; +- +- if (unicam->sensor_embedded_data || pad_id != METADATA_PAD) { +- ret = __subdev_get_format(unicam, &mbus_fmt, pad_id); +- if (ret) { +- unicam_err(unicam, "Failed to get_format - ret %d\n", +- ret); +- return ret; +- } +- +- fmt = find_format_by_code(mbus_fmt.code); +- if (!fmt) { +- /* Find the first format that the sensor and unicam both +- * support +- */ +- fmt = get_first_supported_format(unicam); +- +- if (!fmt) +- /* No compatible formats */ +- return -EINVAL; +- +- mbus_fmt.code = fmt->code; +- ret = __subdev_set_format(unicam, &mbus_fmt, pad_id); +- if (ret) +- return -EINVAL; +- } +- if (mbus_fmt.field != V4L2_FIELD_NONE) { +- /* Interlaced not supported - disable it now. */ +- mbus_fmt.field = V4L2_FIELD_NONE; +- ret = __subdev_set_format(unicam, &mbus_fmt, pad_id); +- if (ret) +- return -EINVAL; +- } +- } else { +- /* Fix this node format as embedded data. */ +- fmt = find_format_by_code(MEDIA_BUS_FMT_SENSOR_DATA); +- } +- +- node->dev = unicam; +- node->pad_id = pad_id; +- node->fmt = fmt; +- if (fmt->fourcc) { +- if (fmt->fourcc != V4L2_META_FMT_SENSOR_DATA) +- node->v_fmt.fmt.pix.pixelformat = fmt->fourcc; +- else +- node->v_fmt.fmt.meta.dataformat = fmt->fourcc; +- } else { +- node->v_fmt.fmt.pix.pixelformat = fmt->repacked_fourcc; +- } +- +- /* Read current subdev format */ +- unicam_reset_format(node); +- +- if (v4l2_subdev_has_op(unicam->sensor, video, s_std)) { +- v4l2_std_id tvnorms; +- +- if (WARN_ON(!v4l2_subdev_has_op(unicam->sensor, video, +- g_tvnorms))) +- /* +- * Subdevice should not advertise s_std but not +- * g_tvnorms +- */ +- return -EINVAL; +- +- ret = v4l2_subdev_call(unicam->sensor, video, +- g_tvnorms, &tvnorms); +- if (WARN_ON(ret)) +- return -EINVAL; +- node->video_dev.tvnorms |= tvnorms; +- } +- +- spin_lock_init(&node->dma_queue_lock); +- mutex_init(&node->lock); +- +- vdev = &node->video_dev; +- if (pad_id == IMAGE_PAD) { +- /* Add controls from the subdevice */ +- ret = v4l2_ctrl_add_handler(&node->ctrl_handler, +- unicam->sensor->ctrl_handler, NULL, +- true); +- if (ret < 0) +- return ret; +- +- /* +- * If the sensor subdevice has any controls, associate the node +- * with the ctrl handler to allow access from userland. +- */ +- if (!list_empty(&node->ctrl_handler.ctrls)) +- vdev->ctrl_handler = &node->ctrl_handler; +- } +- +- q = &node->buffer_queue; +- q->type = type; +- q->io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ; +- q->drv_priv = node; +- q->ops = &unicam_video_qops; +- q->mem_ops = &vb2_dma_contig_memops; +- q->buf_struct_size = sizeof(struct unicam_buffer); +- q->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC; +- q->lock = &node->lock; +- q->min_buffers_needed = 1; +- q->dev = &unicam->pdev->dev; +- +- ret = vb2_queue_init(q); +- if (ret) { +- unicam_err(unicam, "vb2_queue_init() failed\n"); +- return ret; +- } +- +- INIT_LIST_HEAD(&node->dma_queue.active); +- +- vdev->release = video_device_release_empty; +- vdev->fops = &unicam_fops; +- vdev->ioctl_ops = &unicam_ioctl_ops; +- vdev->v4l2_dev = &unicam->v4l2_dev; +- vdev->vfl_dir = VFL_DIR_RX; +- vdev->queue = q; +- vdev->lock = &node->lock; +- vdev->device_caps = (pad_id == IMAGE_PAD) ? +- (V4L2_CAP_VIDEO_CAPTURE | V4L2_CAP_STREAMING) : +- (V4L2_CAP_META_CAPTURE | V4L2_CAP_STREAMING); +- +- /* Define the device names */ +- snprintf(vdev->name, sizeof(vdev->name), "%s-%s", UNICAM_MODULE_NAME, +- node->pad_id == IMAGE_PAD ? "image" : "embedded"); +- +- video_set_drvdata(vdev, node); +- vdev->entity.flags |= MEDIA_ENT_FL_DEFAULT; +- +- node->dummy_buf_cpu_addr = dma_alloc_coherent(&unicam->pdev->dev, +- DUMMY_BUF_SIZE, +- &node->dummy_buf_dma_addr, +- GFP_ATOMIC); +- if (!node->dummy_buf_cpu_addr) { +- unicam_err(unicam, "Unable to allocate dummy buffer.\n"); +- return -ENOMEM; +- } +- +- if (node->pad_id == METADATA_PAD) { +- v4l2_disable_ioctl(vdev, VIDIOC_DQEVENT); +- v4l2_disable_ioctl(vdev, VIDIOC_SUBSCRIBE_EVENT); +- v4l2_disable_ioctl(vdev, VIDIOC_UNSUBSCRIBE_EVENT); +- } +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, video, s_std)) { +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_STD); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_STD); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUMSTD); +- } +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, video, querystd)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_QUERYSTD); +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, video, s_dv_timings)) { +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_EDID); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_EDID); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_DV_TIMINGS_CAP); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_DV_TIMINGS); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_DV_TIMINGS); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUM_DV_TIMINGS); +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_QUERY_DV_TIMINGS); +- } +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, pad, enum_frame_interval)) +- v4l2_disable_ioctl(&node->video_dev, +- VIDIOC_ENUM_FRAMEINTERVALS); +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, video, g_frame_interval)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_PARM); +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, video, s_frame_interval)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_PARM); +- +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, pad, enum_frame_size)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_ENUM_FRAMESIZES); +- +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, pad, set_selection)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_S_SELECTION); +- +- if (node->pad_id == METADATA_PAD || +- !v4l2_subdev_has_op(unicam->sensor, pad, get_selection)) +- v4l2_disable_ioctl(&node->video_dev, VIDIOC_G_SELECTION); +- +- ret = video_register_device(vdev, VFL_TYPE_VIDEO, -1); +- if (ret) { +- unicam_err(unicam, "Unable to register video device.\n"); +- return ret; +- } +- node->registered = 1; +- +- if (unicam->sensor_embedded_data) { +- ret = media_create_pad_link(&unicam->sensor->entity, pad_id, +- &node->video_dev.entity, 0, +- MEDIA_LNK_FL_ENABLED | +- MEDIA_LNK_FL_IMMUTABLE); +- if (ret) +- unicam_err(unicam, "Unable to create pad links.\n"); +- } +- +- return ret; +-} +- +-static void unregister_nodes(struct unicam_device *unicam) +-{ +- struct unicam_node *node; +- int i; +- +- for (i = 0; i < MAX_NODES; i++) { +- node = &unicam->node[i]; +- if (node->dummy_buf_cpu_addr) { +- dma_free_coherent(&unicam->pdev->dev, DUMMY_BUF_SIZE, +- node->dummy_buf_cpu_addr, +- node->dummy_buf_dma_addr); +- } +- if (node->registered) { +- video_unregister_device(&node->video_dev); +- node->registered = 0; +- } +- } +-} +- +-static int unicam_probe_complete(struct unicam_device *unicam) +-{ +- int ret; +- +- v4l2_set_subdev_hostdata(unicam->sensor, unicam); +- +- unicam->v4l2_dev.notify = unicam_notify; +- +- unicam->sensor_config = v4l2_subdev_alloc_pad_config(unicam->sensor); +- if (!unicam->sensor_config) +- return -ENOMEM; +- +- unicam->sensor_embedded_data = (unicam->sensor->entity.num_pads >= 2); +- +- ret = register_node(unicam, &unicam->node[IMAGE_PAD], +- V4L2_BUF_TYPE_VIDEO_CAPTURE, IMAGE_PAD); +- if (ret) { +- unicam_err(unicam, "Unable to register subdev node 0.\n"); +- goto unregister; +- } +- +- ret = register_node(unicam, &unicam->node[METADATA_PAD], +- V4L2_BUF_TYPE_META_CAPTURE, METADATA_PAD); +- if (ret) { +- unicam_err(unicam, "Unable to register subdev node 1.\n"); +- goto unregister; +- } +- +- ret = v4l2_device_register_ro_subdev_nodes(&unicam->v4l2_dev); +- if (ret) { +- unicam_err(unicam, "Unable to register subdev nodes.\n"); +- goto unregister; +- } +- +- return 0; +- +-unregister: +- unregister_nodes(unicam); +- +- return ret; +-} +- +-static int unicam_async_complete(struct v4l2_async_notifier *notifier) +-{ +- struct unicam_device *unicam = container_of(notifier->v4l2_dev, +- struct unicam_device, v4l2_dev); +- +- return unicam_probe_complete(unicam); +-} +- +-static const struct v4l2_async_notifier_operations unicam_async_ops = { +- .bound = unicam_async_bound, +- .complete = unicam_async_complete, +-}; +- +-static int of_unicam_connect_subdevs(struct unicam_device *dev) +-{ +- struct platform_device *pdev = dev->pdev; +- struct device_node *parent, *ep_node = NULL, *remote_ep = NULL, +- *sensor_node = NULL; +- struct v4l2_fwnode_endpoint *ep; +- struct v4l2_async_subdev *asd; +- unsigned int peripheral_data_lanes; +- int ret = -EINVAL; +- unsigned int lane; +- +- parent = pdev->dev.of_node; +- +- asd = &dev->asd; +- ep = &dev->endpoint; +- +- ep_node = of_graph_get_next_endpoint(parent, NULL); +- if (!ep_node) { +- unicam_dbg(3, dev, "can't get next endpoint\n"); +- goto cleanup_exit; +- } +- +- unicam_dbg(3, dev, "ep_node is %s\n", ep_node->name); +- +- v4l2_fwnode_endpoint_parse(of_fwnode_handle(ep_node), ep); +- +- for (lane = 0; lane < ep->bus.mipi_csi2.num_data_lanes; lane++) { +- if (ep->bus.mipi_csi2.data_lanes[lane] != lane + 1) { +- unicam_err(dev, "Local endpoint - data lane reordering not supported\n"); +- goto cleanup_exit; +- } +- } +- +- peripheral_data_lanes = ep->bus.mipi_csi2.num_data_lanes; +- +- sensor_node = of_graph_get_remote_port_parent(ep_node); +- if (!sensor_node) { +- unicam_dbg(3, dev, "can't get remote parent\n"); +- goto cleanup_exit; +- } +- unicam_dbg(3, dev, "sensor_node is %s\n", sensor_node->name); +- asd->match_type = V4L2_ASYNC_MATCH_FWNODE; +- asd->match.fwnode = of_fwnode_handle(sensor_node); +- +- remote_ep = of_graph_get_remote_endpoint(ep_node); +- if (!remote_ep) { +- unicam_dbg(3, dev, "can't get remote-endpoint\n"); +- goto cleanup_exit; +- } +- unicam_dbg(3, dev, "remote_ep is %s\n", remote_ep->name); +- v4l2_fwnode_endpoint_parse(of_fwnode_handle(remote_ep), ep); +- unicam_dbg(3, dev, "parsed remote_ep to endpoint. nr_of_link_frequencies %u, bus_type %u\n", +- ep->nr_of_link_frequencies, ep->bus_type); +- +- switch (ep->bus_type) { +- case V4L2_MBUS_CSI2_DPHY: +- if (ep->bus.mipi_csi2.num_data_lanes > +- peripheral_data_lanes) { +- unicam_err(dev, "Subdevice %s wants too many data lanes (%u > %u)\n", +- sensor_node->name, +- ep->bus.mipi_csi2.num_data_lanes, +- peripheral_data_lanes); +- goto cleanup_exit; +- } +- for (lane = 0; +- lane < ep->bus.mipi_csi2.num_data_lanes; +- lane++) { +- if (ep->bus.mipi_csi2.data_lanes[lane] != lane + 1) { +- unicam_err(dev, "Subdevice %s - incompatible data lane config\n", +- sensor_node->name); +- goto cleanup_exit; +- } +- } +- dev->max_data_lanes = ep->bus.mipi_csi2.num_data_lanes; +- dev->bus_flags = ep->bus.mipi_csi2.flags; +- break; +- case V4L2_MBUS_CCP2: +- if (ep->bus.mipi_csi1.clock_lane != 0 || +- ep->bus.mipi_csi1.data_lane != 1) { +- unicam_err(dev, "Subdevice %s incompatible lane config\n", +- sensor_node->name); +- goto cleanup_exit; +- } +- dev->max_data_lanes = 1; +- dev->bus_flags = ep->bus.mipi_csi1.strobe; +- break; +- default: +- /* Unsupported bus type */ +- unicam_err(dev, "sub-device %s is not a CSI2 or CCP2 device %d\n", +- sensor_node->name, ep->bus_type); +- goto cleanup_exit; +- } +- +- /* Store bus type - CSI2 or CCP2 */ +- dev->bus_type = ep->bus_type; +- unicam_dbg(3, dev, "bus_type is %d\n", dev->bus_type); +- +- /* Store Virtual Channel number */ +- dev->virtual_channel = ep->base.id; +- +- unicam_dbg(3, dev, "v4l2-endpoint: %s\n", +- dev->bus_type == V4L2_MBUS_CSI2_DPHY ? "CSI2" : "CCP2"); +- unicam_dbg(3, dev, "Virtual Channel=%d\n", dev->virtual_channel); +- if (dev->bus_type == V4L2_MBUS_CSI2_DPHY) +- unicam_dbg(3, dev, "flags=0x%08x\n", ep->bus.mipi_csi2.flags); +- unicam_dbg(3, dev, "num_data_lanes=%d\n", dev->max_data_lanes); +- +- unicam_dbg(1, dev, "found sub-device %s\n", sensor_node->name); +- +- v4l2_async_notifier_init(&dev->notifier); +- +- ret = v4l2_async_notifier_add_subdev(&dev->notifier, asd); +- if (ret) { +- unicam_err(dev, "Error adding subdevice - ret %d\n", ret); +- goto cleanup_exit; +- } +- +- dev->notifier.ops = &unicam_async_ops; +- ret = v4l2_async_notifier_register(&dev->v4l2_dev, +- &dev->notifier); +- if (ret) { +- unicam_err(dev, "Error registering async notifier - ret %d\n", +- ret); +- ret = -EINVAL; +- } +- +-cleanup_exit: +- if (remote_ep) +- of_node_put(remote_ep); +- if (sensor_node) +- of_node_put(sensor_node); +- if (ep_node) +- of_node_put(ep_node); +- +- return ret; +-} +- +-static int unicam_probe(struct platform_device *pdev) +-{ +- struct unicam_cfg *unicam_cfg; +- struct unicam_device *unicam; +- struct v4l2_ctrl_handler *hdl; +- struct resource *res; +- int ret; +- +- unicam = devm_kzalloc(&pdev->dev, sizeof(*unicam), GFP_KERNEL); +- if (!unicam) +- return -ENOMEM; +- +- unicam->pdev = pdev; +- unicam_cfg = &unicam->cfg; +- +- res = platform_get_resource(pdev, IORESOURCE_MEM, 0); +- unicam_cfg->base = devm_ioremap_resource(&pdev->dev, res); +- if (IS_ERR(unicam_cfg->base)) { +- unicam_err(unicam, "Failed to get main io block\n"); +- return PTR_ERR(unicam_cfg->base); +- } +- +- res = platform_get_resource(pdev, IORESOURCE_MEM, 1); +- unicam_cfg->clk_gate_base = devm_ioremap_resource(&pdev->dev, res); +- if (IS_ERR(unicam_cfg->clk_gate_base)) { +- unicam_err(unicam, "Failed to get 2nd io block\n"); +- return PTR_ERR(unicam_cfg->clk_gate_base); +- } +- +- unicam->clock = devm_clk_get(&pdev->dev, "lp"); +- if (IS_ERR(unicam->clock)) { +- unicam_err(unicam, "Failed to get clock\n"); +- return PTR_ERR(unicam->clock); +- } +- +- ret = platform_get_irq(pdev, 0); +- if (ret <= 0) { +- dev_err(&pdev->dev, "No IRQ resource\n"); +- return -ENODEV; +- } +- +- ret = devm_request_irq(&pdev->dev, ret, unicam_isr, 0, +- "unicam_capture0", unicam); +- if (ret) { +- dev_err(&pdev->dev, "Unable to request interrupt\n"); +- return -EINVAL; +- } +- +- unicam->mdev.dev = &pdev->dev; +- strscpy(unicam->mdev.model, UNICAM_MODULE_NAME, +- sizeof(unicam->mdev.model)); +- strscpy(unicam->mdev.serial, "", sizeof(unicam->mdev.serial)); +- snprintf(unicam->mdev.bus_info, sizeof(unicam->mdev.bus_info), +- "platform:%s %s", +- pdev->dev.driver->name, dev_name(&pdev->dev)); +- unicam->mdev.hw_revision = 1; +- +- media_entity_pads_init(&unicam->node[IMAGE_PAD].video_dev.entity, 1, +- &unicam->node[IMAGE_PAD].pad); +- media_entity_pads_init(&unicam->node[METADATA_PAD].video_dev.entity, 1, +- &unicam->node[METADATA_PAD].pad); +- media_device_init(&unicam->mdev); +- +- unicam->v4l2_dev.mdev = &unicam->mdev; +- +- ret = v4l2_device_register(&pdev->dev, &unicam->v4l2_dev); +- if (ret) { +- unicam_err(unicam, +- "Unable to register v4l2 device.\n"); +- goto media_cleanup; +- } +- +- ret = media_device_register(&unicam->mdev); +- if (ret < 0) { +- unicam_err(unicam, +- "Unable to register media-controller device.\n"); +- goto probe_out_v4l2_unregister; +- } +- +- /* Reserve space for the controls */ +- hdl = &unicam->node[IMAGE_PAD].ctrl_handler; +- ret = v4l2_ctrl_handler_init(hdl, 16); +- if (ret < 0) +- goto media_unregister; +- +- /* set the driver data in platform device */ +- platform_set_drvdata(pdev, unicam); +- +- ret = of_unicam_connect_subdevs(unicam); +- if (ret) { +- dev_err(&pdev->dev, "Failed to connect subdevs\n"); +- goto free_hdl; +- } +- +- /* Enable the block power domain */ +- pm_runtime_enable(&pdev->dev); +- +- return 0; +- +-free_hdl: +- v4l2_ctrl_handler_free(hdl); +-media_unregister: +- media_device_unregister(&unicam->mdev); +-probe_out_v4l2_unregister: +- v4l2_device_unregister(&unicam->v4l2_dev); +-media_cleanup: +- media_device_cleanup(&unicam->mdev); +- +- return ret; +-} +- +-static int unicam_remove(struct platform_device *pdev) +-{ +- struct unicam_device *unicam = platform_get_drvdata(pdev); +- +- unicam_dbg(2, unicam, "%s\n", __func__); +- +- pm_runtime_disable(&pdev->dev); +- +- v4l2_async_notifier_unregister(&unicam->notifier); +- v4l2_ctrl_handler_free(&unicam->node[IMAGE_PAD].ctrl_handler); +- v4l2_device_unregister(&unicam->v4l2_dev); +- unregister_nodes(unicam); +- if (unicam->sensor_config) +- v4l2_subdev_free_pad_config(unicam->sensor_config); +- media_device_unregister(&unicam->mdev); +- media_device_cleanup(&unicam->mdev); +- +- return 0; +-} +- +-static const struct of_device_id unicam_of_match[] = { +- { .compatible = "brcm,bcm2835-unicam", }, +- { /* sentinel */ }, +-}; +-MODULE_DEVICE_TABLE(of, unicam_of_match); +- +-static struct platform_driver unicam_driver = { +- .probe = unicam_probe, +- .remove = unicam_remove, +- .driver = { +- .name = UNICAM_MODULE_NAME, +- .of_match_table = of_match_ptr(unicam_of_match), +- }, +-}; +- +-module_platform_driver(unicam_driver); +- +-MODULE_AUTHOR("Dave Stevenson <dave.stevenson@raspberrypi.com>"); +-MODULE_DESCRIPTION("BCM2835 Unicam driver"); +-MODULE_LICENSE("GPL"); +-MODULE_VERSION(UNICAM_VERSION); +--- a/drivers/media/platform/bcm2835/vc4-regs-unicam.h ++++ /dev/null +@@ -1,253 +0,0 @@ +-/* SPDX-License-Identifier: GPL-2.0-only */ +- +-/* +- * Copyright (C) 2017-2020 Raspberry Pi Trading. +- * Dave Stevenson <dave.stevenson@raspberrypi.com> +- */ +- +-#ifndef VC4_REGS_UNICAM_H +-#define VC4_REGS_UNICAM_H +- +-/* +- * The following values are taken from files found within the code drop +- * made by Broadcom for the BCM21553 Graphics Driver, predominantly in +- * brcm_usrlib/dag/vmcsx/vcinclude/hardware_vc4.h. +- * They have been modified to be only the register offset. +- */ +-#define UNICAM_CTRL 0x000 +-#define UNICAM_STA 0x004 +-#define UNICAM_ANA 0x008 +-#define UNICAM_PRI 0x00c +-#define UNICAM_CLK 0x010 +-#define UNICAM_CLT 0x014 +-#define UNICAM_DAT0 0x018 +-#define UNICAM_DAT1 0x01c +-#define UNICAM_DAT2 0x020 +-#define UNICAM_DAT3 0x024 +-#define UNICAM_DLT 0x028 +-#define UNICAM_CMP0 0x02c +-#define UNICAM_CMP1 0x030 +-#define UNICAM_CAP0 0x034 +-#define UNICAM_CAP1 0x038 +-#define UNICAM_ICTL 0x100 +-#define UNICAM_ISTA 0x104 +-#define UNICAM_IDI0 0x108 +-#define UNICAM_IPIPE 0x10c +-#define UNICAM_IBSA0 0x110 +-#define UNICAM_IBEA0 0x114 +-#define UNICAM_IBLS 0x118 +-#define UNICAM_IBWP 0x11c +-#define UNICAM_IHWIN 0x120 +-#define UNICAM_IHSTA 0x124 +-#define UNICAM_IVWIN 0x128 +-#define UNICAM_IVSTA 0x12c +-#define UNICAM_ICC 0x130 +-#define UNICAM_ICS 0x134 +-#define UNICAM_IDC 0x138 +-#define UNICAM_IDPO 0x13c +-#define UNICAM_IDCA 0x140 +-#define UNICAM_IDCD 0x144 +-#define UNICAM_IDS 0x148 +-#define UNICAM_DCS 0x200 +-#define UNICAM_DBSA0 0x204 +-#define UNICAM_DBEA0 0x208 +-#define UNICAM_DBWP 0x20c +-#define UNICAM_DBCTL 0x300 +-#define UNICAM_IBSA1 0x304 +-#define UNICAM_IBEA1 0x308 +-#define UNICAM_IDI1 0x30c +-#define UNICAM_DBSA1 0x310 +-#define UNICAM_DBEA1 0x314 +-#define UNICAM_MISC 0x400 +- +-/* +- * The following bitmasks are from the kernel released by Broadcom +- * for Android - https://android.googlesource.com/kernel/bcm/ +- * The Rhea, Hawaii, and Java chips all contain the same VideoCore4 +- * Unicam block as BCM2835, as defined in eg +- * arch/arm/mach-rhea/include/mach/rdb_A0/brcm_rdb_cam.h and similar. +- * Values reworked to use the kernel BIT and GENMASK macros. +- * +- * Some of the bit mnenomics have been amended to match the datasheet. +- */ +-/* UNICAM_CTRL Register */ +-#define UNICAM_CPE BIT(0) +-#define UNICAM_MEM BIT(1) +-#define UNICAM_CPR BIT(2) +-#define UNICAM_CPM_MASK GENMASK(3, 3) +-#define UNICAM_CPM_CSI2 0 +-#define UNICAM_CPM_CCP2 1 +-#define UNICAM_SOE BIT(4) +-#define UNICAM_DCM_MASK GENMASK(5, 5) +-#define UNICAM_DCM_STROBE 0 +-#define UNICAM_DCM_DATA 1 +-#define UNICAM_SLS BIT(6) +-#define UNICAM_PFT_MASK GENMASK(11, 8) +-#define UNICAM_OET_MASK GENMASK(20, 12) +- +-/* UNICAM_STA Register */ +-#define UNICAM_SYN BIT(0) +-#define UNICAM_CS BIT(1) +-#define UNICAM_SBE BIT(2) +-#define UNICAM_PBE BIT(3) +-#define UNICAM_HOE BIT(4) +-#define UNICAM_PLE BIT(5) +-#define UNICAM_SSC BIT(6) +-#define UNICAM_CRCE BIT(7) +-#define UNICAM_OES BIT(8) +-#define UNICAM_IFO BIT(9) +-#define UNICAM_OFO BIT(10) +-#define UNICAM_BFO BIT(11) +-#define UNICAM_DL BIT(12) +-#define UNICAM_PS BIT(13) +-#define UNICAM_IS BIT(14) +-#define UNICAM_PI0 BIT(15) +-#define UNICAM_PI1 BIT(16) +-#define UNICAM_FSI_S BIT(17) +-#define UNICAM_FEI_S BIT(18) +-#define UNICAM_LCI_S BIT(19) +-#define UNICAM_BUF0_RDY BIT(20) +-#define UNICAM_BUF0_NO BIT(21) +-#define UNICAM_BUF1_RDY BIT(22) +-#define UNICAM_BUF1_NO BIT(23) +-#define UNICAM_DI BIT(24) +- +-#define UNICAM_STA_MASK_ALL \ +- (UNICAM_DL + \ +- UNICAM_SBE + \ +- UNICAM_PBE + \ +- UNICAM_HOE + \ +- UNICAM_PLE + \ +- UNICAM_SSC + \ +- UNICAM_CRCE + \ +- UNICAM_IFO + \ +- UNICAM_OFO + \ +- UNICAM_PS + \ +- UNICAM_PI0 + \ +- UNICAM_PI1) +- +-/* UNICAM_ANA Register */ +-#define UNICAM_APD BIT(0) +-#define UNICAM_BPD BIT(1) +-#define UNICAM_AR BIT(2) +-#define UNICAM_DDL BIT(3) +-#define UNICAM_CTATADJ_MASK GENMASK(7, 4) +-#define UNICAM_PTATADJ_MASK GENMASK(11, 8) +- +-/* UNICAM_PRI Register */ +-#define UNICAM_PE BIT(0) +-#define UNICAM_PT_MASK GENMASK(2, 1) +-#define UNICAM_NP_MASK GENMASK(7, 4) +-#define UNICAM_PP_MASK GENMASK(11, 8) +-#define UNICAM_BS_MASK GENMASK(15, 12) +-#define UNICAM_BL_MASK GENMASK(17, 16) +- +-/* UNICAM_CLK Register */ +-#define UNICAM_CLE BIT(0) +-#define UNICAM_CLPD BIT(1) +-#define UNICAM_CLLPE BIT(2) +-#define UNICAM_CLHSE BIT(3) +-#define UNICAM_CLTRE BIT(4) +-#define UNICAM_CLAC_MASK GENMASK(8, 5) +-#define UNICAM_CLSTE BIT(29) +- +-/* UNICAM_CLT Register */ +-#define UNICAM_CLT1_MASK GENMASK(7, 0) +-#define UNICAM_CLT2_MASK GENMASK(15, 8) +- +-/* UNICAM_DATn Registers */ +-#define UNICAM_DLE BIT(0) +-#define UNICAM_DLPD BIT(1) +-#define UNICAM_DLLPE BIT(2) +-#define UNICAM_DLHSE BIT(3) +-#define UNICAM_DLTRE BIT(4) +-#define UNICAM_DLSM BIT(5) +-#define UNICAM_DLFO BIT(28) +-#define UNICAM_DLSTE BIT(29) +- +-#define UNICAM_DAT_MASK_ALL (UNICAM_DLSTE + UNICAM_DLFO) +- +-/* UNICAM_DLT Register */ +-#define UNICAM_DLT1_MASK GENMASK(7, 0) +-#define UNICAM_DLT2_MASK GENMASK(15, 8) +-#define UNICAM_DLT3_MASK GENMASK(23, 16) +- +-/* UNICAM_ICTL Register */ +-#define UNICAM_FSIE BIT(0) +-#define UNICAM_FEIE BIT(1) +-#define UNICAM_IBOB BIT(2) +-#define UNICAM_FCM BIT(3) +-#define UNICAM_TFC BIT(4) +-#define UNICAM_LIP_MASK GENMASK(6, 5) +-#define UNICAM_LCIE_MASK GENMASK(28, 16) +- +-/* UNICAM_IDI0/1 Register */ +-#define UNICAM_ID0_MASK GENMASK(7, 0) +-#define UNICAM_ID1_MASK GENMASK(15, 8) +-#define UNICAM_ID2_MASK GENMASK(23, 16) +-#define UNICAM_ID3_MASK GENMASK(31, 24) +- +-/* UNICAM_ISTA Register */ +-#define UNICAM_FSI BIT(0) +-#define UNICAM_FEI BIT(1) +-#define UNICAM_LCI BIT(2) +- +-#define UNICAM_ISTA_MASK_ALL (UNICAM_FSI + UNICAM_FEI + UNICAM_LCI) +- +-/* UNICAM_IPIPE Register */ +-#define UNICAM_PUM_MASK GENMASK(2, 0) +- /* Unpacking modes */ +- #define UNICAM_PUM_NONE 0 +- #define UNICAM_PUM_UNPACK6 1 +- #define UNICAM_PUM_UNPACK7 2 +- #define UNICAM_PUM_UNPACK8 3 +- #define UNICAM_PUM_UNPACK10 4 +- #define UNICAM_PUM_UNPACK12 5 +- #define UNICAM_PUM_UNPACK14 6 +- #define UNICAM_PUM_UNPACK16 7 +-#define UNICAM_DDM_MASK GENMASK(6, 3) +-#define UNICAM_PPM_MASK GENMASK(9, 7) +- /* Packing modes */ +- #define UNICAM_PPM_NONE 0 +- #define UNICAM_PPM_PACK8 1 +- #define UNICAM_PPM_PACK10 2 +- #define UNICAM_PPM_PACK12 3 +- #define UNICAM_PPM_PACK14 4 +- #define UNICAM_PPM_PACK16 5 +-#define UNICAM_DEM_MASK GENMASK(11, 10) +-#define UNICAM_DEBL_MASK GENMASK(14, 12) +-#define UNICAM_ICM_MASK GENMASK(16, 15) +-#define UNICAM_IDM_MASK GENMASK(17, 17) +- +-/* UNICAM_ICC Register */ +-#define UNICAM_ICFL_MASK GENMASK(4, 0) +-#define UNICAM_ICFH_MASK GENMASK(9, 5) +-#define UNICAM_ICST_MASK GENMASK(12, 10) +-#define UNICAM_ICLT_MASK GENMASK(15, 13) +-#define UNICAM_ICLL_MASK GENMASK(31, 16) +- +-/* UNICAM_DCS Register */ +-#define UNICAM_DIE BIT(0) +-#define UNICAM_DIM BIT(1) +-#define UNICAM_DBOB BIT(3) +-#define UNICAM_FDE BIT(4) +-#define UNICAM_LDP BIT(5) +-#define UNICAM_EDL_MASK GENMASK(15, 8) +- +-/* UNICAM_DBCTL Register */ +-#define UNICAM_DBEN BIT(0) +-#define UNICAM_BUF0_IE BIT(1) +-#define UNICAM_BUF1_IE BIT(2) +- +-/* UNICAM_CMP[0,1] register */ +-#define UNICAM_PCE BIT(31) +-#define UNICAM_GI BIT(9) +-#define UNICAM_CPH BIT(8) +-#define UNICAM_PCVC_MASK GENMASK(7, 6) +-#define UNICAM_PCDT_MASK GENMASK(5, 0) +- +-/* UNICAM_MISC register */ +-#define UNICAM_FL0 BIT(6) +-#define UNICAM_FL1 BIT(9) +- +-#endif |