aboutsummaryrefslogtreecommitdiffstats
path: root/target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch
diff options
context:
space:
mode:
authorÁlvaro Fernández Rojas <noltari@gmail.com>2016-04-07 21:25:10 +0200
committerÁlvaro Fernández Rojas <noltari@gmail.com>2016-04-07 21:25:10 +0200
commita8d4d71c41ff0158c2026cac5981e39702167da9 (patch)
tree0c54ec3eb59f5bab6aa7318d14c92ff875412202 /target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch
parent59e0e88c22007fd77ee9c6c8e02a689889a5f597 (diff)
downloadupstream-a8d4d71c41ff0158c2026cac5981e39702167da9.tar.gz
upstream-a8d4d71c41ff0158c2026cac5981e39702167da9.tar.bz2
upstream-a8d4d71c41ff0158c2026cac5981e39702167da9.zip
brcm2708: update to latest version
As usual these patches were extracted from the raspberry pi repo: https://github.com/raspberrypi/linux/commits/rpi-4.4.y Signed-off-by: Álvaro Fernández Rojas <noltari@gmail.com>
Diffstat (limited to 'target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch')
-rw-r--r--target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch612
1 files changed, 612 insertions, 0 deletions
diff --git a/target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch b/target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch
new file mode 100644
index 0000000000..44b3e5a069
--- /dev/null
+++ b/target/linux/brcm2708/patches-4.4/0119-drm-vc4-Synchronize-validation-code-for-v2-submissio.patch
@@ -0,0 +1,612 @@
+From 76172185d9852c3e2bf7ace7c040b62dd220d9aa Mon Sep 17 00:00:00 2001
+From: Eric Anholt <eric@anholt.net>
+Date: Mon, 7 Dec 2015 12:35:01 -0800
+Subject: [PATCH 119/232] drm/vc4: Synchronize validation code for v2
+ submission upstream.
+
+Signed-off-by: Eric Anholt <eric@anholt.net>
+---
+ drivers/gpu/drm/vc4/vc4_drv.h | 24 +--
+ drivers/gpu/drm/vc4/vc4_gem.c | 14 +-
+ drivers/gpu/drm/vc4/vc4_render_cl.c | 6 +-
+ drivers/gpu/drm/vc4/vc4_validate.c | 287 +++++++++++++++---------------------
+ 4 files changed, 135 insertions(+), 196 deletions(-)
+
+--- a/drivers/gpu/drm/vc4/vc4_drv.h
++++ b/drivers/gpu/drm/vc4/vc4_drv.h
+@@ -189,17 +189,6 @@ to_vc4_encoder(struct drm_encoder *encod
+ #define HVS_READ(offset) readl(vc4->hvs->regs + offset)
+ #define HVS_WRITE(offset, val) writel(val, vc4->hvs->regs + offset)
+
+-enum vc4_bo_mode {
+- VC4_MODE_UNDECIDED,
+- VC4_MODE_RENDER,
+- VC4_MODE_SHADER,
+-};
+-
+-struct vc4_bo_exec_state {
+- struct drm_gem_cma_object *bo;
+- enum vc4_bo_mode mode;
+-};
+-
+ struct vc4_exec_info {
+ /* Sequence number for this bin/render job. */
+ uint64_t seqno;
+@@ -210,7 +199,7 @@ struct vc4_exec_info {
+ /* This is the array of BOs that were looked up at the start of exec.
+ * Command validation will use indices into this array.
+ */
+- struct vc4_bo_exec_state *bo;
++ struct drm_gem_cma_object **bo;
+ uint32_t bo_count;
+
+ /* Pointers for our position in vc4->job_list */
+@@ -238,7 +227,6 @@ struct vc4_exec_info {
+ * command lists.
+ */
+ struct vc4_shader_state {
+- uint8_t packet;
+ uint32_t addr;
+ /* Maximum vertex index referenced by any primitive using this
+ * shader state.
+@@ -254,6 +242,7 @@ struct vc4_exec_info {
+ bool found_tile_binning_mode_config_packet;
+ bool found_start_tile_binning_packet;
+ bool found_increment_semaphore_packet;
++ bool found_flush;
+ uint8_t bin_tiles_x, bin_tiles_y;
+ struct drm_gem_cma_object *tile_bo;
+ uint32_t tile_alloc_offset;
+@@ -265,6 +254,9 @@ struct vc4_exec_info {
+ uint32_t ct0ca, ct0ea;
+ uint32_t ct1ca, ct1ea;
+
++ /* Pointer to the unvalidated bin CL (if present). */
++ void *bin_u;
++
+ /* Pointers to the shader recs. These paddr gets incremented as CL
+ * packets are relocated in validate_gl_shader_state, and the vaddrs
+ * (u and v) get incremented and size decremented as the shader recs
+@@ -455,10 +447,8 @@ vc4_validate_bin_cl(struct drm_device *d
+ int
+ vc4_validate_shader_recs(struct drm_device *dev, struct vc4_exec_info *exec);
+
+-bool vc4_use_bo(struct vc4_exec_info *exec,
+- uint32_t hindex,
+- enum vc4_bo_mode mode,
+- struct drm_gem_cma_object **obj);
++struct drm_gem_cma_object *vc4_use_bo(struct vc4_exec_info *exec,
++ uint32_t hindex);
+
+ int vc4_get_rcl(struct drm_device *dev, struct vc4_exec_info *exec);
+
+--- a/drivers/gpu/drm/vc4/vc4_gem.c
++++ b/drivers/gpu/drm/vc4/vc4_gem.c
+@@ -169,8 +169,8 @@ vc4_save_hang_state(struct drm_device *d
+ }
+
+ for (i = 0; i < exec->bo_count; i++) {
+- drm_gem_object_reference(&exec->bo[i].bo->base);
+- kernel_state->bo[i] = &exec->bo[i].bo->base;
++ drm_gem_object_reference(&exec->bo[i]->base);
++ kernel_state->bo[i] = &exec->bo[i]->base;
+ }
+
+ list_for_each_entry(bo, &exec->unref_list, unref_head) {
+@@ -397,7 +397,7 @@ vc4_update_bo_seqnos(struct vc4_exec_inf
+ unsigned i;
+
+ for (i = 0; i < exec->bo_count; i++) {
+- bo = to_vc4_bo(&exec->bo[i].bo->base);
++ bo = to_vc4_bo(&exec->bo[i]->base);
+ bo->seqno = seqno;
+ }
+
+@@ -467,7 +467,7 @@ vc4_cl_lookup_bos(struct drm_device *dev
+ return -EINVAL;
+ }
+
+- exec->bo = kcalloc(exec->bo_count, sizeof(struct vc4_bo_exec_state),
++ exec->bo = kcalloc(exec->bo_count, sizeof(struct drm_gem_cma_object *),
+ GFP_KERNEL);
+ if (!exec->bo) {
+ DRM_ERROR("Failed to allocate validated BO pointers\n");
+@@ -500,7 +500,7 @@ vc4_cl_lookup_bos(struct drm_device *dev
+ goto fail;
+ }
+ drm_gem_object_reference(bo);
+- exec->bo[i].bo = (struct drm_gem_cma_object *)bo;
++ exec->bo[i] = (struct drm_gem_cma_object *)bo;
+ }
+ spin_unlock(&file_priv->table_lock);
+
+@@ -591,6 +591,8 @@ vc4_get_bcl(struct drm_device *dev, stru
+
+ exec->ct0ca = exec->exec_bo->paddr + bin_offset;
+
++ exec->bin_u = bin;
++
+ exec->shader_rec_v = exec->exec_bo->vaddr + shader_rec_offset;
+ exec->shader_rec_p = exec->exec_bo->paddr + shader_rec_offset;
+ exec->shader_rec_size = args->shader_rec_size;
+@@ -622,7 +624,7 @@ vc4_complete_exec(struct drm_device *dev
+ mutex_lock(&dev->struct_mutex);
+ if (exec->bo) {
+ for (i = 0; i < exec->bo_count; i++)
+- drm_gem_object_unreference(&exec->bo[i].bo->base);
++ drm_gem_object_unreference(&exec->bo[i]->base);
+ kfree(exec->bo);
+ }
+
+--- a/drivers/gpu/drm/vc4/vc4_render_cl.c
++++ b/drivers/gpu/drm/vc4/vc4_render_cl.c
+@@ -436,7 +436,8 @@ static int vc4_rcl_surface_setup(struct
+ if (surf->hindex == ~0)
+ return 0;
+
+- if (!vc4_use_bo(exec, surf->hindex, VC4_MODE_RENDER, obj))
++ *obj = vc4_use_bo(exec, surf->hindex);
++ if (!*obj)
+ return -EINVAL;
+
+ if (surf->flags & VC4_SUBMIT_RCL_SURFACE_READ_IS_FULL_RES) {
+@@ -537,7 +538,8 @@ vc4_rcl_render_config_surface_setup(stru
+ if (surf->hindex == ~0)
+ return 0;
+
+- if (!vc4_use_bo(exec, surf->hindex, VC4_MODE_RENDER, obj))
++ *obj = vc4_use_bo(exec, surf->hindex);
++ if (!*obj)
+ return -EINVAL;
+
+ if (tiling > VC4_TILING_FORMAT_LT) {
+--- a/drivers/gpu/drm/vc4/vc4_validate.c
++++ b/drivers/gpu/drm/vc4/vc4_validate.c
+@@ -94,42 +94,42 @@ size_is_lt(uint32_t width, uint32_t heig
+ height <= 4 * utile_height(cpp));
+ }
+
+-bool
+-vc4_use_bo(struct vc4_exec_info *exec,
+- uint32_t hindex,
+- enum vc4_bo_mode mode,
+- struct drm_gem_cma_object **obj)
++struct drm_gem_cma_object *
++vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex)
+ {
+- *obj = NULL;
++ struct drm_gem_cma_object *obj;
++ struct vc4_bo *bo;
+
+ if (hindex >= exec->bo_count) {
+ DRM_ERROR("BO index %d greater than BO count %d\n",
+ hindex, exec->bo_count);
+- return false;
++ return NULL;
+ }
++ obj = exec->bo[hindex];
++ bo = to_vc4_bo(&obj->base);
+
+- if (exec->bo[hindex].mode != mode) {
+- if (exec->bo[hindex].mode == VC4_MODE_UNDECIDED) {
+- exec->bo[hindex].mode = mode;
+- } else {
+- DRM_ERROR("BO index %d reused with mode %d vs %d\n",
+- hindex, exec->bo[hindex].mode, mode);
+- return false;
+- }
++ if (bo->validated_shader) {
++ DRM_ERROR("Trying to use shader BO as something other than "
++ "a shader\n");
++ return NULL;
+ }
+
+- *obj = exec->bo[hindex].bo;
+- return true;
++ return obj;
++}
++
++static struct drm_gem_cma_object *
++vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index)
++{
++ return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]);
+ }
+
+ static bool
+-vc4_use_handle(struct vc4_exec_info *exec,
+- uint32_t gem_handles_packet_index,
+- enum vc4_bo_mode mode,
+- struct drm_gem_cma_object **obj)
++validate_bin_pos(struct vc4_exec_info *exec, void *untrusted, uint32_t pos)
+ {
+- return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index],
+- mode, obj);
++ /* Note that the untrusted pointer passed to these functions is
++ * incremented past the packet byte.
++ */
++ return (untrusted - 1 == exec->bin_u + pos);
+ }
+
+ static uint32_t
+@@ -202,13 +202,13 @@ vc4_check_tex_size(struct vc4_exec_info
+ }
+
+ static int
+-validate_flush_all(VALIDATE_ARGS)
++validate_flush(VALIDATE_ARGS)
+ {
+- if (exec->found_increment_semaphore_packet) {
+- DRM_ERROR("VC4_PACKET_FLUSH_ALL after "
+- "VC4_PACKET_INCREMENT_SEMAPHORE\n");
++ if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 1)) {
++ DRM_ERROR("Bin CL must end with VC4_PACKET_FLUSH\n");
+ return -EINVAL;
+ }
++ exec->found_flush = true;
+
+ return 0;
+ }
+@@ -233,17 +233,13 @@ validate_start_tile_binning(VALIDATE_ARG
+ static int
+ validate_increment_semaphore(VALIDATE_ARGS)
+ {
+- if (exec->found_increment_semaphore_packet) {
+- DRM_ERROR("Duplicate VC4_PACKET_INCREMENT_SEMAPHORE\n");
++ if (!validate_bin_pos(exec, untrusted, exec->args->bin_cl_size - 2)) {
++ DRM_ERROR("Bin CL must end with "
++ "VC4_PACKET_INCREMENT_SEMAPHORE\n");
+ return -EINVAL;
+ }
+ exec->found_increment_semaphore_packet = true;
+
+- /* Once we've found the semaphore increment, there should be one FLUSH
+- * then the end of the command list. The FLUSH actually triggers the
+- * increment, so we only need to make sure there
+- */
+-
+ return 0;
+ }
+
+@@ -257,11 +253,6 @@ validate_indexed_prim_list(VALIDATE_ARGS
+ uint32_t index_size = (*(uint8_t *)(untrusted + 0) >> 4) ? 2 : 1;
+ struct vc4_shader_state *shader_state;
+
+- if (exec->found_increment_semaphore_packet) {
+- DRM_ERROR("Drawing after VC4_PACKET_INCREMENT_SEMAPHORE\n");
+- return -EINVAL;
+- }
+-
+ /* Check overflow condition */
+ if (exec->shader_state_count == 0) {
+ DRM_ERROR("shader state must precede primitives\n");
+@@ -272,7 +263,8 @@ validate_indexed_prim_list(VALIDATE_ARGS
+ if (max_index > shader_state->max_index)
+ shader_state->max_index = max_index;
+
+- if (!vc4_use_handle(exec, 0, VC4_MODE_RENDER, &ib))
++ ib = vc4_use_handle(exec, 0);
++ if (!ib)
+ return -EINVAL;
+
+ if (offset > ib->base.size ||
+@@ -295,11 +287,6 @@ validate_gl_array_primitive(VALIDATE_ARG
+ uint32_t max_index;
+ struct vc4_shader_state *shader_state;
+
+- if (exec->found_increment_semaphore_packet) {
+- DRM_ERROR("Drawing after VC4_PACKET_INCREMENT_SEMAPHORE\n");
+- return -EINVAL;
+- }
+-
+ /* Check overflow condition */
+ if (exec->shader_state_count == 0) {
+ DRM_ERROR("shader state must precede primitives\n");
+@@ -329,7 +316,6 @@ validate_gl_shader_state(VALIDATE_ARGS)
+ return -EINVAL;
+ }
+
+- exec->shader_state[i].packet = VC4_PACKET_GL_SHADER_STATE;
+ exec->shader_state[i].addr = *(uint32_t *)untrusted;
+ exec->shader_state[i].max_index = 0;
+
+@@ -348,31 +334,6 @@ validate_gl_shader_state(VALIDATE_ARGS)
+ }
+
+ static int
+-validate_nv_shader_state(VALIDATE_ARGS)
+-{
+- uint32_t i = exec->shader_state_count++;
+-
+- if (i >= exec->shader_state_size) {
+- DRM_ERROR("More requests for shader states than declared\n");
+- return -EINVAL;
+- }
+-
+- exec->shader_state[i].packet = VC4_PACKET_NV_SHADER_STATE;
+- exec->shader_state[i].addr = *(uint32_t *)untrusted;
+-
+- if (exec->shader_state[i].addr & 15) {
+- DRM_ERROR("NV shader state address 0x%08x misaligned\n",
+- exec->shader_state[i].addr);
+- return -EINVAL;
+- }
+-
+- *(uint32_t *)validated = (exec->shader_state[i].addr +
+- exec->shader_rec_p);
+-
+- return 0;
+-}
+-
+-static int
+ validate_tile_binning_config(VALIDATE_ARGS)
+ {
+ struct drm_device *dev = exec->exec_bo->base.dev;
+@@ -473,8 +434,8 @@ static const struct cmd_info {
+ } cmd_info[] = {
+ VC4_DEFINE_PACKET(VC4_PACKET_HALT, NULL),
+ VC4_DEFINE_PACKET(VC4_PACKET_NOP, NULL),
+- VC4_DEFINE_PACKET(VC4_PACKET_FLUSH, NULL),
+- VC4_DEFINE_PACKET(VC4_PACKET_FLUSH_ALL, validate_flush_all),
++ VC4_DEFINE_PACKET(VC4_PACKET_FLUSH, validate_flush),
++ VC4_DEFINE_PACKET(VC4_PACKET_FLUSH_ALL, NULL),
+ VC4_DEFINE_PACKET(VC4_PACKET_START_TILE_BINNING,
+ validate_start_tile_binning),
+ VC4_DEFINE_PACKET(VC4_PACKET_INCREMENT_SEMAPHORE,
+@@ -488,7 +449,6 @@ static const struct cmd_info {
+ VC4_DEFINE_PACKET(VC4_PACKET_PRIMITIVE_LIST_FORMAT, NULL),
+
+ VC4_DEFINE_PACKET(VC4_PACKET_GL_SHADER_STATE, validate_gl_shader_state),
+- VC4_DEFINE_PACKET(VC4_PACKET_NV_SHADER_STATE, validate_nv_shader_state),
+
+ VC4_DEFINE_PACKET(VC4_PACKET_CONFIGURATION_BITS, NULL),
+ VC4_DEFINE_PACKET(VC4_PACKET_FLAT_SHADE_FLAGS, NULL),
+@@ -575,8 +535,16 @@ vc4_validate_bin_cl(struct drm_device *d
+ return -EINVAL;
+ }
+
+- if (!exec->found_increment_semaphore_packet) {
+- DRM_ERROR("Bin CL missing VC4_PACKET_INCREMENT_SEMAPHORE\n");
++ /* The bin CL must be ended with INCREMENT_SEMAPHORE and FLUSH. The
++ * semaphore is used to trigger the render CL to start up, and the
++ * FLUSH is what caps the bin lists with
++ * VC4_PACKET_RETURN_FROM_SUB_LIST (so they jump back to the main
++ * render CL when they get called to) and actually triggers the queued
++ * semaphore increment.
++ */
++ if (!exec->found_increment_semaphore_packet || !exec->found_flush) {
++ DRM_ERROR("Bin CL missing VC4_PACKET_INCREMENT_SEMAPHORE + "
++ "VC4_PACKET_FLUSH\n");
+ return -EINVAL;
+ }
+
+@@ -607,7 +575,8 @@ reloc_tex(struct vc4_exec_info *exec,
+ uint32_t cube_map_stride = 0;
+ enum vc4_texture_data_type type;
+
+- if (!vc4_use_bo(exec, texture_handle_index, VC4_MODE_RENDER, &tex))
++ tex = vc4_use_bo(exec, texture_handle_index);
++ if (!tex)
+ return false;
+
+ if (sample->is_direct) {
+@@ -755,51 +724,28 @@ reloc_tex(struct vc4_exec_info *exec,
+ }
+
+ static int
+-validate_shader_rec(struct drm_device *dev,
+- struct vc4_exec_info *exec,
+- struct vc4_shader_state *state)
++validate_gl_shader_rec(struct drm_device *dev,
++ struct vc4_exec_info *exec,
++ struct vc4_shader_state *state)
+ {
+ uint32_t *src_handles;
+ void *pkt_u, *pkt_v;
+- enum shader_rec_reloc_type {
+- RELOC_CODE,
+- RELOC_VBO,
+- };
+- struct shader_rec_reloc {
+- enum shader_rec_reloc_type type;
+- uint32_t offset;
+- };
+- static const struct shader_rec_reloc gl_relocs[] = {
+- { RELOC_CODE, 4 }, /* fs */
+- { RELOC_CODE, 16 }, /* vs */
+- { RELOC_CODE, 28 }, /* cs */
+- };
+- static const struct shader_rec_reloc nv_relocs[] = {
+- { RELOC_CODE, 4 }, /* fs */
+- { RELOC_VBO, 12 }
++ static const uint32_t shader_reloc_offsets[] = {
++ 4, /* fs */
++ 16, /* vs */
++ 28, /* cs */
+ };
+- const struct shader_rec_reloc *relocs;
+- struct drm_gem_cma_object *bo[ARRAY_SIZE(gl_relocs) + 8];
+- uint32_t nr_attributes = 0, nr_fixed_relocs, nr_relocs, packet_size;
++ uint32_t shader_reloc_count = ARRAY_SIZE(shader_reloc_offsets);
++ struct drm_gem_cma_object *bo[shader_reloc_count + 8];
++ uint32_t nr_attributes, nr_relocs, packet_size;
+ int i;
+- struct vc4_validated_shader_info *shader;
+
+- if (state->packet == VC4_PACKET_NV_SHADER_STATE) {
+- relocs = nv_relocs;
+- nr_fixed_relocs = ARRAY_SIZE(nv_relocs);
+-
+- packet_size = 16;
+- } else {
+- relocs = gl_relocs;
+- nr_fixed_relocs = ARRAY_SIZE(gl_relocs);
+-
+- nr_attributes = state->addr & 0x7;
+- if (nr_attributes == 0)
+- nr_attributes = 8;
+- packet_size = gl_shader_rec_size(state->addr);
+- }
+- nr_relocs = nr_fixed_relocs + nr_attributes;
++ nr_attributes = state->addr & 0x7;
++ if (nr_attributes == 0)
++ nr_attributes = 8;
++ packet_size = gl_shader_rec_size(state->addr);
+
++ nr_relocs = ARRAY_SIZE(shader_reloc_offsets) + nr_attributes;
+ if (nr_relocs * 4 > exec->shader_rec_size) {
+ DRM_ERROR("overflowed shader recs reading %d handles "
+ "from %d bytes left\n",
+@@ -829,21 +775,30 @@ validate_shader_rec(struct drm_device *d
+ exec->shader_rec_v += roundup(packet_size, 16);
+ exec->shader_rec_size -= packet_size;
+
+- for (i = 0; i < nr_relocs; i++) {
+- enum vc4_bo_mode mode;
++ if (!(*(uint16_t *)pkt_u & VC4_SHADER_FLAG_FS_SINGLE_THREAD)) {
++ DRM_ERROR("Multi-threaded fragment shaders not supported.\n");
++ return -EINVAL;
++ }
+
+- if (i < nr_fixed_relocs && relocs[i].type == RELOC_CODE)
+- mode = VC4_MODE_SHADER;
+- else
+- mode = VC4_MODE_RENDER;
++ for (i = 0; i < shader_reloc_count; i++) {
++ if (src_handles[i] > exec->bo_count) {
++ DRM_ERROR("Shader handle %d too big\n", src_handles[i]);
++ return -EINVAL;
++ }
+
+- if (!vc4_use_bo(exec, src_handles[i], mode, &bo[i]))
+- return false;
++ bo[i] = exec->bo[src_handles[i]];
++ if (!bo[i])
++ return -EINVAL;
++ }
++ for (i = shader_reloc_count; i < nr_relocs; i++) {
++ bo[i] = vc4_use_bo(exec, src_handles[i]);
++ if (!bo[i])
++ return -EINVAL;
+ }
+
+- for (i = 0; i < nr_fixed_relocs; i++) {
+- struct vc4_bo *vc4_bo;
+- uint32_t o = relocs[i].offset;
++ for (i = 0; i < shader_reloc_count; i++) {
++ struct vc4_validated_shader_info *validated_shader;
++ uint32_t o = shader_reloc_offsets[i];
+ uint32_t src_offset = *(uint32_t *)(pkt_u + o);
+ uint32_t *texture_handles_u;
+ void *uniform_data_u;
+@@ -851,57 +806,50 @@ validate_shader_rec(struct drm_device *d
+
+ *(uint32_t *)(pkt_v + o) = bo[i]->paddr + src_offset;
+
+- switch (relocs[i].type) {
+- case RELOC_CODE:
+- if (src_offset != 0) {
+- DRM_ERROR("Shaders must be at offset 0 "
+- "of the BO.\n");
+- goto fail;
+- }
++ if (src_offset != 0) {
++ DRM_ERROR("Shaders must be at offset 0 of "
++ "the BO.\n");
++ return -EINVAL;
++ }
+
+- vc4_bo = to_vc4_bo(&bo[i]->base);
+- shader = vc4_bo->validated_shader;
+- if (!shader)
+- goto fail;
++ validated_shader = to_vc4_bo(&bo[i]->base)->validated_shader;
++ if (!validated_shader)
++ return -EINVAL;
+
+- if (shader->uniforms_src_size > exec->uniforms_size) {
+- DRM_ERROR("Uniforms src buffer overflow\n");
+- goto fail;
+- }
++ if (validated_shader->uniforms_src_size >
++ exec->uniforms_size) {
++ DRM_ERROR("Uniforms src buffer overflow\n");
++ return -EINVAL;
++ }
+
+- texture_handles_u = exec->uniforms_u;
+- uniform_data_u = (texture_handles_u +
+- shader->num_texture_samples);
+-
+- memcpy(exec->uniforms_v, uniform_data_u,
+- shader->uniforms_size);
+-
+- for (tex = 0;
+- tex < shader->num_texture_samples;
+- tex++) {
+- if (!reloc_tex(exec,
+- uniform_data_u,
+- &shader->texture_samples[tex],
+- texture_handles_u[tex])) {
+- goto fail;
+- }
++ texture_handles_u = exec->uniforms_u;
++ uniform_data_u = (texture_handles_u +
++ validated_shader->num_texture_samples);
++
++ memcpy(exec->uniforms_v, uniform_data_u,
++ validated_shader->uniforms_size);
++
++ for (tex = 0;
++ tex < validated_shader->num_texture_samples;
++ tex++) {
++ if (!reloc_tex(exec,
++ uniform_data_u,
++ &validated_shader->texture_samples[tex],
++ texture_handles_u[tex])) {
++ return -EINVAL;
+ }
++ }
+
+- *(uint32_t *)(pkt_v + o + 4) = exec->uniforms_p;
+-
+- exec->uniforms_u += shader->uniforms_src_size;
+- exec->uniforms_v += shader->uniforms_size;
+- exec->uniforms_p += shader->uniforms_size;
+-
+- break;
++ *(uint32_t *)(pkt_v + o + 4) = exec->uniforms_p;
+
+- case RELOC_VBO:
+- break;
+- }
++ exec->uniforms_u += validated_shader->uniforms_src_size;
++ exec->uniforms_v += validated_shader->uniforms_size;
++ exec->uniforms_p += validated_shader->uniforms_size;
+ }
+
+ for (i = 0; i < nr_attributes; i++) {
+- struct drm_gem_cma_object *vbo = bo[nr_fixed_relocs + i];
++ struct drm_gem_cma_object *vbo =
++ bo[ARRAY_SIZE(shader_reloc_offsets) + i];
+ uint32_t o = 36 + i * 8;
+ uint32_t offset = *(uint32_t *)(pkt_u + o + 0);
+ uint32_t attr_size = *(uint8_t *)(pkt_u + o + 4) + 1;
+@@ -933,9 +881,6 @@ validate_shader_rec(struct drm_device *d
+ }
+
+ return 0;
+-
+-fail:
+- return -EINVAL;
+ }
+
+ int
+@@ -946,7 +891,7 @@ vc4_validate_shader_recs(struct drm_devi
+ int ret = 0;
+
+ for (i = 0; i < exec->shader_state_count; i++) {
+- ret = validate_shader_rec(dev, exec, &exec->shader_state[i]);
++ ret = validate_gl_shader_rec(dev, exec, &exec->shader_state[i]);
+ if (ret)
+ return ret;
+ }