#include #include #include #include #include "render/vulkan.h" #include "types/wlr_matrix.h" static const struct wlr_render_pass_impl render_pass_impl; static struct wlr_vk_render_pass *get_render_pass(struct wlr_render_pass *wlr_pass) { assert(wlr_pass->impl == &render_pass_impl); struct wlr_vk_render_pass *pass = wl_container_of(wlr_pass, pass, base); return pass; } static void bind_pipeline(struct wlr_vk_render_pass *pass, VkPipeline pipeline) { if (pipeline == pass->bound_pipeline) { return; } vkCmdBindPipeline(pass->command_buffer->vk, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline); pass->bound_pipeline = pipeline; } static void get_clip_region(struct wlr_vk_render_pass *pass, const pixman_region32_t *in, pixman_region32_t *out) { if (in != NULL) { pixman_region32_init(out); pixman_region32_copy(out, in); } else { struct wlr_buffer *buffer = pass->render_buffer->wlr_buffer; pixman_region32_init_rect(out, 0, 0, buffer->width, buffer->height); } } static void convert_pixman_box_to_vk_rect(const pixman_box32_t *box, VkRect2D *rect) { *rect = (VkRect2D){ .offset = { .x = box->x1, .y = box->y1 }, .extent = { .width = box->x2 - box->x1, .height = box->y2 - box->y1 }, }; } static float color_to_linear(float non_linear) { // See https://www.w3.org/Graphics/Color/srgb return (non_linear > 0.04045) ? pow((non_linear + 0.055) / 1.055, 2.4) : non_linear / 12.92; } static void mat3_to_mat4(const float mat3[9], float mat4[4][4]) { memset(mat4, 0, sizeof(float) * 16); mat4[0][0] = mat3[0]; mat4[0][1] = mat3[1]; mat4[0][3] = mat3[2]; mat4[1][0] = mat3[3]; mat4[1][1] = mat3[4]; mat4[1][3] = mat3[5]; mat4[2][2] = 1.f; mat4[3][3] = 1.f; } static bool render_pass_submit(struct wlr_render_pass *wlr_pass) { struct wlr_vk_render_pass *pass = get_render_pass(wlr_pass); struct wlr_vk_renderer *renderer = pass->renderer; struct wlr_vk_command_buffer *render_cb = pass->command_buffer; struct wlr_vk_render_buffer *render_buffer = pass->render_buffer; if (vulkan_record_stage_cb(renderer) == VK_NULL_HANDLE) { return false; } struct wlr_vk_command_buffer *stage_cb = renderer->stage.cb; assert(stage_cb != NULL); renderer->stage.cb = NULL; if (render_buffer->blend_image) { // Apply output shader to map blend image to actual output image vkCmdNextSubpass(render_cb->vk, VK_SUBPASS_CONTENTS_INLINE); VkPipeline pipe = render_buffer->render_setup->output_pipe; if (pipe != renderer->bound_pipe) { vkCmdBindPipeline(render_cb->vk, VK_PIPELINE_BIND_POINT_GRAPHICS, pipe); renderer->bound_pipe = pipe; } float final_matrix[9] = { renderer->render_width, 0, -1, 0, renderer->render_height, -1, 0, 0, 0, }; struct wlr_vk_vert_pcr_data vert_pcr_data = { .uv_off = { 0, 0 }, .uv_size = { 1, 1 }, }; mat3_to_mat4(final_matrix, vert_pcr_data.mat4); vkCmdPushConstants(render_cb->vk, renderer->output_pipe_layout, VK_SHADER_STAGE_VERTEX_BIT, 0, sizeof(vert_pcr_data), &vert_pcr_data); vkCmdBindDescriptorSets(render_cb->vk, VK_PIPELINE_BIND_POINT_GRAPHICS, renderer->output_pipe_layout, 0, 1, &render_buffer->blend_descriptor_set, 0, NULL); vkCmdDraw(render_cb->vk, 4, 1, 0, 0); } vkCmdEndRenderPass(render_cb->vk); // insert acquire and release barriers for dmabuf-images uint32_t barrier_count = wl_list_length(&renderer->foreign_textures) + 1; VkImageMemoryBarrier *acquire_barriers = calloc(barrier_count, sizeof(VkImageMemoryBarrier)); VkImageMemoryBarrier *release_barriers = calloc(barrier_count, sizeof(VkImageMemoryBarrier)); VkSemaphoreSubmitInfoKHR *render_wait = calloc(barrier_count * WLR_DMABUF_MAX_PLANES, sizeof(VkSemaphoreSubmitInfoKHR)); if (acquire_barriers == NULL || release_barriers == NULL || render_wait == NULL) { wlr_log_errno(WLR_ERROR, "Allocation failed"); free(acquire_barriers); free(release_barriers); free(render_wait); return false; } struct wlr_vk_texture *texture, *tmp_tex; size_t idx = 0; uint32_t render_wait_len = 0; wl_list_for_each_safe(texture, tmp_tex, &renderer->foreign_textures, foreign_link) { VkImageLayout src_layout = VK_IMAGE_LAYOUT_GENERAL; if (!texture->transitioned) { src_layout = VK_IMAGE_LAYOUT_UNDEFINED; texture->transitioned = true; } // acquire acquire_barriers[idx] = (VkImageMemoryBarrier){ .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, .srcQueueFamilyIndex = VK_QUEUE_FAMILY_FOREIGN_EXT, .dstQueueFamilyIndex = renderer->dev->queue_family, .image = texture->image, .oldLayout = src_layout, .newLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL, .srcAccessMask = 0, // ignored anyways .dstAccessMask = VK_ACCESS_SHADER_READ_BIT, .subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .subresourceRange.layerCount = 1, .subresourceRange.levelCount = 1, }; // release release_barriers[idx] = (VkImageMemoryBarrier){ .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, .srcQueueFamilyIndex = renderer->dev->queue_family, .dstQueueFamilyIndex = VK_QUEUE_FAMILY_FOREIGN_EXT, .image = texture->image, .oldLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL, .newLayout = VK_IMAGE_LAYOUT_GENERAL, .srcAccessMask = VK_ACCESS_SHADER_READ_BIT, .dstAccessMask = 0, // ignored anyways .subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .subresourceRange.layerCount = 1, .subresourceRange.levelCount = 1, }; ++idx; if (!vulkan_sync_foreign_texture(texture)) { wlr_log(WLR_ERROR, "Failed to wait for foreign texture DMA-BUF fence"); } else { for (size_t i = 0; i < WLR_DMABUF_MAX_PLANES; i++) { if (texture->foreign_semaphores[i] != VK_NULL_HANDLE) { assert(render_wait_len < barrier_count * WLR_DMABUF_MAX_PLANES); render_wait[render_wait_len++] = (VkSemaphoreSubmitInfoKHR){ .sType = VK_STRUCTURE_TYPE_SEMAPHORE_SUBMIT_INFO_KHR, .semaphore = texture->foreign_semaphores[i], .stageMask = VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT_KHR, }; } } } wl_list_remove(&texture->foreign_link); texture->owned = false; } // also add acquire/release barriers for the current render buffer VkImageLayout src_layout = VK_IMAGE_LAYOUT_GENERAL; if (!render_buffer->transitioned) { src_layout = VK_IMAGE_LAYOUT_PREINITIALIZED; render_buffer->transitioned = true; } if (render_buffer->blend_image) { // The render pass changes the blend image layout from // color attachment to read only, so on each frame, before // the render pass starts, we change it back VkImageLayout blend_src_layout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL; if (!render_buffer->blend_transitioned) { blend_src_layout = VK_IMAGE_LAYOUT_UNDEFINED; render_buffer->blend_transitioned = true; } VkImageMemoryBarrier blend_acq_barrier = { .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, .srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, .dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED, .image = render_buffer->blend_image, .oldLayout = blend_src_layout, .newLayout = VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL, .srcAccessMask = VK_ACCESS_SHADER_READ_BIT, .dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT, .subresourceRange = { .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .layerCount = 1, .levelCount = 1, }, }; vkCmdPipelineBarrier(stage_cb->vk, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT, VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, 0, 0, NULL, 0, NULL, 1, &blend_acq_barrier); } // acquire render buffer before rendering acquire_barriers[idx] = (VkImageMemoryBarrier){ .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, .srcQueueFamilyIndex = VK_QUEUE_FAMILY_FOREIGN_EXT, .dstQueueFamilyIndex = renderer->dev->queue_family, .image = render_buffer->image, .oldLayout = src_layout, .newLayout = VK_IMAGE_LAYOUT_GENERAL, .srcAccessMask = 0, // ignored anyways .dstAccessMask = VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT, .subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .subresourceRange.layerCount = 1, .subresourceRange.levelCount = 1, }; // release render buffer after rendering release_barriers[idx] = (VkImageMemoryBarrier){ .sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, .srcQueueFamilyIndex = renderer->dev->queue_family, .dstQueueFamilyIndex = VK_QUEUE_FAMILY_FOREIGN_EXT, .image = render_buffer->image, .oldLayout = VK_IMAGE_LAYOUT_GENERAL, .newLayout = VK_IMAGE_LAYOUT_GENERAL, .srcAccessMask = VK_ACCESS_COLOR_ATTACHMENT_READ_BIT | VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT, .dstAccessMask = 0, // ignored anyways .subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .subresourceRange.layerCount = 1, .subresourceRange.levelCount = 1, }; ++idx; vkCmdPipelineBarrier(stage_cb->vk, VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT, VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT | VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT, 0, 0, NULL, 0, NULL, barrier_count, acquire_barriers); vkCmdPipelineBarrier(render_cb->vk, VK_PIPELINE_STAGE_ALL_GRAPHICS_BIT, VK_PIPELINE_STAGE_BOTTOM_OF_PIPE_BIT, 0, 0, NULL, 0, NULL, barrier_count, release_barriers); free(acquire_barriers); free(release_barriers); // No semaphores needed here. // We don't need a semaphore from the stage/transfer submission // to the render submissions since they are on the same queue // and we have a renderpass dependency for that. uint64_t stage_timeline_point = vulkan_end_command_buffer(stage_cb, renderer); if (stage_timeline_point == 0) { return false; } VkCommandBufferSubmitInfoKHR stage_cb_info = { .sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_SUBMIT_INFO_KHR, .commandBuffer = stage_cb->vk, }; VkSemaphoreSubmitInfoKHR stage_signal = { .sType = VK_STRUCTURE_TYPE_SEMAPHORE_SUBMIT_INFO_KHR, .semaphore = renderer->timeline_semaphore, .value = stage_timeline_point, }; VkSubmitInfo2KHR stage_submit = { .sType = VK_STRUCTURE_TYPE_SUBMIT_INFO_2_KHR, .commandBufferInfoCount = 1, .pCommandBufferInfos = &stage_cb_info, .signalSemaphoreInfoCount = 1, .pSignalSemaphoreInfos = &stage_signal, }; VkSemaphoreSubmitInfoKHR stage_wait; if (renderer->stage.last_timeline_point > 0) { stage_wait = (VkSemaphoreSubmitInfoKHR){ .sType = VK_STRUCTURE_TYPE_SEMAPHORE_SUBMIT_INFO_KHR, .semaphore = renderer->timeline_semaphore, .value = renderer->stage.last_timeline_point, .stageMask = VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT_KHR, }; stage_submit.waitSemaphoreInfoCount = 1; stage_submit.pWaitSemaphoreInfos = &stage_wait; } renderer->stage.last_timeline_point = stage_timeline_point; uint64_t render_timeline_point = vulkan_end_command_buffer(render_cb, renderer); if (render_timeline_point == 0) { return false; } uint32_t render_signal_len = 1; VkSemaphoreSubmitInfoKHR render_signal[2] = {0}; render_signal[0] = (VkSemaphoreSubmitInfoKHR){ .sType = VK_STRUCTURE_TYPE_SEMAPHORE_SUBMIT_INFO_KHR, .semaphore = renderer->timeline_semaphore, .value = render_timeline_point, }; if (renderer->dev->implicit_sync_interop) { if (render_cb->binary_semaphore == VK_NULL_HANDLE) { VkExportSemaphoreCreateInfo export_info = { .sType = VK_STRUCTURE_TYPE_EXPORT_SEMAPHORE_CREATE_INFO, .handleTypes = VK_EXTERNAL_SEMAPHORE_HANDLE_TYPE_SYNC_FD_BIT, }; VkSemaphoreCreateInfo semaphore_info = { .sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO, .pNext = &export_info, }; VkResult res = vkCreateSemaphore(renderer->dev->dev, &semaphore_info, NULL, &render_cb->binary_semaphore); if (res != VK_SUCCESS) { wlr_vk_error("vkCreateSemaphore", res); return false; } } render_signal[render_signal_len++] = (VkSemaphoreSubmitInfoKHR){ .sType = VK_STRUCTURE_TYPE_SEMAPHORE_SUBMIT_INFO_KHR, .semaphore = render_cb->binary_semaphore, }; } VkCommandBufferSubmitInfoKHR render_cb_info = { .sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_SUBMIT_INFO_KHR, .commandBuffer = render_cb->vk, }; VkSubmitInfo2KHR render_submit = { .sType = VK_STRUCTURE_TYPE_SUBMIT_INFO_2_KHR, .waitSemaphoreInfoCount = render_wait_len, .pWaitSemaphoreInfos = render_wait, .commandBufferInfoCount = 1, .pCommandBufferInfos = &render_cb_info, .signalSemaphoreInfoCount = render_signal_len, .pSignalSemaphoreInfos = render_signal, }; VkSubmitInfo2KHR submit_info[] = { stage_submit, render_submit }; VkResult res = renderer->dev->api.vkQueueSubmit2KHR(renderer->dev->queue, 2, submit_info, VK_NULL_HANDLE); if (res == VK_ERROR_DEVICE_LOST) { wlr_log(WLR_ERROR, "vkQueueSubmit failed with VK_ERROR_DEVICE_LOST"); wl_signal_emit_mutable(&renderer->wlr_renderer.events.lost, NULL); return false; } else if (res != VK_SUCCESS) { wlr_vk_error("vkQueueSubmit", res); return false; } free(render_wait); struct wlr_vk_shared_buffer *stage_buf, *stage_buf_tmp; wl_list_for_each_safe(stage_buf, stage_buf_tmp, &renderer->stage.buffers, link) { if (stage_buf->allocs.size == 0) { continue; } wl_list_remove(&stage_buf->link); wl_list_insert(&stage_cb->stage_buffers, &stage_buf->link); } if (!vulkan_sync_render_buffer(renderer, render_buffer, render_cb)) { return false; } wlr_buffer_unlock(render_buffer->wlr_buffer); free(pass); return true; } static void render_pass_add_rect(struct wlr_render_pass *wlr_pass, const struct wlr_render_rect_options *options) { struct wlr_vk_render_pass *pass = get_render_pass(wlr_pass); struct wlr_vk_renderer *renderer = pass->renderer; VkCommandBuffer cb = pass->command_buffer->vk; // Input color values are given in sRGB space, shader expects // them in linear space. The shader does all computation in linear // space and expects in inputs in linear space since it outputs // colors in linear space as well (and vulkan then automatically // does the conversion for out sRGB render targets). float linear_color[] = { color_to_linear(options->color.r), color_to_linear(options->color.g), color_to_linear(options->color.b), options->color.a, // no conversion for alpha }; pixman_region32_t clip; get_clip_region(pass, options->clip, &clip); int clip_rects_len; const pixman_box32_t *clip_rects = pixman_region32_rectangles(&clip, &clip_rects_len); switch (options->blend_mode) { case WLR_RENDER_BLEND_MODE_PREMULTIPLIED:; float proj[9], matrix[9]; wlr_matrix_identity(proj); wlr_matrix_project_box(matrix, &options->box, WL_OUTPUT_TRANSFORM_NORMAL, 0, proj); wlr_matrix_multiply(matrix, pass->projection, matrix); struct wlr_vk_vert_pcr_data vert_pcr_data = { .uv_off = { 0, 0 }, .uv_size = { 1, 1 }, }; mat3_to_mat4(matrix, vert_pcr_data.mat4); bind_pipeline(pass, pass->render_buffer->render_setup->quad_pipe); vkCmdPushConstants(cb, renderer->default_pipeline_layout.vk, VK_SHADER_STAGE_VERTEX_BIT, 0, sizeof(vert_pcr_data), &vert_pcr_data); vkCmdPushConstants(cb, renderer->default_pipeline_layout.vk, VK_SHADER_STAGE_FRAGMENT_BIT, sizeof(vert_pcr_data), sizeof(float) * 4, linear_color); for (int i = 0; i < clip_rects_len; i++) { VkRect2D rect; convert_pixman_box_to_vk_rect(&clip_rects[i], &rect); vkCmdSetScissor(cb, 0, 1, &rect); vkCmdDraw(cb, 4, 1, 0, 0); } break; case WLR_RENDER_BLEND_MODE_NONE:; VkClearAttachment clear_att = { .aspectMask = VK_IMAGE_ASPECT_COLOR_BIT, .colorAttachment = 0, .clearValue.color.float32 = { linear_color[0], linear_color[1], linear_color[2], linear_color[3], }, }; VkClearRect clear_rect = { .rect = { .offset = { options->box.x, options->box.y }, .extent = { options->box.width, options->box.height }, }, .layerCount = 1, }; for (int i = 0; i < clip_rects_len; i++) { VkRect2D rect; convert_pixman_box_to_vk_rect(&clip_rects[i], &rect); vkCmdSetScissor(cb, 0, 1, &rect); vkCmdClearAttachments(cb, 1, &clear_att, 1, &clear_rect); } break; } pixman_region32_fini(&clip); } static void render_pass_add_texture(struct wlr_render_pass *wlr_pass, const struct wlr_render_texture_options *options) { struct wlr_vk_render_pass *pass = get_render_pass(wlr_pass); struct wlr_vk_renderer *renderer = pass->renderer; struct wlr_vk_render_buffer *render_buffer = pass->render_buffer; VkCommandBuffer cb = pass->command_buffer->vk; struct wlr_vk_texture *texture = vulkan_get_texture(options->texture); assert(texture->renderer == renderer); if (texture->dmabuf_imported && !texture->owned) { // Store this texture in the list of textures that need to be // acquired before rendering and released after rendering. // We don't do it here immediately since barriers inside // a renderpass are suboptimal (would require additional renderpass // dependency and potentially multiple barriers) and it's // better to issue one barrier for all used textures anyways. texture->owned = true; assert(texture->foreign_link.prev == NULL); assert(texture->foreign_link.next == NULL); wl_list_insert(&renderer->foreign_textures, &texture->foreign_link); } struct wlr_fbox src_box; wlr_render_texture_options_get_src_box(options, &src_box); struct wlr_box dst_box; wlr_render_texture_options_get_dst_box(options, &dst_box); float alpha = wlr_render_texture_options_get_alpha(options); pixman_region32_t clip; get_clip_region(pass, options->clip, &clip); float proj[9], matrix[9]; wlr_matrix_identity(proj); wlr_matrix_project_box(matrix, &dst_box, options->transform, 0, proj); wlr_matrix_multiply(matrix, pass->projection, matrix); struct wlr_vk_vert_pcr_data vert_pcr_data = { .uv_off = { src_box.x / options->texture->width, src_box.y / options->texture->height, }, .uv_size = { src_box.width / options->texture->width, src_box.height / options->texture->height, }, }; mat3_to_mat4(matrix, vert_pcr_data.mat4); VkPipelineLayout pipeline_layout = texture->pipeline_layout->vk; VkPipeline pipeline = vulkan_get_texture_pipeline(texture, render_buffer->render_setup); bind_pipeline(pass, pipeline); vkCmdBindDescriptorSets(cb, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline_layout, 0, 1, &texture->ds, 0, NULL); vkCmdPushConstants(cb, pipeline_layout, VK_SHADER_STAGE_VERTEX_BIT, 0, sizeof(vert_pcr_data), &vert_pcr_data); vkCmdPushConstants(cb, pipeline_layout, VK_SHADER_STAGE_FRAGMENT_BIT, sizeof(vert_pcr_data), sizeof(float), &alpha); int clip_rects_len; const pixman_box32_t *clip_rects = pixman_region32_rectangles(&clip, &clip_rects_len); for (int i = 0; i < clip_rects_len; i++) { VkRect2D rect; convert_pixman_box_to_vk_rect(&clip_rects[i], &rect); vkCmdSetScissor(cb, 0, 1, &rect); vkCmdDraw(cb, 4, 1, 0, 0); } texture->last_used_cb = pass->command_buffer; } static const struct wlr_render_pass_impl render_pass_impl = { .submit = render_pass_submit, .add_rect = render_pass_add_rect, .add_texture = render_pass_add_texture, }; struct wlr_vk_render_pass *vulkan_begin_render_pass(struct wlr_vk_renderer *renderer, struct wlr_vk_render_buffer *buffer) { struct wlr_vk_render_pass *pass = calloc(1, sizeof(*pass)); if (pass == NULL) { return NULL; } wlr_render_pass_init(&pass->base, &render_pass_impl); pass->renderer = renderer; struct wlr_vk_command_buffer *cb = vulkan_acquire_command_buffer(renderer); if (cb == NULL) { free(pass); return NULL; } VkCommandBufferBeginInfo begin_info = { .sType = VK_STRUCTURE_TYPE_COMMAND_BUFFER_BEGIN_INFO, }; VkResult res = vkBeginCommandBuffer(cb->vk, &begin_info); if (res != VK_SUCCESS) { wlr_vk_error("vkBeginCommandBuffer", res); free(pass); return NULL; } int width = buffer->wlr_buffer->width; int height = buffer->wlr_buffer->height; VkRect2D rect = { .extent = { width, height } }; VkRenderPassBeginInfo rp_info = { .sType = VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO, .renderArea = rect, .renderPass = buffer->render_setup->render_pass, .framebuffer = buffer->framebuffer, .clearValueCount = 0, }; vkCmdBeginRenderPass(cb->vk, &rp_info, VK_SUBPASS_CONTENTS_INLINE); vkCmdSetViewport(cb->vk, 0, 1, &(VkViewport){ .width = width, .height = height, .maxDepth = 1, }); // matrix_projection() assumes a GL coordinate system so we need // to pass WL_OUTPUT_TRANSFORM_FLIPPED_180 to adjust it for vulkan. matrix_projection(pass->projection, width, height, WL_OUTPUT_TRANSFORM_FLIPPED_180); wlr_buffer_lock(buffer->wlr_buffer); pass->render_buffer = buffer; pass->command_buffer = cb; return pass; }