/************************************************************************** * * Copyright (C) 2015 Red Hat Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included * in all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #include #include #include "util/u_memory.h" #include "util/u_math.h" #include "util/u_format.h" #include "pipe/p_state.h" #include "testvirgl_encode.h" #include "virgl_protocol.h" #include "tgsi/tgsi_dump.h" #include "tgsi/tgsi_parse.h" static int virgl_encoder_write_cmd_dword(struct virgl_context *ctx, uint32_t dword) { int len = (dword >> 16); if ((ctx->cbuf->cdw + len + 1) > VIRGL_MAX_CMDBUF_DWORDS) ctx->flush(ctx); virgl_encoder_write_dword(ctx->cbuf, dword); return 0; } static void virgl_encoder_write_res(struct virgl_context *ctx, struct virgl_resource *res) { if (res) virgl_encoder_write_dword(ctx->cbuf, res->handle); else virgl_encoder_write_dword(ctx->cbuf, 0); } static void virgl_encoder_attach_res(struct virgl_context *ctx, struct virgl_resource *res) { // struct virgl_winsys *vws = virgl_screen(ctx->base.screen)->vws; // if (res && res->hw_res) // vws->emit_res(vws, ctx->cbuf, res->hw_res, FALSE); } int virgl_encode_bind_object(struct virgl_context *ctx, uint32_t handle, uint32_t object) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_BIND_OBJECT, object, 1)); virgl_encoder_write_dword(ctx->cbuf, handle); return 0; } int virgl_encode_delete_object(struct virgl_context *ctx, uint32_t handle, uint32_t object) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_DESTROY_OBJECT, object, 1)); virgl_encoder_write_dword(ctx->cbuf, handle); return 0; } int virgl_encode_blend_state(struct virgl_context *ctx, uint32_t handle, const struct pipe_blend_state *blend_state) { uint32_t tmp; int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_BLEND, VIRGL_OBJ_BLEND_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); tmp = VIRGL_OBJ_BLEND_S0_INDEPENDENT_BLEND_ENABLE(blend_state->independent_blend_enable) | VIRGL_OBJ_BLEND_S0_LOGICOP_ENABLE(blend_state->logicop_enable) | VIRGL_OBJ_BLEND_S0_DITHER(blend_state->dither) | VIRGL_OBJ_BLEND_S0_ALPHA_TO_COVERAGE(blend_state->alpha_to_coverage) | VIRGL_OBJ_BLEND_S0_ALPHA_TO_ONE(blend_state->alpha_to_one); virgl_encoder_write_dword(ctx->cbuf, tmp); tmp = VIRGL_OBJ_BLEND_S1_LOGICOP_FUNC(blend_state->logicop_func); virgl_encoder_write_dword(ctx->cbuf, tmp); for (i = 0; i < VIRGL_MAX_COLOR_BUFS; i++) { tmp = VIRGL_OBJ_BLEND_S2_RT_BLEND_ENABLE(blend_state->rt[i].blend_enable) | VIRGL_OBJ_BLEND_S2_RT_RGB_FUNC(blend_state->rt[i].rgb_func) | VIRGL_OBJ_BLEND_S2_RT_RGB_SRC_FACTOR(blend_state->rt[i].rgb_src_factor) | VIRGL_OBJ_BLEND_S2_RT_RGB_DST_FACTOR(blend_state->rt[i].rgb_dst_factor)| VIRGL_OBJ_BLEND_S2_RT_ALPHA_FUNC(blend_state->rt[i].alpha_func) | VIRGL_OBJ_BLEND_S2_RT_ALPHA_SRC_FACTOR(blend_state->rt[i].alpha_src_factor) | VIRGL_OBJ_BLEND_S2_RT_ALPHA_DST_FACTOR(blend_state->rt[i].alpha_dst_factor) | VIRGL_OBJ_BLEND_S2_RT_COLORMASK(blend_state->rt[i].colormask); virgl_encoder_write_dword(ctx->cbuf, tmp); } return 0; } int virgl_encode_dsa_state(struct virgl_context *ctx, uint32_t handle, const struct pipe_depth_stencil_alpha_state *dsa_state) { uint32_t tmp; int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_DSA, VIRGL_OBJ_DSA_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); tmp = VIRGL_OBJ_DSA_S0_DEPTH_ENABLE(dsa_state->depth.enabled) | VIRGL_OBJ_DSA_S0_DEPTH_WRITEMASK(dsa_state->depth.writemask) | VIRGL_OBJ_DSA_S0_DEPTH_FUNC(dsa_state->depth.func) | VIRGL_OBJ_DSA_S0_ALPHA_ENABLED(dsa_state->alpha.enabled) | VIRGL_OBJ_DSA_S0_ALPHA_FUNC(dsa_state->alpha.func); virgl_encoder_write_dword(ctx->cbuf, tmp); for (i = 0; i < 2; i++) { tmp = VIRGL_OBJ_DSA_S1_STENCIL_ENABLED(dsa_state->stencil[i].enabled) | VIRGL_OBJ_DSA_S1_STENCIL_FUNC(dsa_state->stencil[i].func) | VIRGL_OBJ_DSA_S1_STENCIL_FAIL_OP(dsa_state->stencil[i].fail_op) | VIRGL_OBJ_DSA_S1_STENCIL_ZPASS_OP(dsa_state->stencil[i].zpass_op) | VIRGL_OBJ_DSA_S1_STENCIL_ZFAIL_OP(dsa_state->stencil[i].zfail_op) | VIRGL_OBJ_DSA_S1_STENCIL_VALUEMASK(dsa_state->stencil[i].valuemask) | VIRGL_OBJ_DSA_S1_STENCIL_WRITEMASK(dsa_state->stencil[i].writemask); virgl_encoder_write_dword(ctx->cbuf, tmp); } virgl_encoder_write_dword(ctx->cbuf, fui(dsa_state->alpha.ref_value)); return 0; } int virgl_encode_rasterizer_state(struct virgl_context *ctx, uint32_t handle, const struct pipe_rasterizer_state *state) { uint32_t tmp; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_RASTERIZER, VIRGL_OBJ_RS_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); tmp = VIRGL_OBJ_RS_S0_FLATSHADE(state->flatshade) | VIRGL_OBJ_RS_S0_DEPTH_CLIP(state->depth_clip) | VIRGL_OBJ_RS_S0_CLIP_HALFZ(state->clip_halfz) | VIRGL_OBJ_RS_S0_RASTERIZER_DISCARD(state->rasterizer_discard) | VIRGL_OBJ_RS_S0_FLATSHADE_FIRST(state->flatshade_first) | VIRGL_OBJ_RS_S0_LIGHT_TWOSIZE(state->light_twoside) | VIRGL_OBJ_RS_S0_SPRITE_COORD_MODE(state->sprite_coord_mode) | VIRGL_OBJ_RS_S0_POINT_QUAD_RASTERIZATION(state->point_quad_rasterization) | VIRGL_OBJ_RS_S0_CULL_FACE(state->cull_face) | VIRGL_OBJ_RS_S0_FILL_FRONT(state->fill_front) | VIRGL_OBJ_RS_S0_FILL_BACK(state->fill_back) | VIRGL_OBJ_RS_S0_SCISSOR(state->scissor) | VIRGL_OBJ_RS_S0_FRONT_CCW(state->front_ccw) | VIRGL_OBJ_RS_S0_CLAMP_VERTEX_COLOR(state->clamp_vertex_color) | VIRGL_OBJ_RS_S0_CLAMP_FRAGMENT_COLOR(state->clamp_fragment_color) | VIRGL_OBJ_RS_S0_OFFSET_LINE(state->offset_line) | VIRGL_OBJ_RS_S0_OFFSET_POINT(state->offset_point) | VIRGL_OBJ_RS_S0_OFFSET_TRI(state->offset_tri) | VIRGL_OBJ_RS_S0_POLY_SMOOTH(state->poly_smooth) | VIRGL_OBJ_RS_S0_POLY_STIPPLE_ENABLE(state->poly_stipple_enable) | VIRGL_OBJ_RS_S0_POINT_SMOOTH(state->point_smooth) | VIRGL_OBJ_RS_S0_POINT_SIZE_PER_VERTEX(state->point_size_per_vertex) | VIRGL_OBJ_RS_S0_MULTISAMPLE(state->multisample) | VIRGL_OBJ_RS_S0_LINE_SMOOTH(state->line_smooth) | VIRGL_OBJ_RS_S0_LINE_STIPPLE_ENABLE(state->line_stipple_enable) | VIRGL_OBJ_RS_S0_LINE_LAST_PIXEL(state->line_last_pixel) | VIRGL_OBJ_RS_S0_HALF_PIXEL_CENTER(state->half_pixel_center) | VIRGL_OBJ_RS_S0_BOTTOM_EDGE_RULE(state->bottom_edge_rule); virgl_encoder_write_dword(ctx->cbuf, tmp); /* S0 */ virgl_encoder_write_dword(ctx->cbuf, fui(state->point_size)); /* S1 */ virgl_encoder_write_dword(ctx->cbuf, state->sprite_coord_enable); /* S2 */ tmp = VIRGL_OBJ_RS_S3_LINE_STIPPLE_PATTERN(state->line_stipple_pattern) | VIRGL_OBJ_RS_S3_LINE_STIPPLE_FACTOR(state->line_stipple_factor) | VIRGL_OBJ_RS_S3_CLIP_PLANE_ENABLE(state->clip_plane_enable); virgl_encoder_write_dword(ctx->cbuf, tmp); /* S3 */ virgl_encoder_write_dword(ctx->cbuf, fui(state->line_width)); /* S4 */ virgl_encoder_write_dword(ctx->cbuf, fui(state->offset_units)); /* S5 */ virgl_encoder_write_dword(ctx->cbuf, fui(state->offset_scale)); /* S6 */ virgl_encoder_write_dword(ctx->cbuf, fui(state->offset_clamp)); /* S7 */ return 0; } int virgl_encode_shader_state(struct virgl_context *ctx, uint32_t handle, uint32_t type, const struct pipe_shader_state *shader, const char *shad_str) { char *str; uint32_t shader_len, len; int i; int ret; uint32_t tmp; int num_tokens; int str_total_size = 65536; if (!shad_str) { num_tokens = tgsi_num_tokens(shader->tokens); str = CALLOC(1, str_total_size); if (!str) return -1; ret = tgsi_dump_str(shader->tokens, TGSI_DUMP_FLOAT_AS_HEX, str, str_total_size); if (ret == -1) { fprintf(stderr, "Failed to translate shader in available space\n"); FREE(str); return -1; } } else { num_tokens = 0; str = (char *)shad_str; } shader_len = strlen(str) + 1; len = ((shader_len + 3) / 4) + VIRGL_OBJ_SHADER_HDR_SIZE(shader->stream_output.num_outputs); virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, type, len)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_dword(ctx->cbuf, num_tokens); virgl_encoder_write_dword(ctx->cbuf, shader->stream_output.num_outputs); if (shader->stream_output.num_outputs) { for (i = 0; i < 4; i++) virgl_encoder_write_dword(ctx->cbuf, shader->stream_output.stride[i]); for (i = 0; i < shader->stream_output.num_outputs; i++) { tmp = VIRGL_OBJ_SHADER_SO_OUTPUT_REGISTER_INDEX(shader->stream_output.output[i].register_index) | VIRGL_OBJ_SHADER_SO_OUTPUT_START_COMPONENT(shader->stream_output.output[i].start_component) | VIRGL_OBJ_SHADER_SO_OUTPUT_NUM_COMPONENTS(shader->stream_output.output[i].num_components) | VIRGL_OBJ_SHADER_SO_OUTPUT_BUFFER(shader->stream_output.output[i].output_buffer) | VIRGL_OBJ_SHADER_SO_OUTPUT_DST_OFFSET(shader->stream_output.output[i].dst_offset); virgl_encoder_write_dword(ctx->cbuf, tmp); } } virgl_encoder_write_block(ctx->cbuf, (uint8_t *)str, shader_len); if (str != shad_str) FREE(str); return 0; } int virgl_encode_clear(struct virgl_context *ctx, unsigned buffers, const union pipe_color_union *color, double depth, unsigned stencil) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CLEAR, 0, VIRGL_OBJ_CLEAR_SIZE)); virgl_encoder_write_dword(ctx->cbuf, buffers); for (i = 0; i < 4; i++) virgl_encoder_write_dword(ctx->cbuf, color->ui[i]); virgl_encoder_write_qword(ctx->cbuf, *(uint64_t *)&depth); virgl_encoder_write_dword(ctx->cbuf, stencil); return 0; } int virgl_encoder_set_framebuffer_state(struct virgl_context *ctx, const struct pipe_framebuffer_state *state) { struct virgl_surface *zsurf = (struct virgl_surface *)state->zsbuf; int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_FRAMEBUFFER_STATE, 0, VIRGL_SET_FRAMEBUFFER_STATE_SIZE(state->nr_cbufs))); virgl_encoder_write_dword(ctx->cbuf, state->nr_cbufs); virgl_encoder_write_dword(ctx->cbuf, zsurf ? zsurf->handle : 0); for (i = 0; i < state->nr_cbufs; i++) { struct virgl_surface *surf = (struct virgl_surface *)state->cbufs[i]; virgl_encoder_write_dword(ctx->cbuf, surf ? surf->handle : 0); if (surf) virgl_encoder_attach_res(ctx, (struct virgl_resource *)surf->base.texture); } if (zsurf) virgl_encoder_attach_res(ctx, (struct virgl_resource *)zsurf->base.texture); return 0; } int virgl_encoder_set_viewport_state(struct virgl_context *ctx, const struct pipe_viewport_state *state) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_VIEWPORT_STATE, 0, VIRGL_SET_VIEWPORT_STATE_SIZE(1))); virgl_encoder_write_dword(ctx->cbuf, 0); for (i = 0; i < 3; i++) virgl_encoder_write_dword(ctx->cbuf, fui(state->scale[i])); for (i = 0; i < 3; i++) virgl_encoder_write_dword(ctx->cbuf, fui(state->translate[i])); return 0; } int virgl_encoder_create_vertex_elements(struct virgl_context *ctx, uint32_t handle, unsigned num_elements, const struct pipe_vertex_element *element) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_VERTEX_ELEMENTS, VIRGL_OBJ_VERTEX_ELEMENTS_SIZE(num_elements))); virgl_encoder_write_dword(ctx->cbuf, handle); for (i = 0; i < num_elements; i++) { virgl_encoder_write_dword(ctx->cbuf, element[i].src_offset); virgl_encoder_write_dword(ctx->cbuf, element[i].instance_divisor); virgl_encoder_write_dword(ctx->cbuf, element[i].vertex_buffer_index); virgl_encoder_write_dword(ctx->cbuf, element[i].src_format); } return 0; } int virgl_encoder_set_vertex_buffers(struct virgl_context *ctx, unsigned num_buffers, const struct pipe_vertex_buffer *buffers) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_VERTEX_BUFFERS, 0, VIRGL_SET_VERTEX_BUFFERS_SIZE(num_buffers))); for (i = 0; i < num_buffers; i++) { struct virgl_resource *res = (struct virgl_resource *)buffers[i].buffer; virgl_encoder_write_dword(ctx->cbuf, buffers[i].stride); virgl_encoder_write_dword(ctx->cbuf, buffers[i].buffer_offset); virgl_encoder_write_res(ctx, res); } return 0; } int virgl_encoder_set_index_buffer(struct virgl_context *ctx, const struct pipe_index_buffer *ib) { int length = VIRGL_SET_INDEX_BUFFER_SIZE(ib); struct virgl_resource *res = NULL; if (ib) res = (struct virgl_resource *)ib->buffer; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_INDEX_BUFFER, 0, length)); virgl_encoder_write_res(ctx, res); if (ib) { virgl_encoder_write_dword(ctx->cbuf, ib->index_size); virgl_encoder_write_dword(ctx->cbuf, ib->offset); } return 0; } int virgl_encoder_draw_vbo(struct virgl_context *ctx, const struct pipe_draw_info *info) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_DRAW_VBO, 0, VIRGL_DRAW_VBO_SIZE)); virgl_encoder_write_dword(ctx->cbuf, info->start); virgl_encoder_write_dword(ctx->cbuf, info->count); virgl_encoder_write_dword(ctx->cbuf, info->mode); virgl_encoder_write_dword(ctx->cbuf, info->indexed); virgl_encoder_write_dword(ctx->cbuf, info->instance_count); virgl_encoder_write_dword(ctx->cbuf, info->index_bias); virgl_encoder_write_dword(ctx->cbuf, info->start_instance); virgl_encoder_write_dword(ctx->cbuf, info->primitive_restart); virgl_encoder_write_dword(ctx->cbuf, info->restart_index); virgl_encoder_write_dword(ctx->cbuf, info->min_index); virgl_encoder_write_dword(ctx->cbuf, info->max_index); virgl_encoder_write_dword(ctx->cbuf, 0); return 0; } int virgl_encoder_create_surface(struct virgl_context *ctx, uint32_t handle, struct virgl_resource *res, const struct pipe_surface *templat) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_SURFACE, VIRGL_OBJ_SURFACE_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_res(ctx, res); virgl_encoder_write_dword(ctx->cbuf, templat->format); if (templat->texture->target == PIPE_BUFFER) { virgl_encoder_write_dword(ctx->cbuf, templat->u.buf.first_element); virgl_encoder_write_dword(ctx->cbuf, templat->u.buf.last_element); } else { virgl_encoder_write_dword(ctx->cbuf, templat->u.tex.level); virgl_encoder_write_dword(ctx->cbuf, templat->u.tex.first_layer | (templat->u.tex.last_layer << 16)); } return 0; } int virgl_encoder_create_so_target(struct virgl_context *ctx, uint32_t handle, struct virgl_resource *res, unsigned buffer_offset, unsigned buffer_size) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_STREAMOUT_TARGET, VIRGL_OBJ_STREAMOUT_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_res(ctx, res); virgl_encoder_write_dword(ctx->cbuf, buffer_offset); virgl_encoder_write_dword(ctx->cbuf, buffer_size); return 0; } static void virgl_encoder_iw_emit_header_1d(struct virgl_context *ctx, struct virgl_resource *res, unsigned level, unsigned usage, const struct pipe_box *box, unsigned stride, unsigned layer_stride) { virgl_encoder_write_res(ctx, res); virgl_encoder_write_dword(ctx->cbuf, level); virgl_encoder_write_dword(ctx->cbuf, usage); virgl_encoder_write_dword(ctx->cbuf, stride); virgl_encoder_write_dword(ctx->cbuf, layer_stride); virgl_encoder_write_dword(ctx->cbuf, box->x); virgl_encoder_write_dword(ctx->cbuf, box->y); virgl_encoder_write_dword(ctx->cbuf, box->z); virgl_encoder_write_dword(ctx->cbuf, box->width); virgl_encoder_write_dword(ctx->cbuf, box->height); virgl_encoder_write_dword(ctx->cbuf, box->depth); } static void virgl_encoder_inline_send_box(struct virgl_context *ctx, struct virgl_resource *res, unsigned level, unsigned usage, const struct pipe_box *box, const void *data, unsigned stride, unsigned layer_stride, int length) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_RESOURCE_INLINE_WRITE, 0, ((length + 3) / 4) + 11)); virgl_encoder_iw_emit_header_1d(ctx, res, level, usage, box, stride, layer_stride); virgl_encoder_write_block(ctx->cbuf, data, length); } int virgl_encoder_inline_write(struct virgl_context *ctx, struct virgl_resource *res, unsigned level, unsigned usage, const struct pipe_box *box, const void *data, unsigned stride, unsigned layer_stride) { uint32_t length, thispass, left_bytes; struct pipe_box mybox = *box; unsigned elsize, size; unsigned layer_size; unsigned stride_internal = stride; unsigned layer_stride_internal = layer_stride; unsigned layer, row; elsize = util_format_get_blocksize(res->base.format); /* total size of data to transfer */ if (!stride) stride_internal = box->width * elsize; layer_size = box->height * stride_internal; if (layer_stride && layer_stride < layer_size) return -1; if (!layer_stride) layer_stride_internal = layer_size; size = layer_stride_internal * box->depth; length = 11 + (size + 3) / 4; /* can we send it all in one cmdbuf? */ if (length < VIRGL_MAX_CMDBUF_DWORDS) { /* is there space in this cmdbuf? if not flush and use another one */ if ((ctx->cbuf->cdw + length + 1) > VIRGL_MAX_CMDBUF_DWORDS) { ctx->flush(ctx); } /* send it all in one go. */ virgl_encoder_inline_send_box(ctx, res, level, usage, &mybox, data, stride, layer_stride, size); return 0; } /* break things down into chunks we can send */ /* send layers in separate chunks */ for (layer = 0; layer < box->depth; layer++) { const void *layer_data = data; mybox.z = layer; mybox.depth = 1; /* send one line in separate chunks */ for (row = 0; row < box->height; row++) { const void *row_data = layer_data; mybox.y = row; mybox.height = 1; mybox.x = 0; left_bytes = box->width * elsize; while (left_bytes) { if (ctx->cbuf->cdw + 12 > VIRGL_MAX_CMDBUF_DWORDS) ctx->flush(ctx); thispass = (VIRGL_MAX_CMDBUF_DWORDS - ctx->cbuf->cdw - 12) * 4; length = MIN2(thispass, left_bytes); mybox.width = length / elsize; virgl_encoder_inline_send_box(ctx, res, level, usage, &mybox, row_data, stride, layer_stride, length); left_bytes -= length; mybox.x += length / elsize; row_data += length; } layer_data += stride_internal; } data += layer_stride_internal; } return 0; } int virgl_encoder_flush_frontbuffer(struct virgl_context *ctx, struct virgl_resource *res) { // virgl_encoder_write_dword(ctx->cbuf, VIRGL_CMD0(VIRGL_CCMD_FLUSH_FRONTUBFFER, 0, 1)); // virgl_encoder_write_dword(ctx->cbuf, res_handle); return 0; } int virgl_encode_sampler_state(struct virgl_context *ctx, uint32_t handle, const struct pipe_sampler_state *state) { uint32_t tmp; int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_SAMPLER_STATE, VIRGL_OBJ_SAMPLER_STATE_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); tmp = VIRGL_OBJ_SAMPLE_STATE_S0_WRAP_S(state->wrap_s) | VIRGL_OBJ_SAMPLE_STATE_S0_WRAP_T(state->wrap_t) | VIRGL_OBJ_SAMPLE_STATE_S0_WRAP_R(state->wrap_r) | VIRGL_OBJ_SAMPLE_STATE_S0_MIN_IMG_FILTER(state->min_img_filter) | VIRGL_OBJ_SAMPLE_STATE_S0_MIN_MIP_FILTER(state->min_mip_filter) | VIRGL_OBJ_SAMPLE_STATE_S0_MAG_IMG_FILTER(state->mag_img_filter) | VIRGL_OBJ_SAMPLE_STATE_S0_COMPARE_MODE(state->compare_mode) | VIRGL_OBJ_SAMPLE_STATE_S0_COMPARE_FUNC(state->compare_func); virgl_encoder_write_dword(ctx->cbuf, tmp); virgl_encoder_write_dword(ctx->cbuf, fui(state->lod_bias)); virgl_encoder_write_dword(ctx->cbuf, fui(state->min_lod)); virgl_encoder_write_dword(ctx->cbuf, fui(state->max_lod)); for (i = 0; i < 4; i++) virgl_encoder_write_dword(ctx->cbuf, state->border_color.ui[i]); return 0; } int virgl_encode_sampler_view(struct virgl_context *ctx, uint32_t handle, struct virgl_resource *res, const struct pipe_sampler_view *state) { uint32_t tmp; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_SAMPLER_VIEW, VIRGL_OBJ_SAMPLER_VIEW_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_res(ctx, res); virgl_encoder_write_dword(ctx->cbuf, state->format); if (1) {//TODOres->u.b.target == PIPE_BUFFER) { virgl_encoder_write_dword(ctx->cbuf, state->u.buf.first_element); virgl_encoder_write_dword(ctx->cbuf, state->u.buf.last_element); } else { virgl_encoder_write_dword(ctx->cbuf, state->u.tex.first_layer | state->u.tex.last_layer << 16); virgl_encoder_write_dword(ctx->cbuf, state->u.tex.first_level | state->u.tex.last_level << 8); } tmp = VIRGL_OBJ_SAMPLER_VIEW_SWIZZLE_R(state->swizzle_r) | VIRGL_OBJ_SAMPLER_VIEW_SWIZZLE_G(state->swizzle_g) | VIRGL_OBJ_SAMPLER_VIEW_SWIZZLE_B(state->swizzle_b) | VIRGL_OBJ_SAMPLER_VIEW_SWIZZLE_A(state->swizzle_a); virgl_encoder_write_dword(ctx->cbuf, tmp); return 0; } int virgl_encode_set_sampler_views(struct virgl_context *ctx, uint32_t shader_type, uint32_t start_slot, uint32_t num_views, struct virgl_sampler_view **views) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_SAMPLER_VIEWS, 0, VIRGL_SET_SAMPLER_VIEWS_SIZE(num_views))); virgl_encoder_write_dword(ctx->cbuf, shader_type); virgl_encoder_write_dword(ctx->cbuf, start_slot); for (i = 0; i < num_views; i++) { uint32_t handle = views[i] ? views[i]->handle : 0; virgl_encoder_write_dword(ctx->cbuf, handle); // if (views[i]) // virgl_encoder_attach_res(ctx, (struct virgl_resource *)views[i]->base.texture); } return 0; } int virgl_encode_bind_sampler_states(struct virgl_context *ctx, uint32_t shader_type, uint32_t start_slot, uint32_t num_handles, uint32_t *handles) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_BIND_SAMPLER_STATES, 0, VIRGL_BIND_SAMPLER_STATES(num_handles))); virgl_encoder_write_dword(ctx->cbuf, shader_type); virgl_encoder_write_dword(ctx->cbuf, start_slot); for (i = 0; i < num_handles; i++) virgl_encoder_write_dword(ctx->cbuf, handles[i]); return 0; } int virgl_encoder_write_constant_buffer(struct virgl_context *ctx, uint32_t shader, uint32_t index, uint32_t size, const void *data) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_CONSTANT_BUFFER, 0, size + 2)); virgl_encoder_write_dword(ctx->cbuf, shader); virgl_encoder_write_dword(ctx->cbuf, index); if (data) virgl_encoder_write_block(ctx->cbuf, data, size * 4); return 0; } int virgl_encoder_set_uniform_buffer(struct virgl_context *ctx, uint32_t shader, uint32_t index, uint32_t offset, uint32_t length, struct virgl_resource *res) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_UNIFORM_BUFFER, 0, VIRGL_SET_UNIFORM_BUFFER_SIZE)); virgl_encoder_write_dword(ctx->cbuf, shader); virgl_encoder_write_dword(ctx->cbuf, index); virgl_encoder_write_dword(ctx->cbuf, offset); virgl_encoder_write_dword(ctx->cbuf, length); virgl_encoder_write_res(ctx, res); return 0; } int virgl_encoder_set_stencil_ref(struct virgl_context *ctx, const struct pipe_stencil_ref *ref) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_STENCIL_REF, 0, VIRGL_SET_STENCIL_REF_SIZE)); virgl_encoder_write_dword(ctx->cbuf, VIRGL_STENCIL_REF_VAL(ref->ref_value[0] , (ref->ref_value[1] << 8))); return 0; } int virgl_encoder_set_blend_color(struct virgl_context *ctx, const struct pipe_blend_color *color) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_BLEND_COLOR, 0, VIRGL_SET_BLEND_COLOR_SIZE)); for (i = 0; i < 4; i++) virgl_encoder_write_dword(ctx->cbuf, fui(color->color[i])); return 0; } int virgl_encoder_set_scissor_state(struct virgl_context *ctx, int start_slot, int num_scissors, const struct pipe_scissor_state *ss) { int s; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_SCISSOR_STATE, 0, VIRGL_SET_SCISSOR_STATE_SIZE(num_scissors))); virgl_encoder_write_dword(ctx->cbuf, start_slot); for (s = 0; s < num_scissors; s++) { virgl_encoder_write_dword(ctx->cbuf, (ss[s].minx | ss[s].miny << 16)); virgl_encoder_write_dword(ctx->cbuf, (ss[s].maxx | ss[s].maxy << 16)); } return 0; } void virgl_encoder_set_polygon_stipple(struct virgl_context *ctx, const struct pipe_poly_stipple *ps) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_POLYGON_STIPPLE, 0, VIRGL_POLYGON_STIPPLE_SIZE)); for (i = 0; i < VIRGL_POLYGON_STIPPLE_SIZE; i++) { virgl_encoder_write_dword(ctx->cbuf, ps->stipple[i]); } } void virgl_encoder_set_sample_mask(struct virgl_context *ctx, unsigned sample_mask) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_SAMPLE_MASK, 0, VIRGL_SET_SAMPLE_MASK_SIZE)); virgl_encoder_write_dword(ctx->cbuf, sample_mask); } void virgl_encoder_set_clip_state(struct virgl_context *ctx, const struct pipe_clip_state *clip) { int i, j; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_CLIP_STATE, 0, VIRGL_SET_CLIP_STATE_SIZE)); for (i = 0; i < VIRGL_MAX_CLIP_PLANES; i++) { for (j = 0; j < 4; j++) { virgl_encoder_write_dword(ctx->cbuf, fui(clip->ucp[i][j])); } } } int virgl_encode_resource_copy_region(struct virgl_context *ctx, struct virgl_resource *dst_res, unsigned dst_level, unsigned dstx, unsigned dsty, unsigned dstz, struct virgl_resource *src_res, unsigned src_level, const struct pipe_box *src_box) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_RESOURCE_COPY_REGION, 0, VIRGL_CMD_RESOURCE_COPY_REGION_SIZE)); virgl_encoder_write_res(ctx, dst_res); virgl_encoder_write_dword(ctx->cbuf, dst_level); virgl_encoder_write_dword(ctx->cbuf, dstx); virgl_encoder_write_dword(ctx->cbuf, dsty); virgl_encoder_write_dword(ctx->cbuf, dstz); virgl_encoder_write_res(ctx, src_res); virgl_encoder_write_dword(ctx->cbuf, src_level); virgl_encoder_write_dword(ctx->cbuf, src_box->x); virgl_encoder_write_dword(ctx->cbuf, src_box->y); virgl_encoder_write_dword(ctx->cbuf, src_box->z); virgl_encoder_write_dword(ctx->cbuf, src_box->width); virgl_encoder_write_dword(ctx->cbuf, src_box->height); virgl_encoder_write_dword(ctx->cbuf, src_box->depth); return 0; } int virgl_encode_blit(struct virgl_context *ctx, struct virgl_resource *dst_res, struct virgl_resource *src_res, const struct pipe_blit_info *blit) { uint32_t tmp; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_BLIT, 0, VIRGL_CMD_BLIT_SIZE)); tmp = VIRGL_CMD_BLIT_S0_MASK(blit->mask) | VIRGL_CMD_BLIT_S0_FILTER(blit->filter) | VIRGL_CMD_BLIT_S0_SCISSOR_ENABLE(blit->scissor_enable); virgl_encoder_write_dword(ctx->cbuf, tmp); virgl_encoder_write_dword(ctx->cbuf, (blit->scissor.minx | blit->scissor.miny << 16)); virgl_encoder_write_dword(ctx->cbuf, (blit->scissor.maxx | blit->scissor.maxy << 16)); virgl_encoder_write_res(ctx, dst_res); virgl_encoder_write_dword(ctx->cbuf, blit->dst.level); virgl_encoder_write_dword(ctx->cbuf, blit->dst.format); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.x); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.y); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.z); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.width); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.height); virgl_encoder_write_dword(ctx->cbuf, blit->dst.box.depth); virgl_encoder_write_res(ctx, src_res); virgl_encoder_write_dword(ctx->cbuf, blit->src.level); virgl_encoder_write_dword(ctx->cbuf, blit->src.format); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.x); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.y); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.z); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.width); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.height); virgl_encoder_write_dword(ctx->cbuf, blit->src.box.depth); return 0; } int virgl_encoder_create_query(struct virgl_context *ctx, uint32_t handle, uint query_type, struct virgl_resource *res, uint32_t offset) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_OBJECT, VIRGL_OBJECT_QUERY, VIRGL_OBJ_QUERY_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_dword(ctx->cbuf, query_type); virgl_encoder_write_dword(ctx->cbuf, offset); virgl_encoder_write_res(ctx, res); return 0; } int virgl_encoder_begin_query(struct virgl_context *ctx, uint32_t handle) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_BEGIN_QUERY, 0, 1)); virgl_encoder_write_dword(ctx->cbuf, handle); return 0; } int virgl_encoder_end_query(struct virgl_context *ctx, uint32_t handle) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_END_QUERY, 0, 1)); virgl_encoder_write_dword(ctx->cbuf, handle); return 0; } int virgl_encoder_get_query_result(struct virgl_context *ctx, uint32_t handle, boolean wait) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_GET_QUERY_RESULT, 0, 2)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_dword(ctx->cbuf, wait ? 1 : 0); return 0; } int virgl_encoder_render_condition(struct virgl_context *ctx, uint32_t handle, boolean condition, uint mode) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_RENDER_CONDITION, 0, VIRGL_RENDER_CONDITION_SIZE)); virgl_encoder_write_dword(ctx->cbuf, handle); virgl_encoder_write_dword(ctx->cbuf, condition); virgl_encoder_write_dword(ctx->cbuf, mode); return 0; } int virgl_encoder_set_so_targets(struct virgl_context *ctx, unsigned num_targets, struct pipe_stream_output_target **targets, unsigned append_bitmask) { int i; virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_STREAMOUT_TARGETS, 0, num_targets + 1)); virgl_encoder_write_dword(ctx->cbuf, append_bitmask); for (i = 0; i < num_targets; i++) { struct virgl_so_target *tg = (struct virgl_so_target *)targets[i]; virgl_encoder_write_dword(ctx->cbuf, tg->handle); } return 0; } int virgl_encoder_set_sub_ctx(struct virgl_context *ctx, uint32_t sub_ctx_id) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_SET_SUB_CTX, 0, 1)); virgl_encoder_write_dword(ctx->cbuf, sub_ctx_id); return 0; } int virgl_encoder_create_sub_ctx(struct virgl_context *ctx, uint32_t sub_ctx_id) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_CREATE_SUB_CTX, 0, 1)); virgl_encoder_write_dword(ctx->cbuf, sub_ctx_id); return 0; } int virgl_encoder_destroy_sub_ctx(struct virgl_context *ctx, uint32_t sub_ctx_id) { virgl_encoder_write_cmd_dword(ctx, VIRGL_CMD0(VIRGL_CCMD_DESTROY_SUB_CTX, 0, 1)); virgl_encoder_write_dword(ctx->cbuf, sub_ctx_id); return 0; }