mirror of
https://git.ffmpeg.org/ffmpeg.git
synced 2026-04-18 08:30:24 +00:00
Instead of implicitly testing for NaN values. This is mostly a straightforward translation, but we need some slight extra boilerplate to ensure the mask is correctly updated when e.g. commuting past a swizzle. Signed-off-by: Niklas Haas <git@haasn.dev>
1230 lines
47 KiB
C
1230 lines
47 KiB
C
/**
|
|
* Copyright (C) 2026 Lynne
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/mem.h"
|
|
#include "libavutil/refstruct.h"
|
|
|
|
#include "../ops_internal.h"
|
|
#include "../swscale_internal.h"
|
|
|
|
#include "ops.h"
|
|
|
|
#if HAVE_SPIRV_HEADERS_SPIRV_H || HAVE_SPIRV_UNIFIED1_SPIRV_H
|
|
#include "spvasm.h"
|
|
#endif
|
|
|
|
static void ff_sws_vk_uninit(AVRefStructOpaque opaque, void *obj)
|
|
{
|
|
FFVulkanOpsCtx *s = obj;
|
|
|
|
#if CONFIG_LIBSHADERC || CONFIG_LIBGLSLANG
|
|
if (s->spvc)
|
|
s->spvc->uninit(&s->spvc);
|
|
#endif
|
|
ff_vk_uninit(&s->vkctx);
|
|
}
|
|
|
|
int ff_sws_vk_init(SwsContext *sws, AVBufferRef *dev_ref)
|
|
{
|
|
int err;
|
|
SwsInternal *c = sws_internal(sws);
|
|
|
|
if (!c->hw_priv) {
|
|
c->hw_priv = av_refstruct_alloc_ext(sizeof(FFVulkanOpsCtx), 0, NULL,
|
|
ff_sws_vk_uninit);
|
|
if (!c->hw_priv)
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
|
|
FFVulkanOpsCtx *s = c->hw_priv;
|
|
if (s->vkctx.device_ref && s->vkctx.device_ref->data != dev_ref->data) {
|
|
/* Reinitialize with new context */
|
|
ff_vk_uninit(&s->vkctx);
|
|
} else if (s->vkctx.device_ref && s->vkctx.device_ref->data == dev_ref->data) {
|
|
return 0;
|
|
}
|
|
|
|
err = ff_vk_init(&s->vkctx, sws, dev_ref, NULL);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
s->qf = ff_vk_qf_find(&s->vkctx, VK_QUEUE_COMPUTE_BIT, 0);
|
|
if (!s->qf) {
|
|
av_log(sws, AV_LOG_ERROR, "Device has no compute queues\n");
|
|
return AVERROR(ENOTSUP);
|
|
}
|
|
|
|
#if CONFIG_LIBSHADERC || CONFIG_LIBGLSLANG
|
|
if (!s->spvc) {
|
|
s->spvc = ff_vk_spirv_init();
|
|
if (!s->spvc)
|
|
return AVERROR(ENOMEM);
|
|
}
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define MAX_DITHER_BUFS 4
|
|
|
|
typedef struct VulkanPriv {
|
|
FFVulkanOpsCtx *s;
|
|
FFVkExecPool e;
|
|
FFVulkanShader shd;
|
|
FFVkBuffer dither_buf[MAX_DITHER_BUFS];
|
|
int nb_dither_buf;
|
|
enum FFVkShaderRepFormat src_rep;
|
|
enum FFVkShaderRepFormat dst_rep;
|
|
} VulkanPriv;
|
|
|
|
static void process(const SwsFrame *dst, const SwsFrame *src, int y, int h,
|
|
const SwsPass *pass)
|
|
{
|
|
VulkanPriv *p = (VulkanPriv *) pass->priv;
|
|
FFVkExecContext *ec = ff_vk_exec_get(&p->s->vkctx, &p->e);
|
|
FFVulkanFunctions *vk = &p->s->vkctx.vkfn;
|
|
ff_vk_exec_start(&p->s->vkctx, ec);
|
|
|
|
AVFrame *src_f = (AVFrame *) src->avframe;
|
|
AVFrame *dst_f = (AVFrame *) dst->avframe;
|
|
ff_vk_exec_add_dep_frame(&p->s->vkctx, ec, src_f,
|
|
VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT,
|
|
VK_PIPELINE_STAGE_2_COMPUTE_SHADER_BIT);
|
|
ff_vk_exec_add_dep_frame(&p->s->vkctx, ec, dst_f,
|
|
VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT,
|
|
VK_PIPELINE_STAGE_2_COMPUTE_SHADER_BIT);
|
|
|
|
VkImageView src_views[AV_NUM_DATA_POINTERS];
|
|
VkImageView dst_views[AV_NUM_DATA_POINTERS];
|
|
ff_vk_create_imageviews(&p->s->vkctx, ec, src_views, src_f, p->src_rep);
|
|
ff_vk_create_imageviews(&p->s->vkctx, ec, dst_views, dst_f, p->dst_rep);
|
|
|
|
ff_vk_shader_update_img_array(&p->s->vkctx, ec, &p->shd, src_f, src_views,
|
|
0, 0, VK_IMAGE_LAYOUT_GENERAL, VK_NULL_HANDLE);
|
|
ff_vk_shader_update_img_array(&p->s->vkctx, ec, &p->shd, dst_f, dst_views,
|
|
0, 1, VK_IMAGE_LAYOUT_GENERAL, VK_NULL_HANDLE);
|
|
|
|
int nb_img_bar = 0;
|
|
VkImageMemoryBarrier2 img_bar[8];
|
|
ff_vk_frame_barrier(&p->s->vkctx, ec, src_f, img_bar, &nb_img_bar,
|
|
VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT,
|
|
VK_PIPELINE_STAGE_2_COMPUTE_SHADER_BIT,
|
|
VK_ACCESS_SHADER_READ_BIT,
|
|
VK_IMAGE_LAYOUT_GENERAL,
|
|
VK_QUEUE_FAMILY_IGNORED);
|
|
ff_vk_frame_barrier(&p->s->vkctx, ec, dst_f, img_bar, &nb_img_bar,
|
|
VK_PIPELINE_STAGE_2_ALL_COMMANDS_BIT,
|
|
VK_PIPELINE_STAGE_2_COMPUTE_SHADER_BIT,
|
|
VK_ACCESS_SHADER_WRITE_BIT,
|
|
VK_IMAGE_LAYOUT_GENERAL,
|
|
VK_QUEUE_FAMILY_IGNORED);
|
|
vk->CmdPipelineBarrier2(ec->buf, &(VkDependencyInfo) {
|
|
.sType = VK_STRUCTURE_TYPE_DEPENDENCY_INFO,
|
|
.pImageMemoryBarriers = img_bar,
|
|
.imageMemoryBarrierCount = nb_img_bar,
|
|
});
|
|
|
|
ff_vk_exec_bind_shader(&p->s->vkctx, ec, &p->shd);
|
|
|
|
vk->CmdDispatch(ec->buf,
|
|
FFALIGN(src_f->width, p->shd.lg_size[0])/p->shd.lg_size[0],
|
|
FFALIGN(src_f->height, p->shd.lg_size[1])/p->shd.lg_size[1],
|
|
1);
|
|
|
|
ff_vk_exec_submit(&p->s->vkctx, ec);
|
|
ff_vk_exec_wait(&p->s->vkctx, ec);
|
|
}
|
|
|
|
static void free_fn(void *priv)
|
|
{
|
|
VulkanPriv *p = priv;
|
|
ff_vk_exec_pool_free(&p->s->vkctx, &p->e);
|
|
ff_vk_shader_free(&p->s->vkctx, &p->shd);
|
|
for (int i = 0; i < p->nb_dither_buf; i++)
|
|
ff_vk_free_buf(&p->s->vkctx, &p->dither_buf[i]);
|
|
av_refstruct_unref(&p->s);
|
|
av_free(priv);
|
|
}
|
|
|
|
static int create_dither_bufs(FFVulkanOpsCtx *s, VulkanPriv *p, SwsOpList *ops)
|
|
{
|
|
int err;
|
|
p->nb_dither_buf = 0;
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
if (op->op != SWS_OP_DITHER)
|
|
continue;
|
|
av_assert0(p->nb_dither_buf + 1 <= MAX_DITHER_BUFS);
|
|
|
|
int size = (1 << op->dither.size_log2);
|
|
int idx = p->nb_dither_buf;
|
|
err = ff_vk_create_buf(&s->vkctx, &p->dither_buf[idx],
|
|
size*size*sizeof(float), NULL, NULL,
|
|
VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT,
|
|
VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT |
|
|
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT);
|
|
if (err < 0)
|
|
goto fail;
|
|
p->nb_dither_buf++;
|
|
|
|
float *dither_data;
|
|
err = ff_vk_map_buffer(&s->vkctx, &p->dither_buf[idx],
|
|
(uint8_t **)&dither_data, 0);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
for (int i = 0; i < size; i++) {
|
|
for (int j = 0; j < size; j++) {
|
|
const AVRational r = op->dither.matrix[i*size + j];
|
|
dither_data[i*size + j] = r.num/(float)r.den;
|
|
}
|
|
}
|
|
|
|
ff_vk_unmap_buffer(&s->vkctx, &p->dither_buf[idx], 1);
|
|
}
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
for (int i = 0; i < p->nb_dither_buf; i++)
|
|
ff_vk_free_buf(&p->s->vkctx, &p->dither_buf[i]);
|
|
return err;
|
|
}
|
|
|
|
#if HAVE_SPIRV_HEADERS_SPIRV_H || HAVE_SPIRV_UNIFIED1_SPIRV_H
|
|
struct DitherData {
|
|
int size;
|
|
int arr_1d_id;
|
|
int arr_2d_id;
|
|
int struct_id;
|
|
int struct_ptr_id;
|
|
int id;
|
|
int mask_id;
|
|
};
|
|
|
|
typedef struct SPIRVIDs {
|
|
int in_vars[3 + MAX_DITHER_BUFS];
|
|
|
|
int glfn;
|
|
int ep;
|
|
|
|
/* Types */
|
|
int void_type;
|
|
int b_type;
|
|
int u32_type;
|
|
int i32_type;
|
|
int f32_type;
|
|
int void_fn_type;
|
|
|
|
/* Define vector types */
|
|
int bvec2_type;
|
|
int u32vec2_type;
|
|
int i32vec2_type;
|
|
|
|
int u32vec3_type;
|
|
|
|
int u32vec4_type;
|
|
int f32vec4_type;
|
|
|
|
/* Constants */
|
|
int u32_p;
|
|
int f32_p;
|
|
int f32_0;
|
|
int u32_cid[5];
|
|
|
|
int const_ids[128];
|
|
int nb_const_ids;
|
|
|
|
int linear_deco_off[16];
|
|
int linear_deco_ops[16];
|
|
int nb_linear_ops;
|
|
|
|
struct DitherData dither[MAX_DITHER_BUFS];
|
|
int dither_ptr_elem_id;
|
|
int nb_dither_bufs;
|
|
|
|
int out_img_type;
|
|
int out_img_array_id;
|
|
|
|
int in_img_type;
|
|
int in_img_array_id;
|
|
|
|
/* Pointer types for images */
|
|
int u32vec3_tptr;
|
|
int out_img_tptr;
|
|
int out_img_sptr;
|
|
|
|
int in_img_tptr;
|
|
int in_img_sptr;
|
|
} SPIRVIDs;
|
|
|
|
/* Section 1: Function to define all shader header data, and decorations */
|
|
static void define_shader_header(FFVulkanShader *shd, SwsOpList *ops,
|
|
SPICtx *spi, SPIRVIDs *id)
|
|
{
|
|
spi_OpCapability(spi, SpvCapabilityShader); /* Shader type */
|
|
|
|
/* Declare required capabilities */
|
|
spi_OpCapability(spi, SpvCapabilityInt16);
|
|
spi_OpCapability(spi, SpvCapabilityInt8);
|
|
spi_OpCapability(spi, SpvCapabilityImageQuery);
|
|
spi_OpCapability(spi, SpvCapabilityStorageImageReadWithoutFormat);
|
|
spi_OpCapability(spi, SpvCapabilityStorageImageWriteWithoutFormat);
|
|
spi_OpCapability(spi, SpvCapabilityStorageBuffer8BitAccess);
|
|
/* Import the GLSL set of functions (used for min/max) */
|
|
id->glfn = spi_OpExtInstImport(spi, "GLSL.std.450");
|
|
|
|
/* Next section starts here */
|
|
spi_OpMemoryModel(spi, SpvAddressingModelLogical, SpvMemoryModelGLSL450);
|
|
|
|
/* Entrypoint */
|
|
id->ep = spi_OpEntryPoint(spi, SpvExecutionModelGLCompute, "main",
|
|
id->in_vars, 3 + id->nb_dither_bufs);
|
|
spi_OpExecutionMode(spi, id->ep, SpvExecutionModeLocalSize,
|
|
shd->lg_size, 3);
|
|
|
|
/* gl_GlobalInvocationID descriptor decorations */
|
|
spi_OpDecorate(spi, id->in_vars[0], SpvDecorationBuiltIn,
|
|
SpvBuiltInGlobalInvocationId);
|
|
|
|
/* Input image descriptor decorations */
|
|
spi_OpDecorate(spi, id->in_vars[1], SpvDecorationNonWritable);
|
|
spi_OpDecorate(spi, id->in_vars[1], SpvDecorationDescriptorSet, 0);
|
|
spi_OpDecorate(spi, id->in_vars[1], SpvDecorationBinding, 0);
|
|
|
|
/* Output image descriptor decorations */
|
|
spi_OpDecorate(spi, id->in_vars[2], SpvDecorationNonReadable);
|
|
spi_OpDecorate(spi, id->in_vars[2], SpvDecorationDescriptorSet, 0);
|
|
spi_OpDecorate(spi, id->in_vars[2], SpvDecorationBinding, 1);
|
|
|
|
for (int i = 0; i < id->nb_dither_bufs; i++) {
|
|
spi_OpDecorate(spi, id->dither[i].arr_1d_id, SpvDecorationArrayStride,
|
|
sizeof(float));
|
|
spi_OpDecorate(spi, id->dither[i].arr_2d_id, SpvDecorationArrayStride,
|
|
id->dither[i].size*sizeof(float));
|
|
spi_OpDecorate(spi, id->dither[i].struct_id, SpvDecorationBlock);
|
|
spi_OpMemberDecorate(spi, id->dither[i].struct_id, 0, SpvDecorationOffset, 0);
|
|
spi_OpDecorate(spi, id->dither[i].id, SpvDecorationDescriptorSet, 1);
|
|
spi_OpDecorate(spi, id->dither[i].id, SpvDecorationBinding, i);
|
|
}
|
|
|
|
/* All linear arithmetic ops must be decorated with NoContraction */
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
if (op->op != SWS_OP_LINEAR)
|
|
continue;
|
|
av_assert0((id->nb_linear_ops + 1) <= FF_ARRAY_ELEMS(id->linear_deco_off));
|
|
|
|
int nb_ops = 0;
|
|
for (int j = 0; j < 4; j++) {
|
|
nb_ops += !!op->lin.m[j][0].num;
|
|
nb_ops += op->lin.m[j][0].num && op->lin.m[j][4].num;
|
|
for (int i = 1; i < 4; i++) {
|
|
nb_ops += !!op->lin.m[j][i].num;
|
|
nb_ops += op->lin.m[j][i].num &&
|
|
(op->lin.m[j][0].num || op->lin.m[j][4].num);
|
|
}
|
|
}
|
|
|
|
id->linear_deco_off[id->nb_linear_ops] = spi_reserve(spi, nb_ops*4*3);
|
|
id->linear_deco_ops[id->nb_linear_ops] = nb_ops;
|
|
id->nb_linear_ops++;
|
|
}
|
|
}
|
|
|
|
/* Section 2: Define all types and constants */
|
|
static void define_shader_consts(SwsOpList *ops, SPICtx *spi, SPIRVIDs *id)
|
|
{
|
|
/* Define scalar types */
|
|
id->void_type = spi_OpTypeVoid(spi);
|
|
id->b_type = spi_OpTypeBool(spi);
|
|
int u32_type =
|
|
id->u32_type = spi_OpTypeInt(spi, 32, 0);
|
|
id->i32_type = spi_OpTypeInt(spi, 32, 1);
|
|
int f32_type =
|
|
id->f32_type = spi_OpTypeFloat(spi, 32);
|
|
id->void_fn_type = spi_OpTypeFunction(spi, id->void_type, NULL, 0);
|
|
|
|
/* Define vector types */
|
|
id->bvec2_type = spi_OpTypeVector(spi, id->b_type, 2);
|
|
id->u32vec2_type = spi_OpTypeVector(spi, u32_type, 2);
|
|
id->i32vec2_type = spi_OpTypeVector(spi, id->i32_type, 2);
|
|
|
|
id->u32vec3_type = spi_OpTypeVector(spi, u32_type, 3);
|
|
|
|
id->u32vec4_type = spi_OpTypeVector(spi, u32_type, 4);
|
|
id->f32vec4_type = spi_OpTypeVector(spi, f32_type, 4);
|
|
|
|
/* Constants */
|
|
id->u32_p = spi_OpUndef(spi, u32_type);
|
|
id->f32_p = spi_OpUndef(spi, f32_type);
|
|
id->f32_0 = spi_OpConstantFloat(spi, f32_type, 0);
|
|
for (int i = 0; i < 5; i++)
|
|
id->u32_cid[i] = spi_OpConstantUInt(spi, u32_type, i);
|
|
|
|
/* Operation constants */
|
|
id->nb_const_ids = 0;
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
/* Make sure there's always enough space for the maximum number of
|
|
* constants a single operation needs (currently linear, 20 consts). */
|
|
av_assert0((id->nb_const_ids + 20) <= FF_ARRAY_ELEMS(id->const_ids));
|
|
const SwsOp *op = &ops->ops[n];
|
|
switch (op->op) {
|
|
case SWS_OP_CONVERT:
|
|
if (ff_sws_pixel_type_is_int(op->convert.to) && op->convert.expand) {
|
|
AVRational m = ff_sws_pixel_expand(op->type, op->convert.to);
|
|
int tmp = spi_OpConstantUInt(spi, id->u32_type, m.num);
|
|
tmp = spi_OpConstantComposite(spi, id->u32vec4_type,
|
|
tmp, tmp, tmp, tmp);
|
|
id->const_ids[id->nb_const_ids++] = tmp;
|
|
}
|
|
break;
|
|
case SWS_OP_CLEAR:
|
|
for (int i = 0; i < 4; i++) {
|
|
if (!SWS_COMP_TEST(op->clear.mask, i))
|
|
continue;
|
|
AVRational cv = op->clear.value[i];
|
|
if (op->type == SWS_PIXEL_F32) {
|
|
float q = (float)cv.num/cv.den;
|
|
id->const_ids[id->nb_const_ids++] =
|
|
spi_OpConstantFloat(spi, f32_type, q);
|
|
} else {
|
|
av_assert0(cv.den == 1);
|
|
id->const_ids[id->nb_const_ids++] =
|
|
spi_OpConstantUInt(spi, u32_type, cv.num);
|
|
}
|
|
}
|
|
break;
|
|
case SWS_OP_LSHIFT:
|
|
case SWS_OP_RSHIFT: {
|
|
int tmp = spi_OpConstantUInt(spi, u32_type, op->shift.amount);
|
|
tmp = spi_OpConstantComposite(spi, id->u32vec4_type,
|
|
tmp, tmp, tmp, tmp);
|
|
id->const_ids[id->nb_const_ids++] = tmp;
|
|
break;
|
|
}
|
|
case SWS_OP_SCALE: {
|
|
int tmp;
|
|
if (op->type == SWS_PIXEL_F32) {
|
|
float q = op->scale.factor.num/(float)op->scale.factor.den;
|
|
tmp = spi_OpConstantFloat(spi, f32_type, q);
|
|
tmp = spi_OpConstantComposite(spi, id->f32vec4_type,
|
|
tmp, tmp, tmp, tmp);
|
|
} else {
|
|
av_assert0(op->scale.factor.den == 1);
|
|
tmp = spi_OpConstantUInt(spi, u32_type, op->scale.factor.num);
|
|
tmp = spi_OpConstantComposite(spi, id->u32vec4_type,
|
|
tmp, tmp, tmp, tmp);
|
|
}
|
|
id->const_ids[id->nb_const_ids++] = tmp;
|
|
break;
|
|
}
|
|
case SWS_OP_MIN:
|
|
case SWS_OP_MAX:
|
|
for (int i = 0; i < 4; i++) {
|
|
int tmp;
|
|
AVRational cl = op->clamp.limit[i];
|
|
if (!op->clamp.limit[i].den) {
|
|
continue;
|
|
} else if (op->type == SWS_PIXEL_F32) {
|
|
float q = (float)cl.num/((float)cl.den);
|
|
tmp = spi_OpConstantFloat(spi, f32_type, q);
|
|
} else {
|
|
av_assert0(cl.den == 1);
|
|
tmp = spi_OpConstantUInt(spi, u32_type, cl.num);
|
|
}
|
|
id->const_ids[id->nb_const_ids++] = tmp;
|
|
}
|
|
break;
|
|
case SWS_OP_DITHER:
|
|
for (int i = 0; i < 4; i++) {
|
|
if (op->dither.y_offset[i] < 0)
|
|
continue;
|
|
int tmp = spi_OpConstantUInt(spi, u32_type, op->dither.y_offset[i]);
|
|
id->const_ids[id->nb_const_ids++] = tmp;
|
|
}
|
|
break;
|
|
case SWS_OP_LINEAR: {
|
|
for (int i = 0; i < 4; i++) {
|
|
float val;
|
|
if (op->lin.m[i][0].num) {
|
|
val = op->lin.m[i][0].num/(float)op->lin.m[i][0].den;
|
|
id->const_ids[id->nb_const_ids++] =
|
|
spi_OpConstantFloat(spi, f32_type, val);
|
|
}
|
|
if (op->lin.m[i][4].num) {
|
|
val = op->lin.m[i][4].num/(float)op->lin.m[i][4].den;
|
|
id->const_ids[id->nb_const_ids++] =
|
|
spi_OpConstantFloat(spi, f32_type, val);
|
|
}
|
|
for (int j = 1; j < 4; j++) {
|
|
if (!op->lin.m[i][j].num)
|
|
continue;
|
|
val = op->lin.m[i][j].num/(float)op->lin.m[i][j].den;
|
|
id->const_ids[id->nb_const_ids++] =
|
|
spi_OpConstantFloat(spi, f32_type, val);
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Section 3: Define bindings */
|
|
static void define_shader_bindings(SwsOpList *ops, SPICtx *spi, SPIRVIDs *id,
|
|
int in_img_count, int out_img_count)
|
|
{
|
|
id->dither_ptr_elem_id = spi_OpTypePointer(spi, SpvStorageClassUniform,
|
|
id->f32_type);
|
|
|
|
struct DitherData *dither = id->dither;
|
|
for (int i = 0; i < id->nb_dither_bufs; i++) {
|
|
int size_id = spi_OpConstantUInt(spi, id->u32_type, dither[i].size);
|
|
dither[i].mask_id = spi_OpConstantUInt(spi, id->u32_type, dither[i].size - 1);
|
|
spi_OpTypeArray(spi, id->f32_type, dither[i].arr_1d_id, size_id);
|
|
spi_OpTypeArray(spi, dither[i].arr_1d_id, dither[i].arr_2d_id, size_id);
|
|
spi_OpTypeStruct(spi, dither[i].struct_id, dither[i].arr_2d_id);
|
|
dither[i].struct_ptr_id = spi_OpTypePointer(spi, SpvStorageClassUniform,
|
|
dither[i].struct_id);
|
|
dither[i].id = spi_OpVariable(spi, dither[i].id, dither[i].struct_ptr_id,
|
|
SpvStorageClassUniform, 0);
|
|
}
|
|
|
|
const SwsOp *op_w = ff_sws_op_list_output(ops);
|
|
const SwsOp *op_r = ff_sws_op_list_input(ops);
|
|
|
|
/* Define image types for descriptors */
|
|
id->out_img_type = spi_OpTypeImage(spi,
|
|
op_w->type == SWS_PIXEL_F32 ?
|
|
id->f32_type : id->u32_type,
|
|
2, 0, 0, 0, 2, SpvImageFormatUnknown);
|
|
id->out_img_array_id = spi_OpTypeArray(spi, id->out_img_type, spi_get_id(spi),
|
|
id->u32_cid[out_img_count]);
|
|
|
|
id->in_img_type = 0;
|
|
id->in_img_array_id = 0;
|
|
if (op_r) {
|
|
/* If the formats match, we have to reuse the types due to SPIR-V not
|
|
* allowing redundant type defines */
|
|
int match = ((op_w->type == SWS_PIXEL_F32) ==
|
|
(op_r->type == SWS_PIXEL_F32));
|
|
id->in_img_type = match ? id->out_img_type :
|
|
spi_OpTypeImage(spi,
|
|
op_r->type == SWS_PIXEL_F32 ?
|
|
id->f32_type : id->u32_type,
|
|
2, 0, 0, 0, 2, SpvImageFormatUnknown);
|
|
id->in_img_array_id = spi_OpTypeArray(spi, id->in_img_type, spi_get_id(spi),
|
|
id->u32_cid[in_img_count]);
|
|
}
|
|
|
|
/* Pointer types for images */
|
|
id->u32vec3_tptr = spi_OpTypePointer(spi, SpvStorageClassInput,
|
|
id->u32vec3_type);
|
|
id->out_img_tptr = spi_OpTypePointer(spi, SpvStorageClassUniformConstant,
|
|
id->out_img_array_id);
|
|
id->out_img_sptr = spi_OpTypePointer(spi, SpvStorageClassUniformConstant,
|
|
id->out_img_type);
|
|
|
|
id->in_img_tptr = 0;
|
|
id->in_img_sptr = 0;
|
|
if (op_r) {
|
|
id->in_img_tptr= spi_OpTypePointer(spi, SpvStorageClassUniformConstant,
|
|
id->in_img_array_id);
|
|
id->in_img_sptr= spi_OpTypePointer(spi, SpvStorageClassUniformConstant,
|
|
id->in_img_type);
|
|
}
|
|
|
|
/* Define inputs */
|
|
spi_OpVariable(spi, id->in_vars[0], id->u32vec3_tptr,
|
|
SpvStorageClassInput, 0);
|
|
if (op_r) {
|
|
spi_OpVariable(spi, id->in_vars[1], id->in_img_tptr,
|
|
SpvStorageClassUniformConstant, 0);
|
|
}
|
|
spi_OpVariable(spi, id->in_vars[2], id->out_img_tptr,
|
|
SpvStorageClassUniformConstant, 0);
|
|
}
|
|
|
|
static int add_ops_spirv(VulkanPriv *p, FFVulkanOpsCtx *s,
|
|
SwsOpList *ops, FFVulkanShader *shd)
|
|
{
|
|
uint8_t spvbuf[1024*16];
|
|
SPICtx spi_context = { 0 }, *spi = &spi_context;
|
|
SPIRVIDs spid_data = { 0 }, *id = &spid_data;
|
|
spi_init(spi, spvbuf, sizeof(spvbuf));
|
|
|
|
/* Interlaced formats are not currently supported */
|
|
if (ops->src.interlaced || ops->dst.interlaced)
|
|
return AVERROR(ENOTSUP);
|
|
|
|
ff_vk_shader_load(shd, VK_SHADER_STAGE_COMPUTE_BIT, NULL,
|
|
(uint32_t []) { 32, 32, 1 }, 0);
|
|
shd->precompiled = 0;
|
|
|
|
/* Image ops, to determine types */
|
|
const SwsOp *op_w = ff_sws_op_list_output(ops);
|
|
int out_img_count = op_w->rw.packed ? 1 : op_w->rw.elems;
|
|
p->dst_rep = op_w->type == SWS_PIXEL_F32 ? FF_VK_REP_FLOAT : FF_VK_REP_UINT;
|
|
|
|
const SwsOp *op_r = ff_sws_op_list_input(ops);
|
|
int in_img_count = op_r ? op_r->rw.packed ? 1 : op_r->rw.elems : 0;
|
|
if (op_r)
|
|
p->src_rep = op_r->type == SWS_PIXEL_F32 ? FF_VK_REP_FLOAT : FF_VK_REP_UINT;
|
|
|
|
FFVulkanDescriptorSetBinding desc_set[MAX_DITHER_BUFS] = {
|
|
{
|
|
.type = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
|
|
.stages = VK_SHADER_STAGE_COMPUTE_BIT,
|
|
.elems = 4,
|
|
},
|
|
{
|
|
.type = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
|
|
.stages = VK_SHADER_STAGE_COMPUTE_BIT,
|
|
.elems = 4,
|
|
},
|
|
};
|
|
ff_vk_shader_add_descriptor_set(&s->vkctx, shd, desc_set, 2, 0, 0);
|
|
|
|
/* Create dither buffers */
|
|
int err = create_dither_bufs(s, p, ops);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
/* Entrypoint inputs: gl_GlobalInvocationID, input and output images, dither */
|
|
id->in_vars[0] = spi_get_id(spi);
|
|
id->in_vars[1] = spi_get_id(spi);
|
|
id->in_vars[2] = spi_get_id(spi);
|
|
|
|
/* Create dither buffer descriptor set */
|
|
id->nb_dither_bufs = 0;
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
if (op->op != SWS_OP_DITHER)
|
|
continue;
|
|
|
|
id->dither[id->nb_dither_bufs].size = 1 << op->dither.size_log2;
|
|
id->dither[id->nb_dither_bufs].arr_1d_id = spi_get_id(spi);
|
|
id->dither[id->nb_dither_bufs].arr_2d_id = spi_get_id(spi);
|
|
id->dither[id->nb_dither_bufs].struct_id = spi_get_id(spi);
|
|
id->dither[id->nb_dither_bufs].id = spi_get_id(spi);
|
|
id->in_vars[3 + id->nb_dither_bufs] = id->dither[id->nb_dither_bufs].id;
|
|
|
|
desc_set[id->nb_dither_bufs++] = (FFVulkanDescriptorSetBinding) {
|
|
.type = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER,
|
|
.stages = VK_SHADER_STAGE_COMPUTE_BIT,
|
|
};
|
|
}
|
|
if (id->nb_dither_bufs)
|
|
ff_vk_shader_add_descriptor_set(&s->vkctx, shd, desc_set,
|
|
id->nb_dither_bufs, 1, 0);
|
|
|
|
/* Define shader header sections */
|
|
define_shader_header(shd, ops, spi, id);
|
|
define_shader_consts(ops, spi, id);
|
|
define_shader_bindings(ops, spi, id, in_img_count, out_img_count);
|
|
|
|
/* Main function starts here */
|
|
spi_OpFunction(spi, id->ep, id->void_type, 0, id->void_fn_type);
|
|
spi_OpLabel(spi, spi_get_id(spi));
|
|
|
|
/* Load input image handles */
|
|
int in_img[4] = { 0 };
|
|
for (int i = 0; i < in_img_count; i++) {
|
|
/* Deref array and then the pointer */
|
|
int img = spi_OpAccessChain(spi, id->in_img_sptr,
|
|
id->in_vars[1], id->u32_cid[i]);
|
|
in_img[i] = spi_OpLoad(spi, id->in_img_type, img,
|
|
SpvMemoryAccessMaskNone, 0);
|
|
}
|
|
|
|
/* Load output image handles */
|
|
int out_img[4];
|
|
for (int i = 0; i < out_img_count; i++) {
|
|
int img = spi_OpAccessChain(spi, id->out_img_sptr,
|
|
id->in_vars[2], id->u32_cid[i]);
|
|
out_img[i] = spi_OpLoad(spi, id->out_img_type, img,
|
|
SpvMemoryAccessMaskNone, 0);
|
|
}
|
|
|
|
/* Load gl_GlobalInvocationID */
|
|
int gid = spi_OpLoad(spi, id->u32vec3_type, id->in_vars[0],
|
|
SpvMemoryAccessMaskNone, 0);
|
|
|
|
/* ivec2(gl_GlobalInvocationID.xy) */
|
|
gid = spi_OpVectorShuffle(spi, id->u32vec2_type, gid, gid, 0, 1);
|
|
int gi2 = spi_OpBitcast(spi, id->i32vec2_type, gid);
|
|
|
|
/* imageSize(out_img[0]); */
|
|
int img1_s = spi_OpImageQuerySize(spi, id->i32vec2_type, out_img[0]);
|
|
int scmp = spi_OpSGreaterThanEqual(spi, id->bvec2_type, gi2, img1_s);
|
|
scmp = spi_OpAny(spi, id->b_type, scmp);
|
|
|
|
/* if (out of bounds) return */
|
|
int quit_label = spi_get_id(spi), merge_label = spi_get_id(spi);
|
|
spi_OpSelectionMerge(spi, merge_label, SpvSelectionControlMaskNone);
|
|
spi_OpBranchConditional(spi, scmp, quit_label, merge_label, 0);
|
|
|
|
spi_OpLabel(spi, quit_label);
|
|
spi_OpReturn(spi); /* Quit if out of bounds here */
|
|
spi_OpLabel(spi, merge_label);
|
|
|
|
/* Initialize main data state */
|
|
int data;
|
|
if (ops->ops[0].type == SWS_PIXEL_F32)
|
|
data = spi_OpCompositeConstruct(spi, id->f32vec4_type,
|
|
id->f32_p, id->f32_p,
|
|
id->f32_p, id->f32_p);
|
|
else
|
|
data = spi_OpCompositeConstruct(spi, id->u32vec4_type,
|
|
id->u32_p, id->u32_p,
|
|
id->u32_p, id->u32_p);
|
|
|
|
/* Keep track of which constant/buffer to use */
|
|
int nb_const_ids = 0;
|
|
int nb_dither_bufs = 0;
|
|
int nb_linear_ops = 0;
|
|
|
|
/* Operations */
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
SwsPixelType cur_type = op->op == SWS_OP_CONVERT ?
|
|
op->convert.to : op->type;
|
|
int type_v = cur_type == SWS_PIXEL_F32 ?
|
|
id->f32vec4_type : id->u32vec4_type;
|
|
int type_s = cur_type == SWS_PIXEL_F32 ?
|
|
id->f32_type : id->u32_type;
|
|
int uid = cur_type == SWS_PIXEL_F32 ?
|
|
id->f32_p : id->u32_p;
|
|
|
|
switch (op->op) {
|
|
case SWS_OP_READ:
|
|
if (op->rw.frac || op->rw.filter) {
|
|
return AVERROR(ENOTSUP);
|
|
} else if (op->rw.packed) {
|
|
data = spi_OpImageRead(spi, type_v, in_img[ops->plane_src[0]],
|
|
gid, SpvImageOperandsMaskNone);
|
|
} else {
|
|
int tmp[4] = { uid, uid, uid, uid };
|
|
for (int i = 0; i < op->rw.elems; i++) {
|
|
tmp[i] = spi_OpImageRead(spi, type_v,
|
|
in_img[ops->plane_src[i]], gid,
|
|
SpvImageOperandsMaskNone);
|
|
tmp[i] = spi_OpCompositeExtract(spi, type_s, tmp[i], 0);
|
|
}
|
|
data = spi_OpCompositeConstruct(spi, type_v,
|
|
tmp[0], tmp[1], tmp[2], tmp[3]);
|
|
}
|
|
break;
|
|
case SWS_OP_WRITE:
|
|
if (op->rw.frac || op->rw.filter) {
|
|
return AVERROR(ENOTSUP);
|
|
} else if (op->rw.packed) {
|
|
spi_OpImageWrite(spi, out_img[ops->plane_dst[0]], gid, data,
|
|
SpvImageOperandsMaskNone);
|
|
} else {
|
|
for (int i = 0; i < op->rw.elems; i++) {
|
|
int tmp = spi_OpCompositeExtract(spi, type_s, data, i);
|
|
tmp = spi_OpCompositeConstruct(spi, type_v, tmp, tmp, tmp, tmp);
|
|
spi_OpImageWrite(spi, out_img[ops->plane_dst[i]], gid, tmp,
|
|
SpvImageOperandsMaskNone);
|
|
}
|
|
}
|
|
break;
|
|
case SWS_OP_CLEAR:
|
|
for (int i = 0; i < 4; i++) {
|
|
if (!op->clear.value[i].den)
|
|
continue;
|
|
data = spi_OpCompositeInsert(spi, type_v,
|
|
id->const_ids[nb_const_ids++],
|
|
data, i);
|
|
}
|
|
break;
|
|
case SWS_OP_SWIZZLE:
|
|
data = spi_OpVectorShuffle(spi, type_v, data, data,
|
|
op->swizzle.in[0],
|
|
op->swizzle.in[1],
|
|
op->swizzle.in[2],
|
|
op->swizzle.in[3]);
|
|
break;
|
|
case SWS_OP_CONVERT:
|
|
if (ff_sws_pixel_type_is_int(cur_type) && op->convert.expand)
|
|
data = spi_OpIMul(spi, type_v, data, id->const_ids[nb_const_ids++]);
|
|
else if (op->type == SWS_PIXEL_F32 && type_s == id->u32_type)
|
|
data = spi_OpConvertFToU(spi, type_v, data);
|
|
else if (op->type != SWS_PIXEL_F32 && type_s == id->f32_type)
|
|
data = spi_OpConvertUToF(spi, type_v, data);
|
|
break;
|
|
case SWS_OP_LSHIFT:
|
|
data = spi_OpShiftLeftLogical(spi, type_v, data,
|
|
id->const_ids[nb_const_ids++]);
|
|
break;
|
|
case SWS_OP_RSHIFT:
|
|
data = spi_OpShiftRightLogical(spi, type_v, data,
|
|
id->const_ids[nb_const_ids++]);
|
|
break;
|
|
case SWS_OP_SCALE:
|
|
if (op->type == SWS_PIXEL_F32)
|
|
data = spi_OpFMul(spi, type_v, data,
|
|
id->const_ids[nb_const_ids++]);
|
|
else
|
|
data = spi_OpIMul(spi, type_v, data,
|
|
id->const_ids[nb_const_ids++]);
|
|
break;
|
|
case SWS_OP_MIN:
|
|
case SWS_OP_MAX: {
|
|
int t = op->type == SWS_PIXEL_F32 ?
|
|
op->op == SWS_OP_MIN ? GLSLstd450FMin : GLSLstd450FMax :
|
|
op->op == SWS_OP_MIN ? GLSLstd450UMin : GLSLstd450UMax;
|
|
for (int i = 0; i < 4; i++) {
|
|
if (!op->clamp.limit[i].den)
|
|
continue;
|
|
int tmp = spi_OpCompositeExtract(spi, type_s, data, i);
|
|
tmp = spi_OpExtInst(spi, type_s, id->glfn, t,
|
|
tmp, id->const_ids[nb_const_ids++]);
|
|
data = spi_OpCompositeInsert(spi, type_v, tmp, data, i);
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_DITHER: {
|
|
int did = nb_dither_bufs++;
|
|
int x_id = spi_OpCompositeExtract(spi, id->u32_type, gid, 0);
|
|
int y_pos = spi_OpCompositeExtract(spi, id->u32_type, gid, 1);
|
|
x_id = spi_OpBitwiseAnd(spi, id->u32_type, x_id,
|
|
id->dither[did].mask_id);
|
|
for (int i = 0; i < 4; i++) {
|
|
if (op->dither.y_offset[i] < 0)
|
|
continue;
|
|
|
|
int y_id = spi_OpIAdd(spi, id->u32_type, y_pos,
|
|
id->const_ids[nb_const_ids++]);
|
|
y_id = spi_OpBitwiseAnd(spi, id->u32_type, y_id,
|
|
id->dither[did].mask_id);
|
|
|
|
int ptr = spi_OpAccessChain(spi, id->dither_ptr_elem_id,
|
|
id->dither[did].id, id->u32_cid[0],
|
|
y_id, x_id);
|
|
int val = spi_OpLoad(spi, id->f32_type, ptr,
|
|
SpvMemoryAccessMaskNone, 0);
|
|
|
|
int tmp = spi_OpCompositeExtract(spi, type_s, data, i);
|
|
tmp = spi_OpFAdd(spi, type_s, tmp, val);
|
|
data = spi_OpCompositeInsert(spi, type_v, tmp, data, i);
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_LINEAR: {
|
|
int tmp[4];
|
|
tmp[0] = spi_OpCompositeExtract(spi, type_s, data, 0);
|
|
tmp[1] = spi_OpCompositeExtract(spi, type_s, data, 1);
|
|
tmp[2] = spi_OpCompositeExtract(spi, type_s, data, 2);
|
|
tmp[3] = spi_OpCompositeExtract(spi, type_s, data, 3);
|
|
|
|
int off = spi_reserve(spi, 0); /* Current offset */
|
|
spi->off = id->linear_deco_off[nb_linear_ops];
|
|
for (int i = 0; i < id->linear_deco_ops[nb_linear_ops]; i++)
|
|
spi_OpDecorate(spi, spi->id + i, SpvDecorationNoContraction);
|
|
spi->off = off;
|
|
|
|
int res[4];
|
|
for (int j = 0; j < 4; j++) {
|
|
res[j] = op->type == SWS_PIXEL_F32 ? id->f32_0 : id->u32_cid[0];
|
|
if (op->lin.m[j][0].num)
|
|
res[j] = spi_OpFMul(spi, type_s, tmp[0],
|
|
id->const_ids[nb_const_ids++]);
|
|
|
|
if (op->lin.m[j][0].num && op->lin.m[j][4].num)
|
|
res[j] = spi_OpFAdd(spi, type_s,
|
|
id->const_ids[nb_const_ids++], res[j]);
|
|
else if (op->lin.m[j][4].num)
|
|
res[j] = id->const_ids[nb_const_ids++];
|
|
|
|
for (int i = 1; i < 4; i++) {
|
|
if (!op->lin.m[j][i].num)
|
|
continue;
|
|
|
|
int v = spi_OpFMul(spi, type_s, tmp[i],
|
|
id->const_ids[nb_const_ids++]);
|
|
if (op->lin.m[j][0].num || op->lin.m[j][4].num)
|
|
res[j] = spi_OpFAdd(spi, type_s, res[j], v);
|
|
else
|
|
res[j] = v;
|
|
}
|
|
}
|
|
data = spi_OpCompositeConstruct(spi, type_v,
|
|
res[0], res[1], res[2], res[3]);
|
|
nb_linear_ops++;
|
|
break;
|
|
}
|
|
default:
|
|
return AVERROR(ENOTSUP);
|
|
}
|
|
}
|
|
|
|
/* Return and finalize */
|
|
spi_OpReturn(spi);
|
|
spi_OpFunctionEnd(spi);
|
|
|
|
int len = spi_end(spi);
|
|
if (len < 0)
|
|
return AVERROR_INVALIDDATA;
|
|
|
|
return ff_vk_shader_link(&s->vkctx, shd, spvbuf, len, "main");
|
|
}
|
|
#endif
|
|
|
|
#if CONFIG_LIBSHADERC || CONFIG_LIBGLSLANG
|
|
static void add_desc_read_write(FFVulkanDescriptorSetBinding *out_desc,
|
|
enum FFVkShaderRepFormat *out_rep,
|
|
const SwsOp *op)
|
|
{
|
|
const char *img_type = op->type == SWS_PIXEL_F32 ? "rgba32f" :
|
|
op->type == SWS_PIXEL_U32 ? "rgba32ui" :
|
|
op->type == SWS_PIXEL_U16 ? "rgba16ui" :
|
|
"rgba8ui";
|
|
|
|
*out_desc = (FFVulkanDescriptorSetBinding) {
|
|
.name = op->op == SWS_OP_WRITE ? "dst_img" : "src_img",
|
|
.type = VK_DESCRIPTOR_TYPE_STORAGE_IMAGE,
|
|
.mem_layout = img_type,
|
|
.mem_quali = op->op == SWS_OP_WRITE ? "writeonly" : "readonly",
|
|
.dimensions = 2,
|
|
.elems = 4,
|
|
.stages = VK_SHADER_STAGE_COMPUTE_BIT,
|
|
};
|
|
|
|
*out_rep = op->type == SWS_PIXEL_F32 ? FF_VK_REP_FLOAT : FF_VK_REP_UINT;
|
|
}
|
|
|
|
#define QSTR "(%i/%i%s)"
|
|
#define QTYPE(Q) (Q).num, (Q).den, cur_type == SWS_PIXEL_F32 ? ".0f" : ""
|
|
|
|
static int add_ops_glsl(VulkanPriv *p, FFVulkanOpsCtx *s,
|
|
SwsOpList *ops, FFVulkanShader *shd)
|
|
{
|
|
int err;
|
|
uint8_t *spv_data;
|
|
size_t spv_len;
|
|
void *spv_opaque = NULL;
|
|
|
|
/* Interlaced formats are not currently supported */
|
|
if (ops->src.interlaced || ops->dst.interlaced)
|
|
return AVERROR(ENOTSUP);
|
|
|
|
err = ff_vk_shader_init(&s->vkctx, shd, "sws_pass",
|
|
VK_SHADER_STAGE_COMPUTE_BIT,
|
|
NULL, 0, 32, 32, 1, 0);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
int nb_desc = 0;
|
|
FFVulkanDescriptorSetBinding buf_desc[8];
|
|
|
|
const SwsOp *read = ff_sws_op_list_input(ops);
|
|
const SwsOp *write = ff_sws_op_list_output(ops);
|
|
if (read)
|
|
add_desc_read_write(&buf_desc[nb_desc++], &p->src_rep, read);
|
|
add_desc_read_write(&buf_desc[nb_desc++], &p->dst_rep, write);
|
|
ff_vk_shader_add_descriptor_set(&s->vkctx, shd, buf_desc, nb_desc, 0, 0);
|
|
|
|
err = create_dither_bufs(s, p, ops);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
nb_desc = 0;
|
|
char dither_buf_name[MAX_DITHER_BUFS][64];
|
|
char dither_mat_name[MAX_DITHER_BUFS][64];
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
if (op->op != SWS_OP_DITHER)
|
|
continue;
|
|
int size = (1 << op->dither.size_log2);
|
|
av_assert0(size < 8192);
|
|
snprintf(dither_buf_name[nb_desc], 64, "dither_buf%i", n);
|
|
snprintf(dither_mat_name[nb_desc], 64, "float dither_mat%i[%i][%i];",
|
|
n, size, size);
|
|
buf_desc[nb_desc] = (FFVulkanDescriptorSetBinding) {
|
|
.name = dither_buf_name[nb_desc],
|
|
.type = VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER,
|
|
.stages = VK_SHADER_STAGE_COMPUTE_BIT,
|
|
.mem_layout = "scalar",
|
|
.buf_content = dither_mat_name[nb_desc],
|
|
};
|
|
nb_desc++;
|
|
}
|
|
if (nb_desc)
|
|
ff_vk_shader_add_descriptor_set(&s->vkctx, shd, buf_desc,
|
|
nb_desc, 1, 0);
|
|
|
|
GLSLC(0, void main() );
|
|
GLSLC(0, { );
|
|
GLSLC(1, ivec2 pos = ivec2(gl_GlobalInvocationID.xy); );
|
|
GLSLC(1, ivec2 size = imageSize(src_img[0]); );
|
|
GLSLC(1, if (any(greaterThanEqual(pos, size))) );
|
|
GLSLC(2, return; );
|
|
GLSLC(0, );
|
|
GLSLC(1, u8vec4 u8; );
|
|
GLSLC(1, u16vec4 u16; );
|
|
GLSLC(1, u32vec4 u32; );
|
|
GLSLC(1, precise f32vec4 f32; );
|
|
GLSLC(1, precise f32vec4 tmp; );
|
|
GLSLC(0, );
|
|
|
|
for (int n = 0; n < ops->num_ops; n++) {
|
|
const SwsOp *op = &ops->ops[n];
|
|
SwsPixelType cur_type = op->op == SWS_OP_CONVERT ? op->convert.to :
|
|
op->type;
|
|
const char *type_name = ff_sws_pixel_type_name(cur_type);
|
|
const char *type_v = cur_type == SWS_PIXEL_F32 ? "f32vec4" :
|
|
cur_type == SWS_PIXEL_U32 ? "u32vec4" :
|
|
cur_type == SWS_PIXEL_U16 ? "u16vec4" : "u8vec4";
|
|
const char *type_s = cur_type == SWS_PIXEL_F32 ? "float" :
|
|
cur_type == SWS_PIXEL_U32 ? "uint32_t" :
|
|
cur_type == SWS_PIXEL_U16 ? "uint16_t" : "uint8_t";
|
|
av_bprintf(&shd->src, " // %s\n", ff_sws_op_type_name(op->op));
|
|
|
|
switch (op->op) {
|
|
case SWS_OP_READ: {
|
|
if (op->rw.frac || op->rw.filter) {
|
|
return AVERROR(ENOTSUP);
|
|
} else if (op->rw.packed) {
|
|
GLSLF(1, %s = %s(imageLoad(src_img[%i], pos)); ,
|
|
type_name, type_v, ops->plane_src[0]);
|
|
} else {
|
|
for (int i = 0; i < op->rw.elems; i++)
|
|
GLSLF(1, %s.%c = %s(imageLoad(src_img[%i], pos)[0]); ,
|
|
type_name, "xyzw"[i], type_s, ops->plane_src[i]);
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_WRITE: {
|
|
if (op->rw.frac || op->rw.filter) {
|
|
return AVERROR(ENOTSUP);
|
|
} else if (op->rw.packed) {
|
|
GLSLF(1, imageStore(dst_img[%i], pos, %s(%s)); ,
|
|
ops->plane_dst[0], type_v, type_name);
|
|
} else {
|
|
for (int i = 0; i < op->rw.elems; i++)
|
|
GLSLF(1, imageStore(dst_img[%i], pos, %s(%s[%i])); ,
|
|
ops->plane_dst[i], type_v, type_name, i);
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_SWIZZLE: {
|
|
av_bprintf(&shd->src, " %s = %s.", type_name, type_name);
|
|
for (int i = 0; i < 4; i++)
|
|
av_bprintf(&shd->src, "%c", "xyzw"[op->swizzle.in[i]]);
|
|
av_bprintf(&shd->src, ";\n");
|
|
break;
|
|
}
|
|
case SWS_OP_CLEAR: {
|
|
for (int i = 0; i < 4; i++) {
|
|
if (!SWS_COMP_TEST(op->clear.mask, i))
|
|
continue;
|
|
av_bprintf(&shd->src, " %s.%c = %s"QSTR";\n", type_name,
|
|
"xyzw"[i], type_s, QTYPE(op->clear.value[i]));
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_SCALE:
|
|
av_bprintf(&shd->src, " %s = %s * "QSTR";\n",
|
|
type_name, type_name, QTYPE(op->scale.factor));
|
|
break;
|
|
case SWS_OP_MIN:
|
|
case SWS_OP_MAX:
|
|
for (int i = 0; i < 4; i++) {
|
|
if (!op->clamp.limit[i].den)
|
|
continue;
|
|
av_bprintf(&shd->src, " %s.%c = %s(%s.%c, "QSTR");\n",
|
|
type_name, "xyzw"[i],
|
|
op->op == SWS_OP_MIN ? "min" : "max",
|
|
type_name, "xyzw"[i], QTYPE(op->clamp.limit[i]));
|
|
}
|
|
break;
|
|
case SWS_OP_LSHIFT:
|
|
case SWS_OP_RSHIFT:
|
|
av_bprintf(&shd->src, " %s %s= %i;\n", type_name,
|
|
op->op == SWS_OP_LSHIFT ? "<<" : ">>", op->shift.amount);
|
|
break;
|
|
case SWS_OP_CONVERT:
|
|
if (ff_sws_pixel_type_is_int(cur_type) && op->convert.expand) {
|
|
const AVRational sc = ff_sws_pixel_expand(op->type, op->convert.to);
|
|
av_bprintf(&shd->src, " %s = %s((%s*%i)/%i);\n",
|
|
type_name, type_v, ff_sws_pixel_type_name(op->type),
|
|
sc.num, sc.den);
|
|
} else {
|
|
av_bprintf(&shd->src, " %s = %s(%s);\n",
|
|
type_name, type_v, ff_sws_pixel_type_name(op->type));
|
|
}
|
|
break;
|
|
case SWS_OP_DITHER: {
|
|
int size = (1 << op->dither.size_log2);
|
|
for (int i = 0; i < 4; i++) {
|
|
if (op->dither.y_offset[i] < 0)
|
|
continue;
|
|
av_bprintf(&shd->src, " %s.%c += dither_mat%i[(pos.y + %i) & %i]"
|
|
"[pos.x & %i];\n",
|
|
type_name, "xyzw"[i], n,
|
|
op->dither.y_offset[i], size - 1,
|
|
size - 1);
|
|
}
|
|
break;
|
|
}
|
|
case SWS_OP_LINEAR:
|
|
for (int i = 0; i < 4; i++) {
|
|
if (op->lin.m[i][4].num)
|
|
av_bprintf(&shd->src, " tmp.%c = "QSTR";\n", "xyzw"[i],
|
|
QTYPE(op->lin.m[i][4]));
|
|
else
|
|
av_bprintf(&shd->src, " tmp.%c = 0;\n", "xyzw"[i]);
|
|
for (int j = 0; j < 4; j++) {
|
|
if (!op->lin.m[i][j].num)
|
|
continue;
|
|
av_bprintf(&shd->src, " tmp.%c += f32.%c*"QSTR";\n",
|
|
"xyzw"[i], "xyzw"[j], QTYPE(op->lin.m[i][j]));
|
|
}
|
|
}
|
|
av_bprintf(&shd->src, " f32 = tmp;\n");
|
|
break;
|
|
default:
|
|
return AVERROR(ENOTSUP);
|
|
}
|
|
}
|
|
|
|
GLSLC(0, } );
|
|
|
|
err = s->spvc->compile_shader(&s->vkctx, s->spvc, shd,
|
|
&spv_data, &spv_len, "main",
|
|
&spv_opaque);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
err = ff_vk_shader_link(&s->vkctx, shd, spv_data, spv_len, "main");
|
|
|
|
if (spv_opaque)
|
|
s->spvc->free_shader(s->spvc, &spv_opaque);
|
|
|
|
if (err < 0)
|
|
return err;
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int compile(SwsContext *sws, SwsOpList *ops, SwsCompiledOp *out, int glsl)
|
|
{
|
|
int err;
|
|
SwsInternal *c = sws_internal(sws);
|
|
FFVulkanOpsCtx *s = c->hw_priv;
|
|
if (!s)
|
|
return AVERROR(ENOTSUP);
|
|
|
|
VulkanPriv *p = av_mallocz(sizeof(*p));
|
|
if (!p)
|
|
return AVERROR(ENOMEM);
|
|
p->s = av_refstruct_ref(c->hw_priv);
|
|
|
|
err = ff_vk_exec_pool_init(&s->vkctx, s->qf, &p->e, 1,
|
|
0, 0, 0, NULL);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
if (ops->src.format == AV_PIX_FMT_BGR0 ||
|
|
ops->src.format == AV_PIX_FMT_BGRA ||
|
|
ops->dst.format == AV_PIX_FMT_BGR0 ||
|
|
ops->dst.format == AV_PIX_FMT_BGRA) {
|
|
VkFormatProperties2 prop = {
|
|
.sType = VK_STRUCTURE_TYPE_FORMAT_PROPERTIES_2,
|
|
};
|
|
FFVulkanFunctions *vk = &s->vkctx.vkfn;
|
|
vk->GetPhysicalDeviceFormatProperties2(s->vkctx.hwctx->phys_dev,
|
|
VK_FORMAT_B8G8R8A8_UNORM,
|
|
&prop);
|
|
if (!(prop.formatProperties.optimalTilingFeatures &
|
|
VK_FORMAT_FEATURE_2_STORAGE_WRITE_WITHOUT_FORMAT_BIT)) {
|
|
err = AVERROR(ENOTSUP);
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
if (glsl) {
|
|
err = AVERROR(ENOTSUP);
|
|
#if CONFIG_LIBSHADERC || CONFIG_LIBGLSLANG
|
|
err = add_ops_glsl(p, s, ops, &p->shd);
|
|
#endif
|
|
} else {
|
|
err = AVERROR(ENOTSUP);
|
|
#if HAVE_SPIRV_HEADERS_SPIRV_H || HAVE_SPIRV_UNIFIED1_SPIRV_H
|
|
err = add_ops_spirv(p, s, ops, &p->shd);
|
|
#endif
|
|
}
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
err = ff_vk_shader_register_exec(&s->vkctx, &p->e, &p->shd);
|
|
if (err < 0)
|
|
goto fail;
|
|
|
|
for (int i = 0; i < p->nb_dither_buf; i++)
|
|
ff_vk_shader_update_desc_buffer(&s->vkctx, &p->e.contexts[0], &p->shd,
|
|
1, i, 0, &p->dither_buf[i],
|
|
0, VK_WHOLE_SIZE, VK_FORMAT_UNDEFINED);
|
|
|
|
*out = (SwsCompiledOp) {
|
|
.opaque = true,
|
|
.func_opaque = process,
|
|
.priv = p,
|
|
.free = free_fn,
|
|
};
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
free_fn(p);
|
|
return err;
|
|
}
|
|
|
|
#if HAVE_SPIRV_HEADERS_SPIRV_H || HAVE_SPIRV_UNIFIED1_SPIRV_H
|
|
static int compile_spirv(SwsContext *sws, SwsOpList *ops, SwsCompiledOp *out)
|
|
{
|
|
return compile(sws, ops, out, 0);
|
|
}
|
|
|
|
const SwsOpBackend backend_spirv = {
|
|
.name = "spirv",
|
|
.compile = compile_spirv,
|
|
.hw_format = AV_PIX_FMT_VULKAN,
|
|
};
|
|
#endif
|
|
|
|
#if CONFIG_LIBSHADERC || CONFIG_LIBGLSLANG
|
|
static int compile_glsl(SwsContext *sws, SwsOpList *ops, SwsCompiledOp *out)
|
|
{
|
|
return compile(sws, ops, out, 1);
|
|
}
|
|
|
|
const SwsOpBackend backend_glsl = {
|
|
.name = "glsl",
|
|
.compile = compile_glsl,
|
|
.hw_format = AV_PIX_FMT_VULKAN,
|
|
};
|
|
#endif
|