13 uint inputs_stack_offsets,
14 uint result_stack_offset)
16 uint a_stack_offset, b_stack_offset, c_stack_offset;
31 uint inputs_stack_offsets,
32 uint outputs_stack_offsets,
35 uint value_stack_offset, vector_stack_offset;
36 uint a_stack_offset, b_stack_offset, param1_stack_offset;
38 inputs_stack_offsets, &a_stack_offset, &b_stack_offset, ¶m1_stack_offset);
@ NODE_VECTOR_MATH_FACEFORWARD
@ NODE_VECTOR_MATH_MULTIPLY_ADD
local_group_size(16, 16) .push_constant(Type b
const KernelGlobalsCPU *ccl_restrict KernelGlobals
#define ccl_device_noinline
#define CCL_NAMESPACE_END
ccl_device_noinline int svm_node_vector_math(KernelGlobals kg, ccl_private ShaderData *sd, ccl_private float *stack, uint type, uint inputs_stack_offsets, uint outputs_stack_offsets, int offset)
CCL_NAMESPACE_BEGIN ccl_device_noinline void svm_node_math(KernelGlobals kg, ccl_private ShaderData *sd, ccl_private float *stack, uint type, uint inputs_stack_offsets, uint result_stack_offset)
ccl_device_inline void stack_store_float3(ccl_private float *stack, uint a, float3 f)
CCL_NAMESPACE_BEGIN ccl_device_inline float3 stack_load_float3(ccl_private float *stack, uint a)
ccl_device_inline uint4 read_node(KernelGlobals kg, ccl_private int *offset)
ccl_device_forceinline void svm_unpack_node_uchar3(uint i, ccl_private uint *x, ccl_private uint *y, ccl_private uint *z)
ccl_device_forceinline void svm_unpack_node_uchar2(uint i, ccl_private uint *x, ccl_private uint *y)
ccl_device_inline void stack_store_float(ccl_private float *stack, uint a, float f)
ccl_device_inline float stack_load_float(ccl_private float *stack, uint a)
ccl_device_inline bool stack_valid(uint a)
CCL_NAMESPACE_BEGIN ccl_device void svm_vector_math(ccl_private float *value, ccl_private float3 *vector, NodeVectorMathType type, float3 a, float3 b, float3 c, float param1)
ccl_device float svm_math(NodeMathType type, float a, float b, float c)