35 const int destination_stride = destination.
stride != 0 ? destination.
stride :
42 DCHECK_EQ(destination.
stride, 0) <<
"Custom stride for float destination is not implemented.";
79 #define DEFINE_PASS_ACCESSOR(pass, kernel_pass) \
80 void PassAccessorGPU::get_pass_##pass(const RenderBuffers *render_buffers, \
81 const BufferParams &buffer_params, \
82 const Destination &destination) const \
84 run_film_convert_kernels( \
85 DEVICE_KERNEL_FILM_CONVERT_##kernel_pass, render_buffers, buffer_params, destination); \
106 #undef DEFINE_PASS_ACCESSOR
virtual bool synchronize()=0
virtual bool enqueue(DeviceKernel kernel, const int work_size, DeviceKernelArguments const &args)=0
void run_film_convert_kernels(DeviceKernel kernel, const RenderBuffers *render_buffers, const BufferParams &buffer_params, const Destination &destination) const
PassAccessorGPU(DeviceQueue *queue, const PassAccessInfo &pass_access_info, float exposure, int num_samples)
device_ptr d_pixels_half_rgba
virtual void init_kernel_film_convert(KernelFilmConvert *kfilm_convert, const BufferParams &buffer_params, const Destination &destination) const
device_vector< float > buffer
device_ptr device_pointer
#define CCL_NAMESPACE_END
SyclQueue void void size_t num_bytes SyclQueue void const char void *memory_device_pointer KernelContext int kernel
ccl_gpu_kernel_postfix ccl_global const int ccl_global float const int work_size
ccl_gpu_kernel_postfix ccl_global float int int int int float bool int offset
ccl_gpu_kernel_postfix ccl_global float int int int int ccl_global const float int int int int int int int int int int int int num_samples
#define DEFINE_PASS_ACCESSOR(pass, kernel_pass)