This includes much improved GPU rendering performance, viewport interactivity, new shadow catcher, revamped sampling settings, subsurface scattering anisotropy, new GPU volume sampling, improved PMJ sampling pattern, and more. Some features have also been removed or changed, breaking backwards compatibility. Including the removal of the OpenCL backend, for which alternatives are under development. Release notes and code docs: https://wiki.blender.org/wiki/Reference/Release_Notes/3.0/Cycles https://wiki.blender.org/wiki/Source/Render/Cycles Credits: * Sergey Sharybin * Brecht Van Lommel * Patrick Mours (OptiX backend) * Christophe Hery (subsurface scattering anisotropy) * William Leeson (PMJ sampling pattern) * Alaska (various fixes and tweaks) * Thomas Dinges (various fixes) For the full commit history, see the cycles-x branch. This squashes together all the changes since intermediate changes would often fail building or tests. Ref T87839, T87837, T87836 Fixes T90734, T89353, T80267, T80267, T77185, T69800
69 lines
2.4 KiB
C
69 lines
2.4 KiB
C
/*
|
|
* Copyright 2014 Blender Foundation
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#ifndef __UTIL_ATOMIC_H__
|
|
#define __UTIL_ATOMIC_H__
|
|
|
|
#ifndef __KERNEL_GPU__
|
|
|
|
/* Using atomic ops header from Blender. */
|
|
# include "atomic_ops.h"
|
|
|
|
# define atomic_add_and_fetch_float(p, x) atomic_add_and_fetch_fl((p), (x))
|
|
# define atomic_compare_and_swap_float(p, old_val, new_val) \
|
|
atomic_cas_float((p), (old_val), (new_val))
|
|
|
|
# define atomic_fetch_and_inc_uint32(p) atomic_fetch_and_add_uint32((p), 1)
|
|
# define atomic_fetch_and_dec_uint32(p) atomic_fetch_and_add_uint32((p), -1)
|
|
|
|
# define CCL_LOCAL_MEM_FENCE 0
|
|
# define ccl_barrier(flags) ((void)0)
|
|
|
|
#else /* __KERNEL_GPU__ */
|
|
|
|
# ifdef __KERNEL_CUDA__
|
|
|
|
# define atomic_add_and_fetch_float(p, x) (atomicAdd((float *)(p), (float)(x)) + (float)(x))
|
|
|
|
# define atomic_fetch_and_add_uint32(p, x) atomicAdd((unsigned int *)(p), (unsigned int)(x))
|
|
# define atomic_fetch_and_sub_uint32(p, x) atomicSub((unsigned int *)(p), (unsigned int)(x))
|
|
# define atomic_fetch_and_inc_uint32(p) atomic_fetch_and_add_uint32((p), 1)
|
|
# define atomic_fetch_and_dec_uint32(p) atomic_fetch_and_sub_uint32((p), 1)
|
|
# define atomic_fetch_and_or_uint32(p, x) atomicOr((unsigned int *)(p), (unsigned int)(x))
|
|
|
|
ccl_device_inline float atomic_compare_and_swap_float(volatile float *dest,
|
|
const float old_val,
|
|
const float new_val)
|
|
{
|
|
union {
|
|
unsigned int int_value;
|
|
float float_value;
|
|
} new_value, prev_value, result;
|
|
prev_value.float_value = old_val;
|
|
new_value.float_value = new_val;
|
|
result.int_value = atomicCAS((unsigned int *)dest, prev_value.int_value, new_value.int_value);
|
|
return result.float_value;
|
|
}
|
|
|
|
# define CCL_LOCAL_MEM_FENCE
|
|
# define ccl_barrier(flags) __syncthreads()
|
|
|
|
# endif /* __KERNEL_CUDA__ */
|
|
|
|
#endif /* __KERNEL_GPU__ */
|
|
|
|
#endif /* __UTIL_ATOMIC_H__ */
|