This frequently showed up in profiling but shouldn't. This also updates the code to use atomics for more correctness and adds multi-threading for better performance.
48 lines
1.3 KiB
C++
48 lines
1.3 KiB
C++
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
|
|
#include <mutex>
|
|
|
|
#include "BLI_array.hh"
|
|
#include "BLI_index_range.hh"
|
|
#include "BLI_span.hh"
|
|
#include "BLI_task.hh"
|
|
#include "BLI_vector.hh"
|
|
|
|
namespace blender {
|
|
|
|
static RawVector<RawArray<int64_t, 0>> arrays;
|
|
static std::mutex current_array_mutex;
|
|
std::atomic<int64_t> IndexRange::s_current_array_size = 0;
|
|
std::atomic<int64_t *> IndexRange::s_current_array = nullptr;
|
|
|
|
Span<int64_t> IndexRange::as_span_internal() const
|
|
{
|
|
int64_t min_required_size = start_ + size_;
|
|
|
|
std::lock_guard<std::mutex> lock(current_array_mutex);
|
|
|
|
/* Double checked lock. */
|
|
if (min_required_size <= s_current_array_size) {
|
|
return Span<int64_t>(s_current_array + start_, size_);
|
|
}
|
|
|
|
/* Isolate, because a mutex is locked. */
|
|
threading::isolate_task([&]() {
|
|
int64_t new_size = std::max<int64_t>(1000, power_of_2_max_u(min_required_size));
|
|
RawArray<int64_t, 0> new_array(new_size);
|
|
threading::parallel_for(IndexRange(new_size), 4096, [&](const IndexRange range) {
|
|
for (const int64_t i : range) {
|
|
new_array[i] = i;
|
|
}
|
|
});
|
|
arrays.append(std::move(new_array));
|
|
|
|
s_current_array.store(arrays.last().data(), std::memory_order_release);
|
|
s_current_array_size.store(new_size, std::memory_order_release);
|
|
});
|
|
|
|
return Span<int64_t>(s_current_array + start_, size_);
|
|
}
|
|
|
|
} // namespace blender
|