/* SPDX-FileCopyrightText: 2023 Blender Authors * * SPDX-License-Identifier: GPL-2.0-or-later */ #include "BLI_array_utils.hh" #include "BLI_rand.hh" #include "BLI_task.hh" #include "BKE_attribute_math.hh" #include "BKE_geometry_set.hh" #include "BKE_instances.hh" namespace blender::bke { InstanceReference::InstanceReference(GeometrySet geometry_set) : type_(Type::GeometrySet), geometry_set_(std::make_unique(std::move(geometry_set))) { } void InstanceReference::ensure_owns_direct_data() { if (type_ != Type::GeometrySet) { return; } geometry_set_->ensure_owns_direct_data(); } bool InstanceReference::owns_direct_data() const { if (type_ != Type::GeometrySet) { /* The object and collection instances are not direct data. */ return true; } return geometry_set_->owns_direct_data(); } bool operator==(const InstanceReference &a, const InstanceReference &b) { if (a.geometry_set_ && b.geometry_set_) { return *a.geometry_set_ == *b.geometry_set_; } return a.type_ == b.type_ && a.data_ == b.data_; } Instances::Instances() { CustomData_reset(&attributes_); } Instances::Instances(Instances &&other) : references_(std::move(other.references_)), reference_handles_(std::move(other.reference_handles_)), transforms_(std::move(other.transforms_)), almost_unique_ids_(std::move(other.almost_unique_ids_)), attributes_(other.attributes_) { CustomData_reset(&other.attributes_); } Instances::Instances(const Instances &other) : references_(other.references_), reference_handles_(other.reference_handles_), transforms_(other.transforms_), almost_unique_ids_(other.almost_unique_ids_) { CustomData_copy(&other.attributes_, &attributes_, CD_MASK_ALL, other.instances_num()); } Instances::~Instances() { CustomData_free(&attributes_, this->instances_num()); } Instances &Instances::operator=(const Instances &other) { if (this == &other) { return *this; } std::destroy_at(this); new (this) Instances(other); return *this; } Instances &Instances::operator=(Instances &&other) { if (this == &other) { return *this; } std::destroy_at(this); new (this) Instances(std::move(other)); return *this; } void Instances::resize(int capacity) { const int old_size = this->instances_num(); reference_handles_.resize(capacity); transforms_.resize(capacity); CustomData_realloc(&attributes_, old_size, capacity, CD_SET_DEFAULT); } void Instances::add_instance(const int instance_handle, const float4x4 &transform) { BLI_assert(instance_handle >= 0); BLI_assert(instance_handle < references_.size()); const int old_size = this->instances_num(); reference_handles_.append(instance_handle); transforms_.append(transform); CustomData_realloc(&attributes_, old_size, transforms_.size()); } Span Instances::reference_handles() const { return reference_handles_; } MutableSpan Instances::reference_handles() { return reference_handles_; } MutableSpan Instances::transforms() { return transforms_; } Span Instances::transforms() const { return transforms_; } GeometrySet &Instances::geometry_set_from_reference(const int reference_index) { /* If this assert fails, it means #ensure_geometry_instances must be called first or that the * reference can't be converted to a geometry set. */ BLI_assert(references_[reference_index].type() == InstanceReference::Type::GeometrySet); return references_[reference_index].geometry_set(); } std::optional Instances::find_reference_handle(const InstanceReference &query) { for (const int i : references_.index_range()) { const InstanceReference &reference = references_[i]; if (reference == query) { return i; } } return std::nullopt; } int Instances::add_reference(const InstanceReference &reference) { if (std::optional handle = this->find_reference_handle(reference)) { return *handle; } return references_.append_and_get_index(reference); } Span Instances::references() const { return references_; } void Instances::remove(const IndexMask &mask, const AnonymousAttributePropagationInfo &propagation_info) { const std::optional masked_range = mask.to_range(); if (masked_range.has_value() && masked_range->start() == 0) { /* Deleting from the end of the array can be much faster since no data has to be shifted. */ this->resize(mask.size()); this->remove_unused_references(); return; } const int new_size = mask.size(); Instances new_instances; new_instances.references_ = std::move(references_); new_instances.reference_handles_.resize(new_size); new_instances.transforms_.resize(new_size); array_utils::gather( reference_handles_.as_span(), mask, new_instances.reference_handles_.as_mutable_span()); array_utils::gather(transforms_.as_span(), mask, new_instances.transforms_.as_mutable_span()); gather_attributes(this->attributes(), ATTR_DOMAIN_INSTANCE, propagation_info, {"position"}, mask, new_instances.attributes_for_write()); *this = std::move(new_instances); this->remove_unused_references(); } void Instances::remove_unused_references() { const int tot_instances = this->instances_num(); const int tot_references_before = references_.size(); if (tot_instances == 0) { /* If there are no instances, no reference is needed. */ references_.clear(); return; } if (tot_references_before == 1) { /* There is only one reference and at least one instance. So the only existing reference is * used. Nothing to do here. */ return; } Array usage_by_handle(tot_references_before, false); std::mutex mutex; /* Loop over all instances to see which references are used. */ threading::parallel_for(IndexRange(tot_instances), 1000, [&](IndexRange range) { /* Use local counter to avoid lock contention. */ Array local_usage_by_handle(tot_references_before, false); for (const int i : range) { const int handle = reference_handles_[i]; BLI_assert(handle >= 0 && handle < tot_references_before); local_usage_by_handle[handle] = true; } std::lock_guard lock{mutex}; for (const int i : IndexRange(tot_references_before)) { usage_by_handle[i] |= local_usage_by_handle[i]; } }); if (!usage_by_handle.as_span().contains(false)) { /* All references are used. */ return; } /* Create new references and a mapping for the handles. */ Vector handle_mapping; Vector new_references; int next_new_handle = 0; bool handles_have_to_be_updated = false; for (const int old_handle : IndexRange(tot_references_before)) { if (!usage_by_handle[old_handle]) { /* Add some dummy value. It won't be read again. */ handle_mapping.append(-1); } else { const InstanceReference &reference = references_[old_handle]; handle_mapping.append(next_new_handle); new_references.append(reference); if (old_handle != next_new_handle) { handles_have_to_be_updated = true; } next_new_handle++; } } references_ = new_references; if (!handles_have_to_be_updated) { /* All remaining handles are the same as before, so they don't have to be updated. This happens * when unused handles are only at the end. */ return; } /* Update handles of instances. */ threading::parallel_for(IndexRange(tot_instances), 1000, [&](IndexRange range) { for (const int i : range) { reference_handles_[i] = handle_mapping[reference_handles_[i]]; } }); } int Instances::instances_num() const { return transforms_.size(); } int Instances::references_num() const { return references_.size(); } bool Instances::owns_direct_data() const { for (const InstanceReference &reference : references_) { if (!reference.owns_direct_data()) { return false; } } return true; } void Instances::ensure_owns_direct_data() { for (const InstanceReference &const_reference : references_) { /* `const` cast is fine because we are not changing anything that would change the hash of the * reference. */ InstanceReference &reference = const_cast(const_reference); reference.ensure_owns_direct_data(); } } static Array generate_unique_instance_ids(Span original_ids) { Array unique_ids(original_ids.size()); Set used_unique_ids; used_unique_ids.reserve(original_ids.size()); Vector instances_with_id_collision; for (const int instance_index : original_ids.index_range()) { const int original_id = original_ids[instance_index]; if (used_unique_ids.add(original_id)) { /* The original id has not been used by another instance yet. */ unique_ids[instance_index] = original_id; } else { /* The original id of this instance collided with a previous instance, it needs to be looked * at again in a second pass. Don't generate a new random id here, because this might collide * with other existing ids. */ instances_with_id_collision.append(instance_index); } } Map generator_by_original_id; for (const int instance_index : instances_with_id_collision) { const int original_id = original_ids[instance_index]; RandomNumberGenerator &rng = generator_by_original_id.lookup_or_add_cb(original_id, [&]() { RandomNumberGenerator rng; rng.seed_random(original_id); return rng; }); const int max_iteration = 100; for (int iteration = 0;; iteration++) { /* Try generating random numbers until an unused one has been found. */ const int random_id = rng.get_int32(); if (used_unique_ids.add(random_id)) { /* This random id is not used by another instance. */ unique_ids[instance_index] = random_id; break; } if (iteration == max_iteration) { /* It seems to be very unlikely that we ever run into this case (assuming there are less * than 2^30 instances). However, if that happens, it's better to use an id that is not * unique than to be stuck in an infinite loop. */ unique_ids[instance_index] = original_id; break; } } } return unique_ids; } Span Instances::almost_unique_ids() const { std::lock_guard lock(almost_unique_ids_mutex_); bke::AttributeReader instance_ids_attribute = this->attributes().lookup("id"); if (instance_ids_attribute) { Span instance_ids = instance_ids_attribute.varray.get_internal_span(); if (almost_unique_ids_.size() != instance_ids.size()) { almost_unique_ids_ = generate_unique_instance_ids(instance_ids); } } else { almost_unique_ids_.reinitialize(this->instances_num()); for (const int i : almost_unique_ids_.index_range()) { almost_unique_ids_[i] = i; } } return almost_unique_ids_; } } // namespace blender::bke