diff --git a/.idea/editor.xml b/.idea/editor.xml
index 5fff85e..dfdaaeb 100644
--- a/.idea/editor.xml
+++ b/.idea/editor.xml
@@ -240,5 +240,247 @@
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
\ No newline at end of file
diff --git a/CMakeLists.txt b/CMakeLists.txt
index 33f88bd..8abd3a5 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -27,7 +27,7 @@ macro(compile_options target_name)
sanitizers(${target_name})
endmacro()
-project(blt-gp VERSION 0.3.19)
+project(blt-gp VERSION 0.3.20)
include(CTest)
diff --git a/include/blt/gp/tree.h b/include/blt/gp/tree.h
index 8f8b086..bc4fa1b 100644
--- a/include/blt/gp/tree.h
+++ b/include/blt/gp/tree.h
@@ -102,6 +102,17 @@ namespace blt::gp
stack_allocator values;
};
+ inline size_t accumulate_type_sizes(const detail::op_iter_t begin, const detail::op_iter_t end)
+ {
+ size_t total = 0;
+ for (auto it = begin; it != end; ++it)
+ {
+ if (it->is_value())
+ total += it->type_size();
+ }
+ return total;
+ }
+
template
class evaluation_ref
{
@@ -348,6 +359,22 @@ namespace blt::gp
[[nodiscard]] std::optional select_subtree_traverse(type_id type, u32 max_tries = 5, double terminal_chance = 0.1,
double depth_multiplier = 0.6) const;
+ /**
+ * Copies the subtree found at point into the provided out params
+ * @param point subtree point
+ * @param operators vector for storing subtree operators
+ * @param stack stack for storing subtree values
+ */
+ void copy_subtree(subtree_point_t point, std::vector& operators, stack_allocator& stack);
+
+ /**
+ * Swaps the subtrees between this tree and the other tree
+ * @param our_subtree
+ * @param other_tree
+ * @param other_subtree
+ */
+ void swap_subtrees(subtree_point_t our_subtree, tree_t& other_tree, subtree_point_t other_subtree);
+
/**
* User function for evaluating this tree using a context reference. This function should only be used if the tree is expecting the context value
* This function returns a copy of your value, if it is too large for the stack, or you otherwise need a reference, please use the corresponding
@@ -614,12 +641,12 @@ namespace blt::gp
return {ind, ++pos};
}
- tree_t& operator*()
+ tree_t& operator*() const
{
return ind[pos].tree;
}
- tree_t& operator->()
+ tree_t& operator->() const
{
return ind[pos].tree;
}
diff --git a/src/transformers.cpp b/src/transformers.cpp
index b97a276..2a114e1 100644
--- a/src/transformers.cpp
+++ b/src/transformers.cpp
@@ -35,19 +35,9 @@ namespace blt::gp
return new_tree;
}
- inline size_t accumulate_type_sizes(detail::op_iter_t begin, detail::op_iter_t end)
- {
- size_t total = 0;
- for (auto it = begin; it != end; ++it)
- {
- if (it->is_value())
- total += it->type_size();
- }
- return total;
- }
-
+ // TODO: consolidate the two copies of this. other is in tree.cpp
template
- u8* get_thread_pointer_for_size(size_t bytes)
+ static u8* get_thread_pointer_for_size(const size_t bytes)
{
thread_local expanding_buffer buffer;
if (bytes > buffer.size())
@@ -64,9 +54,6 @@ namespace blt::gp
if (p1.get_operations().size() < config.min_tree_size || p2.get_operations().size() < config.min_tree_size)
return false;
- auto& c1_ops = c1.get_operations();
- auto& c2_ops = c2.get_operations();
-
std::optional point;
if (config.traverse)
@@ -77,72 +64,13 @@ namespace blt::gp
if (!point)
return false;
- const auto selection = program.get_random().get_u32(0, 2);
-
- // Used to make copies of operators. Statically stored for memory caching purposes.
- // Thread local as this function cannot have external modifications / will be called from multiple threads.
- thread_local tracked_vector c1_operators;
- thread_local tracked_vector c2_operators;
- c1_operators.clear();
- c2_operators.clear();
-
// TODO: more crossover!
- switch (selection)
+ switch (program.get_random().get_u32(0, 2))
{
case 0:
case 1:
- {
- // basic crossover
- const auto crossover_point_begin_itr = c1_ops.begin() + point->p1_crossover_point;
- const auto crossover_point_end_itr = c1_ops.begin() + c1.find_endpoint(point->p1_crossover_point);
-
- const auto found_point_begin_itr = c2_ops.begin() + point->p2_crossover_point;
- const auto found_point_end_itr = c2_ops.begin() + c2.find_endpoint(point->p2_crossover_point);
-
- stack_allocator& c1_stack = c1.get_values();
- stack_allocator& c2_stack = c2.get_values();
-
- for (const auto& op : iterate(crossover_point_begin_itr, crossover_point_end_itr))
- c1_operators.push_back(op);
- for (const auto& op : iterate(found_point_begin_itr, found_point_end_itr))
- c2_operators.push_back(op);
-
- const size_t c1_stack_after_bytes = accumulate_type_sizes(crossover_point_end_itr, c1_ops.end());
- const size_t c1_stack_for_bytes = accumulate_type_sizes(crossover_point_begin_itr, crossover_point_end_itr);
- const size_t c2_stack_after_bytes = accumulate_type_sizes(found_point_end_itr, c2_ops.end());
- const size_t c2_stack_for_bytes = accumulate_type_sizes(found_point_begin_itr, found_point_end_itr);
- const auto c1_total = static_cast(c1_stack_after_bytes + c1_stack_for_bytes);
- const auto c2_total = static_cast(c2_stack_after_bytes + c2_stack_for_bytes);
- const auto copy_ptr_c1 = get_thread_pointer_for_size(c1_total);
- const auto copy_ptr_c2 = get_thread_pointer_for_size(c2_total);
-
- c1_stack.reserve(c1_stack.bytes_in_head() - c1_stack_for_bytes + c2_stack_for_bytes);
- c2_stack.reserve(c2_stack.bytes_in_head() - c2_stack_for_bytes + c1_stack_for_bytes);
-
- c1_stack.copy_to(copy_ptr_c1, c1_total);
- c1_stack.pop_bytes(c1_total);
-
- c2_stack.copy_to(copy_ptr_c2, c2_total);
- c2_stack.pop_bytes(c2_total);
-
- c2_stack.copy_from(copy_ptr_c1, c1_stack_for_bytes);
- c2_stack.copy_from(copy_ptr_c2 + c2_stack_for_bytes, c2_stack_after_bytes);
-
- c1_stack.copy_from(copy_ptr_c2, c2_stack_for_bytes);
- c1_stack.copy_from(copy_ptr_c1 + c1_stack_for_bytes, c1_stack_after_bytes);
-
- // now swap the operators
- auto insert_point_c1 = crossover_point_begin_itr - 1;
- auto insert_point_c2 = found_point_begin_itr - 1;
-
- // invalidates [begin, end()) so the insert points should be fine
- c1_ops.erase(crossover_point_begin_itr, crossover_point_end_itr);
- c2_ops.erase(found_point_begin_itr, found_point_end_itr);
-
- c1_ops.insert(++insert_point_c1, c2_operators.begin(), c2_operators.end());
- c2_ops.insert(++insert_point_c2, c1_operators.begin(), c1_operators.end());
- }
- break;
+ // TODO: use the tree's selector methods!
+ c1.swap_subtrees({point->p1_crossover_point, 0}, c2, {point->p2_crossover_point, 0});
break;
default:
#if BLT_DEBUG_LEVEL > 0
diff --git a/src/tree.cpp b/src/tree.cpp
index a6487e2..cd61338 100644
--- a/src/tree.cpp
+++ b/src/tree.cpp
@@ -24,17 +24,13 @@
namespace blt::gp
{
- // this one will copy previous bytes over
- template
- blt::span get_pointer_for_size(blt::size_t size)
+ template
+ static u8* get_thread_pointer_for_size(const size_t bytes)
{
- static blt::span buffer{nullptr, 0};
- if (buffer.size() < size)
- {
- delete[] buffer.data();
- buffer = {new blt::u8[size], size};
- }
- return buffer;
+ thread_local expanding_buffer buffer;
+ if (bytes > buffer.size())
+ buffer.resize(bytes);
+ return buffer.data();
}
std::ostream& create_indent(std::ostream& out, blt::size_t amount, bool pretty_print)
@@ -198,7 +194,8 @@ namespace blt::gp
return {static_cast(point), info.return_type};
if (m_program->get_random().choice(terminal_chance))
return {static_cast(point), info.return_type};
- } while (true);
+ }
+ while (true);
}
std::optional tree_t::select_subtree(const type_id type, const u32 max_tries, const double terminal_chance) const
@@ -252,6 +249,121 @@ namespace blt::gp
return {};
}
+ void tree_t::copy_subtree(const subtree_point_t point, std::vector& operators, stack_allocator& stack)
+ {
+ const auto point_begin_itr = operations.begin() + point.pos;
+ const auto point_end_itr = operations.begin() + find_endpoint(point.pos);
+
+ const size_t after_bytes = accumulate_type_sizes(point_end_itr, operations.end());
+
+ operators.reserve(operators.size() + std::distance(point_begin_itr, point_end_itr));
+ size_t for_bytes = 0;
+ for (auto& it : iterate(point_begin_itr, point_end_itr))
+ {
+ if (it.is_value())
+ {
+ if (it.get_flags().is_ephemeral() && it.has_ephemeral_drop())
+ {
+ auto& ptr = values.access_pointer_forward(for_bytes, it.type_size());
+ ++*ptr;
+ }
+ for_bytes += it.type_size();
+ }
+ operators.emplace_back(it);
+ }
+
+ auto* after_ptr = get_thread_pointer_for_size(after_bytes);
+
+ // TODO: this is inefficient, make a copy range function in stack
+ values.copy_to(after_ptr, after_bytes);
+ values.pop_bytes(after_bytes);
+ stack.copy_from(values, for_bytes);
+ values.copy_from(after_ptr, after_bytes);
+ }
+
+ void tree_t::swap_subtrees(const subtree_point_t our_subtree, tree_t& other_tree, const subtree_point_t other_subtree)
+ {
+ const auto our_point_begin_itr = operations.begin() + our_subtree.pos;
+ const auto our_point_end_itr = operations.begin() + find_endpoint(our_subtree.pos);
+
+ const auto other_point_begin_itr = other_tree.operations.begin() + other_subtree.pos;
+ const auto other_point_end_itr = other_tree.operations.begin() + other_tree.find_endpoint(other_subtree.pos);
+
+ thread_local tracked_vector c1_operators;
+ thread_local tracked_vector c2_operators;
+ c1_operators.clear();
+ c2_operators.clear();
+
+ c1_operators.reserve(std::distance(our_point_begin_itr, our_point_end_itr));
+ c2_operators.reserve(std::distance(other_point_begin_itr, other_point_end_itr));
+
+ // i don't think this is required for swapping values, since the total number of additions is net zero
+ // the tree isn't destroyed at any point.
+
+ size_t for_our_bytes = 0;
+ for (const auto& it : iterate(our_point_begin_itr, our_point_end_itr))
+ {
+ if (it.is_value())
+ {
+ // if (it.get_flags().is_ephemeral() && it.has_ephemeral_drop())
+ // {
+ // auto& ptr = values.access_pointer_forward(for_our_bytes, it.type_size());
+ // ++*ptr;
+ // }
+ for_our_bytes += it.type_size();
+ }
+ c1_operators.emplace_back(it);
+ }
+
+ size_t for_other_bytes = 0;
+ for (const auto& it : iterate(other_point_begin_itr, other_point_end_itr))
+ {
+ if (it.is_value())
+ {
+ // if (it.get_flags().is_ephemeral() && it.has_ephemeral_drop())
+ // {
+ // auto& ptr = values.access_pointer_forward(for_other_bytes, it.type_size());
+ // ++*ptr;
+ // }
+ for_other_bytes += it.type_size();
+ }
+ c2_operators.emplace_back(it);
+ }
+
+ const size_t c1_stack_after_bytes = accumulate_type_sizes(our_point_end_itr, operations.end());
+ const size_t c2_stack_after_bytes = accumulate_type_sizes(other_point_end_itr, other_tree.operations.end());
+ const auto c1_total = static_cast(c1_stack_after_bytes + for_our_bytes);
+ const auto c2_total = static_cast(c2_stack_after_bytes + for_other_bytes);
+ const auto copy_ptr_c1 = get_thread_pointer_for_size(c1_total);
+ const auto copy_ptr_c2 = get_thread_pointer_for_size(c2_total);
+
+ values.reserve(values.bytes_in_head() - for_our_bytes + for_other_bytes);
+ other_tree.values.reserve(other_tree.values.bytes_in_head() - for_other_bytes + for_our_bytes);
+
+ values.copy_to(copy_ptr_c1, c1_total);
+ values.pop_bytes(c1_total);
+
+ other_tree.values.copy_to(copy_ptr_c2, c2_total);
+ other_tree.values.pop_bytes(c2_total);
+
+ other_tree.values.copy_from(copy_ptr_c1, for_our_bytes);
+ other_tree.values.copy_from(copy_ptr_c2 + for_other_bytes, c2_stack_after_bytes);
+
+ values.copy_from(copy_ptr_c2, for_other_bytes);
+ values.copy_from(copy_ptr_c1 + for_our_bytes, c1_stack_after_bytes);
+
+ // now swap the operators
+ auto insert_point_c1 = our_point_begin_itr - 1;
+ auto insert_point_c2 = other_point_begin_itr - 1;
+
+ // invalidates [begin, end()) so the insert points should be fine
+ operations.erase(our_point_begin_itr, our_point_end_itr);
+ other_tree.operations.erase(other_point_begin_itr, other_point_end_itr);
+
+ operations.insert(++insert_point_c1, c2_operators.begin(), c2_operators.end());
+ other_tree.operations.insert(++insert_point_c2, c1_operators.begin(), c1_operators.end());
+ }
+
ptrdiff_t tree_t::find_endpoint(ptrdiff_t start) const
{
@@ -387,11 +499,11 @@ namespace blt::gp
{
auto& ptr = values.access_pointer_forward(total_bytes, op.type_size());
--*ptr;
- BLT_TRACE(ptr->load());
+ // BLT_TRACE(ptr->load());
if (*ptr == 0)
{
// BLT_TRACE("Deleting pointers!");
- delete ptr.get();
+ // delete ptr.get();
}
}
total_bytes += op.type_size();