540 lines
21 KiB
C++
540 lines
21 KiB
C++
/*
|
|
* Copyright (C) 2016 ScyllaDB
|
|
*/
|
|
|
|
/*
|
|
* This file is part of Scylla.
|
|
*
|
|
* Scylla is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU Affero General Public License as published by
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* Scylla is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with Scylla. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
|
|
#include <seastar/core/thread.hh>
|
|
#include <seastar/tests/test-utils.hh>
|
|
|
|
#include "mutation_source_test.hh"
|
|
#include "streamed_mutation.hh"
|
|
#include "frozen_mutation.hh"
|
|
#include "tests/test_services.hh"
|
|
#include "schema_builder.hh"
|
|
#include "total_order_check.hh"
|
|
|
|
#include "disk-error-handler.hh"
|
|
|
|
thread_local disk_error_signal_type commit_error;
|
|
thread_local disk_error_signal_type general_disk_error;
|
|
|
|
void check_order_of_fragments(streamed_mutation sm)
|
|
{
|
|
stdx::optional<position_in_partition> previous;
|
|
position_in_partition::less_compare cmp(*sm.schema());
|
|
auto mf = sm().get0();
|
|
while (mf) {
|
|
if (previous) {
|
|
BOOST_REQUIRE(cmp(*previous, mf->position()));
|
|
}
|
|
previous = position_in_partition(mf->position());
|
|
mf = sm().get0();
|
|
}
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_mutation_from_streamed_mutation_from_mutation) {
|
|
return seastar::async([] {
|
|
for_each_mutation([&] (const mutation& m) {
|
|
auto get_sm = [&] {
|
|
return streamed_mutation_from_mutation(mutation(m));
|
|
};
|
|
|
|
check_order_of_fragments(get_sm());
|
|
auto mopt = mutation_from_streamed_mutation(get_sm()).get0();
|
|
BOOST_REQUIRE(mopt);
|
|
BOOST_REQUIRE_EQUAL(m, *mopt);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_abandoned_streamed_mutation_from_mutation) {
|
|
return seastar::async([] {
|
|
for_each_mutation([&] (const mutation& m) {
|
|
auto sm = streamed_mutation_from_mutation(mutation(m));
|
|
sm().get();
|
|
sm().get();
|
|
// We rely on AddressSanitizer telling us if nothing was leaked.
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_mutation_merger) {
|
|
return seastar::async([] {
|
|
for_each_mutation_pair([&] (const mutation& m1, const mutation& m2, are_equal) {
|
|
if (m1.schema()->version() != m2.schema()->version()) {
|
|
return;
|
|
}
|
|
|
|
auto m12 = m1;
|
|
m12.apply(m2);
|
|
|
|
auto get_sm = [&] {
|
|
std::vector<streamed_mutation> sms;
|
|
sms.emplace_back(streamed_mutation_from_mutation(mutation(m1)));
|
|
sms.emplace_back(streamed_mutation_from_mutation(mutation(m2.schema(), m1.decorated_key(), m2.partition())));
|
|
return merge_mutations(std::move(sms));
|
|
};
|
|
|
|
check_order_of_fragments(get_sm());
|
|
auto mopt = mutation_from_streamed_mutation(get_sm()).get0();
|
|
BOOST_REQUIRE(mopt);
|
|
BOOST_REQUIRE(m12.partition().difference(m1.schema(), mopt->partition()).empty());
|
|
BOOST_REQUIRE(mopt->partition().difference(m1.schema(), m12.partition()).empty());
|
|
});
|
|
});
|
|
}
|
|
|
|
// A StreamedMutationConsumer which distributes fragments randomly into several mutations.
|
|
class fragment_scatterer {
|
|
std::vector<mutation>& _mutations;
|
|
size_t _next = 0;
|
|
private:
|
|
template<typename Func>
|
|
void for_each_target(Func&& func) {
|
|
// round-robin
|
|
func(_mutations[_next % _mutations.size()]);
|
|
++_next;
|
|
}
|
|
public:
|
|
fragment_scatterer(std::vector<mutation>& muts)
|
|
: _mutations(muts)
|
|
{ }
|
|
|
|
stop_iteration consume(tombstone t) {
|
|
for_each_target([&] (mutation& m) {
|
|
m.partition().apply(t);
|
|
});
|
|
return stop_iteration::no;
|
|
}
|
|
|
|
stop_iteration consume(range_tombstone&& rt) {
|
|
for_each_target([&] (mutation& m) {
|
|
m.partition().apply_row_tombstone(*m.schema(), std::move(rt));
|
|
});
|
|
return stop_iteration::no;
|
|
}
|
|
|
|
stop_iteration consume(static_row&& sr) {
|
|
for_each_target([&] (mutation& m) {
|
|
m.partition().static_row().apply(*m.schema(), column_kind::static_column, std::move(sr.cells()));
|
|
});
|
|
return stop_iteration::no;
|
|
}
|
|
|
|
stop_iteration consume(clustering_row&& cr) {
|
|
for_each_target([&] (mutation& m) {
|
|
auto& dr = m.partition().clustered_row(*m.schema(), std::move(cr.key()));
|
|
dr.apply(cr.tomb());
|
|
dr.apply(cr.marker());
|
|
dr.cells().apply(*m.schema(), column_kind::regular_column, std::move(cr.cells()));
|
|
});
|
|
return stop_iteration::no;
|
|
}
|
|
|
|
stop_iteration consume_end_of_partition() {
|
|
return stop_iteration::no;
|
|
}
|
|
};
|
|
|
|
SEASTAR_TEST_CASE(test_mutation_merger_conforms_to_mutation_source) {
|
|
return seastar::async([] {
|
|
run_mutation_source_tests([](schema_ptr s, const std::vector<mutation>& partitions) -> mutation_source {
|
|
// We create a mutation source which combines N memtables.
|
|
// The input fragments are spread among the memtables according to some selection logic,
|
|
|
|
const int n = 5;
|
|
|
|
std::vector<lw_shared_ptr<memtable>> memtables;
|
|
for (int i = 0; i < n; ++i) {
|
|
memtables.push_back(make_lw_shared<memtable>(s));
|
|
}
|
|
|
|
for (auto&& m : partitions) {
|
|
std::vector<mutation> muts;
|
|
for (int i = 0; i < n; ++i) {
|
|
muts.push_back(mutation(m.decorated_key(), m.schema()));
|
|
}
|
|
fragment_scatterer c{muts};
|
|
auto sm = streamed_mutation_from_mutation(m);
|
|
do_consume_streamed_mutation_flattened(sm, c).get();
|
|
for (int i = 0; i < n; ++i) {
|
|
memtables[i]->apply(std::move(muts[i]));
|
|
}
|
|
}
|
|
|
|
return mutation_source([memtables] (schema_ptr s,
|
|
const dht::partition_range& range,
|
|
const query::partition_slice& slice,
|
|
const io_priority_class& pc,
|
|
tracing::trace_state_ptr trace_state,
|
|
streamed_mutation::forwarding fwd)
|
|
{
|
|
std::vector<mutation_reader> readers;
|
|
for (int i = 0; i < n; ++i) {
|
|
readers.push_back(memtables[i]->make_reader(s, range, slice, pc, trace_state, fwd));
|
|
}
|
|
return make_combined_reader(std::move(readers));
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_freezing_streamed_mutations) {
|
|
return seastar::async([] {
|
|
storage_service_for_tests ssft;
|
|
|
|
for_each_mutation([&] (const mutation& m) {
|
|
auto fm = freeze(streamed_mutation_from_mutation(mutation(m))).get0();
|
|
|
|
auto m1 = fm.unfreeze(m.schema());
|
|
BOOST_REQUIRE_EQUAL(m, m1);
|
|
|
|
auto fm1 = freeze(m);
|
|
BOOST_REQUIRE(fm.representation() == fm1.representation());
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_fragmenting_and_freezing_streamed_mutations) {
|
|
return seastar::async([] {
|
|
storage_service_for_tests ssft;
|
|
|
|
for_each_mutation([&] (const mutation& m) {
|
|
std::vector<frozen_mutation> fms;
|
|
|
|
fragment_and_freeze(streamed_mutation_from_mutation(mutation(m)), [&] (auto fm, bool frag) {
|
|
BOOST_REQUIRE(!frag);
|
|
fms.emplace_back(std::move(fm));
|
|
return make_ready_future<>();
|
|
}, std::numeric_limits<size_t>::max()).get0();
|
|
|
|
BOOST_REQUIRE_EQUAL(fms.size(), 1);
|
|
|
|
auto m1 = fms.back().unfreeze(m.schema());
|
|
BOOST_REQUIRE_EQUAL(m, m1);
|
|
|
|
fms.clear();
|
|
|
|
stdx::optional<bool> fragmented;
|
|
fragment_and_freeze(streamed_mutation_from_mutation(mutation(m)), [&] (auto fm, bool frag) {
|
|
BOOST_REQUIRE(!fragmented || *fragmented == frag);
|
|
*fragmented = frag;
|
|
fms.emplace_back(std::move(fm));
|
|
return make_ready_future<>();
|
|
}, 1).get0();
|
|
|
|
auto expected_fragments = m.partition().clustered_rows().calculate_size()
|
|
+ m.partition().row_tombstones().size()
|
|
+ !m.partition().static_row().empty();
|
|
BOOST_REQUIRE_EQUAL(fms.size(), std::max(expected_fragments, size_t(1)));
|
|
BOOST_REQUIRE(expected_fragments < 2 || *fragmented);
|
|
|
|
auto m2 = fms.back().unfreeze(m.schema());
|
|
fms.pop_back();
|
|
while (!fms.empty()) {
|
|
m2.partition().apply(*m.schema(), fms.back().partition(), *m.schema());
|
|
fms.pop_back();
|
|
}
|
|
BOOST_REQUIRE_EQUAL(m, m2);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_range_tombstones_stream) {
|
|
return seastar::async([] {
|
|
auto s = schema_builder("ks", "cf")
|
|
.with_column("pk", int32_type, column_kind::partition_key)
|
|
.with_column("ck1", int32_type, column_kind::clustering_key)
|
|
.with_column("ck2", int32_type, column_kind::clustering_key)
|
|
.with_column("r", int32_type)
|
|
.build();
|
|
|
|
auto pk = partition_key::from_single_value(*s, int32_type->decompose(0));
|
|
auto create_ck = [&] (std::vector<int> v) {
|
|
std::vector<bytes> vs;
|
|
boost::transform(v, std::back_inserter(vs), [] (int x) { return int32_type->decompose(x); });
|
|
return clustering_key_prefix::from_exploded(*s, std::move(vs));
|
|
};
|
|
|
|
tombstone t0(0, { });
|
|
tombstone t1(1, { });
|
|
|
|
auto rt1 = range_tombstone(create_ck({ 1 }), t0, bound_kind::incl_start, create_ck({ 1, 3 }), bound_kind::incl_end);
|
|
auto rt2 = range_tombstone(create_ck({ 1, 1 }), t1, bound_kind::incl_start, create_ck({ 1, 3 }), bound_kind::excl_end);
|
|
auto rt3 = range_tombstone(create_ck({ 1, 1 }), t0, bound_kind::incl_start, create_ck({ 2 }), bound_kind::incl_end);
|
|
auto rt4 = range_tombstone(create_ck({ 2 }), t0, bound_kind::excl_start, create_ck({ 2, 2 }), bound_kind::incl_end);
|
|
|
|
mutation_fragment cr1 = clustering_row(create_ck({ 0, 0 }));
|
|
mutation_fragment cr2 = clustering_row(create_ck({ 1, 0 }));
|
|
mutation_fragment cr3 = clustering_row(create_ck({ 1, 1 }));
|
|
auto cr4 = rows_entry(create_ck({ 1, 2 }));
|
|
auto cr5 = rows_entry(create_ck({ 1, 3 }));
|
|
|
|
range_tombstone_stream rts(*s);
|
|
rts.apply(range_tombstone(rt1));
|
|
rts.apply(range_tombstone(rt2));
|
|
rts.apply(range_tombstone(rt4));
|
|
|
|
mutation_fragment_opt mf = rts.get_next(cr1);
|
|
BOOST_REQUIRE(!mf);
|
|
|
|
mf = rts.get_next(cr2);
|
|
BOOST_REQUIRE(mf && mf->is_range_tombstone());
|
|
auto expected1 = range_tombstone(create_ck({ 1 }), t0, bound_kind::incl_start, create_ck({ 1, 1 }), bound_kind::excl_end);
|
|
BOOST_REQUIRE(mf->as_range_tombstone().equal(*s, expected1));
|
|
|
|
mf = rts.get_next(cr2);
|
|
BOOST_REQUIRE(!mf);
|
|
|
|
mf = rts.get_next(mutation_fragment(range_tombstone(rt3)));
|
|
BOOST_REQUIRE(mf && mf->is_range_tombstone());
|
|
BOOST_REQUIRE(mf->as_range_tombstone().equal(*s, rt2));
|
|
|
|
mf = rts.get_next(cr3);
|
|
BOOST_REQUIRE(!mf);
|
|
|
|
mf = rts.get_next(cr4);
|
|
BOOST_REQUIRE(!mf);
|
|
|
|
mf = rts.get_next(cr5);
|
|
BOOST_REQUIRE(mf && mf->is_range_tombstone());
|
|
auto expected2 = range_tombstone(create_ck({ 1, 3 }), t0, bound_kind::incl_start, create_ck({ 1, 3 }), bound_kind::incl_end);
|
|
BOOST_REQUIRE(mf->as_range_tombstone().equal(*s, expected2));
|
|
|
|
mf = rts.get_next();
|
|
BOOST_REQUIRE(mf && mf->is_range_tombstone());
|
|
BOOST_REQUIRE(mf->as_range_tombstone().equal(*s, rt4));
|
|
|
|
mf = rts.get_next();
|
|
BOOST_REQUIRE(!mf);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_mutation_hash) {
|
|
return seastar::async([] {
|
|
for_each_mutation_pair([] (auto&& m1, auto&& m2, are_equal eq) {
|
|
auto get_hash = [] (streamed_mutation m) {
|
|
md5_hasher h;
|
|
m.key().feed_hash(h, *m.schema());
|
|
|
|
mutation_hasher<md5_hasher> mh(*m.schema(), h);
|
|
consume(m, std::move(mh)).get0();
|
|
return h.finalize();
|
|
};
|
|
auto h1 = get_hash(streamed_mutation_from_mutation(mutation(m1)));
|
|
auto h2 = get_hash(streamed_mutation_from_mutation(mutation(m2)));
|
|
if (eq) {
|
|
if (h1 != h2) {
|
|
BOOST_FAIL(sprint("Hash should be equal for %s and %s", m1, m2));
|
|
}
|
|
} else {
|
|
// We're using a strong hasher, collision should be unlikely
|
|
if (h1 == h2) {
|
|
BOOST_FAIL(sprint("Hash should be different for %s and %s", m1, m2));
|
|
}
|
|
}
|
|
});
|
|
});
|
|
}
|
|
|
|
static
|
|
composite cell_name(const schema& s, const clustering_key& ck, const column_definition& col) {
|
|
if (s.is_dense()) {
|
|
return composite::serialize_value(ck.components(s), s.is_compound());
|
|
} else {
|
|
const bytes_view column_name = col.name();
|
|
return composite::serialize_value(boost::range::join(
|
|
boost::make_iterator_range(ck.begin(s), ck.end(s)),
|
|
boost::make_iterator_range(&column_name, &column_name + 1)),
|
|
s.is_compound());
|
|
}
|
|
}
|
|
|
|
static
|
|
composite cell_name_for_static_column(const schema& s, const column_definition& cdef) {
|
|
const bytes_view column_name = cdef.name();
|
|
return composite::serialize_static(s, boost::make_iterator_range(&column_name, &column_name + 1));
|
|
}
|
|
|
|
inline
|
|
composite composite_for_key(const schema& s, const clustering_key& ck) {
|
|
return composite::serialize_value(ck.components(s), s.is_compound());
|
|
}
|
|
|
|
inline
|
|
composite composite_before_key(const schema& s, const clustering_key& ck) {
|
|
return composite::serialize_value(ck.components(s), s.is_compound(), composite::eoc::start);
|
|
}
|
|
|
|
inline
|
|
composite composite_after_prefixed(const schema& s, const clustering_key& ck) {
|
|
return composite::serialize_value(ck.components(s), s.is_compound(), composite::eoc::end);
|
|
}
|
|
|
|
inline
|
|
position_in_partition position_for_row(const clustering_key& ck) {
|
|
return position_in_partition(position_in_partition::clustering_row_tag_t(), ck);
|
|
}
|
|
|
|
inline
|
|
position_in_partition position_before(const clustering_key& ck) {
|
|
return position_in_partition(position_in_partition::range_tag_t(), bound_view(ck, bound_kind::incl_start));
|
|
}
|
|
|
|
inline
|
|
position_in_partition position_after_prefixed(const clustering_key& ck) {
|
|
return position_in_partition(position_in_partition::range_tag_t(), bound_view(ck, bound_kind::incl_end));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_ordering_of_position_in_partition_and_composite_view) {
|
|
return seastar::async([] {
|
|
auto s = schema_builder("ks", "cf")
|
|
.with_column("pk", int32_type, column_kind::partition_key)
|
|
.with_column("ck1", int32_type, column_kind::clustering_key)
|
|
.with_column("ck2", int32_type, column_kind::clustering_key)
|
|
.with_column("s1", int32_type, column_kind::static_column)
|
|
.with_column("v", int32_type)
|
|
.build();
|
|
|
|
const column_definition& v_def = *s->get_column_definition("v");
|
|
const column_definition& s_def = *s->get_column_definition("s1");
|
|
|
|
auto make_ck = [&] (int ck1, int ck2) {
|
|
std::vector<data_value> cells;
|
|
cells.push_back(data_value(ck1));
|
|
cells.push_back(data_value(ck2));
|
|
return clustering_key::from_deeply_exploded(*s, cells);
|
|
};
|
|
|
|
auto ck1 = make_ck(1, 2);
|
|
auto ck2 = make_ck(2, 1);
|
|
auto ck3 = make_ck(2, 3);
|
|
auto ck4 = make_ck(3, 1);
|
|
|
|
using cmp = position_in_partition::composite_tri_compare;
|
|
total_order_check<cmp, position_in_partition, composite>(cmp(*s))
|
|
.next(cell_name_for_static_column(*s, s_def))
|
|
.equal_to(position_range::full().start())
|
|
.next(position_before(ck1))
|
|
.equal_to(composite_before_key(*s, ck1))
|
|
.equal_to(composite_for_key(*s, ck1))
|
|
.equal_to(position_for_row(ck1))
|
|
.next(cell_name(*s, ck1, v_def))
|
|
.next(position_after_prefixed(ck1))
|
|
.equal_to(composite_after_prefixed(*s, ck1))
|
|
.next(position_before(ck2))
|
|
.equal_to(composite_before_key(*s, ck2))
|
|
.equal_to(composite_for_key(*s, ck2))
|
|
.equal_to(position_for_row(ck2))
|
|
.next(cell_name(*s, ck2, v_def))
|
|
.next(position_after_prefixed(ck2))
|
|
.equal_to(composite_after_prefixed(*s, ck2))
|
|
.next(position_before(ck3))
|
|
.equal_to(composite_before_key(*s, ck3))
|
|
.equal_to(composite_for_key(*s, ck3))
|
|
.equal_to(position_for_row(ck3))
|
|
.next(cell_name(*s, ck3, v_def))
|
|
.next(position_after_prefixed(ck3))
|
|
.equal_to(composite_after_prefixed(*s, ck3))
|
|
.next(position_before(ck4))
|
|
.equal_to(composite_before_key(*s, ck4))
|
|
.equal_to(composite_for_key(*s, ck4))
|
|
.equal_to(position_for_row(ck4))
|
|
.next(cell_name(*s, ck4, v_def))
|
|
.next(position_after_prefixed(ck4))
|
|
.equal_to(composite_after_prefixed(*s, ck4))
|
|
.next(position_range::full().end())
|
|
.check();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_ordering_of_position_in_partition_and_composite_view_in_a_dense_table) {
|
|
return seastar::async([] {
|
|
auto s = schema_builder("ks", "cf")
|
|
.with_column("pk", int32_type, column_kind::partition_key)
|
|
.with_column("ck1", int32_type, column_kind::clustering_key)
|
|
.with_column("ck2", int32_type, column_kind::clustering_key)
|
|
.with_column("v", int32_type)
|
|
.set_is_dense(true)
|
|
.build();
|
|
|
|
auto make_ck = [&] (int ck1, stdx::optional<int> ck2 = stdx::nullopt) {
|
|
std::vector<data_value> cells;
|
|
cells.push_back(data_value(ck1));
|
|
if (ck2) {
|
|
cells.push_back(data_value(ck2));
|
|
}
|
|
return clustering_key::from_deeply_exploded(*s, cells);
|
|
};
|
|
|
|
auto ck1 = make_ck(1);
|
|
auto ck2 = make_ck(1, 2);
|
|
auto ck3 = make_ck(2);
|
|
auto ck4 = make_ck(2, 3);
|
|
auto ck5 = make_ck(2, 4);
|
|
auto ck6 = make_ck(3);
|
|
|
|
using cmp = position_in_partition::composite_tri_compare;
|
|
total_order_check<cmp, position_in_partition, composite>(cmp(*s))
|
|
.next(composite())
|
|
.next(position_range::full().start())
|
|
.next(position_before(ck1))
|
|
.equal_to(composite_before_key(*s, ck1))
|
|
.equal_to(composite_for_key(*s, ck1))
|
|
.equal_to(position_for_row(ck1))
|
|
// .next(position_after(ck1)) // FIXME: #1446
|
|
.next(position_before(ck2))
|
|
.equal_to(composite_before_key(*s, ck2))
|
|
.equal_to(composite_for_key(*s, ck2))
|
|
.equal_to(position_for_row(ck2))
|
|
.next(position_after_prefixed(ck2))
|
|
.equal_to(composite_after_prefixed(*s, ck2))
|
|
.next(position_after_prefixed(ck1)) // prefix of ck2
|
|
.equal_to(composite_after_prefixed(*s, ck1))
|
|
.next(position_before(ck3))
|
|
.equal_to(composite_before_key(*s, ck3))
|
|
.equal_to(composite_for_key(*s, ck3))
|
|
.equal_to(position_for_row(ck3))
|
|
// .next(position_after(ck3)) // FIXME: #1446
|
|
.next(position_before(ck4))
|
|
.equal_to(composite_before_key(*s, ck4))
|
|
.equal_to(composite_for_key(*s, ck4))
|
|
.equal_to(position_for_row(ck4))
|
|
.next(position_after_prefixed(ck4))
|
|
.equal_to(composite_after_prefixed(*s, ck4))
|
|
.next(position_before(ck5))
|
|
.equal_to(composite_before_key(*s, ck5))
|
|
.equal_to(composite_for_key(*s, ck5))
|
|
.equal_to(position_for_row(ck5))
|
|
.next(position_after_prefixed(ck5))
|
|
.equal_to(composite_after_prefixed(*s, ck5))
|
|
.next(position_after_prefixed(ck3)) // prefix of ck4-ck5
|
|
.equal_to(composite_after_prefixed(*s, ck3))
|
|
.next(position_before(ck6))
|
|
.equal_to(composite_before_key(*s, ck6))
|
|
.equal_to(composite_for_key(*s, ck6))
|
|
.equal_to(position_for_row(ck6))
|
|
.next(position_after_prefixed(ck6))
|
|
.equal_to(composite_after_prefixed(*s, ck6))
|
|
.next(position_range::full().end())
|
|
.check();
|
|
});
|
|
}
|