The reader_lifecycle_policy API was created around the idea of shard readers (optionally) being saved and reused on the next page. To do this, the lifecycle policy has to also be able to control the lifecycle of by-reference parameters of readers: the slice and the range. This was possible from day 1, as the readers are created through the lifecycle policy, which can intercept and replace the said parameters with copies that are created in stable storage. There was one whole in the design though: fast-forwarding, which can change the range of the read, without the lifecycle policy knowing about this. In practice this results in fast-forwarded readers being saved together with the wrong range, their range reference becoming stale. The only lifecycle implementation prone to this is the one in `multishard_mutation_query.cc`, as it is the only one actually saving readers. It will fast-forward its reader when the query happens over multiple ranges. There were no problems related to this so far because no one passes more than one range to said functions, but this is incidental. This patch solves this by adding an `update_read_range()` method to the lifecycle policy, allowing the shard reader to update the read range when being fast forwarded. To allow the shard reader to also have control over the lifecycle of this range, a shared pointer is used. This control is required because when an `evictable_reader` is the top-level reader on the shard, it can invoke `create_reader()` with an edited range after `update_read_range()`, replacing the fast-forwarded-to range with a new one, yanking it out from under the feet of the evictable reader itself. By using a shared pointer here, we can ensure the range stays alive while it is the current one.
475 lines
19 KiB
C++
475 lines
19 KiB
C++
/*
|
|
* Copyright (C) 2018-present ScyllaDB
|
|
*/
|
|
|
|
/*
|
|
* This file is part of Scylla.
|
|
*
|
|
* Scylla is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU Affero General Public License as published by
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* Scylla is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with Scylla. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include <seastar/util/closeable.hh>
|
|
|
|
#include "mutation_compactor.hh"
|
|
#include "mutation_reader.hh"
|
|
|
|
#include <boost/intrusive/set.hpp>
|
|
|
|
#include <variant>
|
|
|
|
namespace query {
|
|
|
|
template <typename Consumer>
|
|
class clustering_position_tracker {
|
|
Consumer _consumer;
|
|
lw_shared_ptr<std::optional<clustering_key_prefix>> _last_ckey;
|
|
|
|
public:
|
|
clustering_position_tracker(Consumer&& consumer, lw_shared_ptr<std::optional<clustering_key_prefix>> last_ckey)
|
|
: _consumer(std::forward<Consumer>(consumer))
|
|
, _last_ckey(std::move(last_ckey)) {
|
|
}
|
|
|
|
void consume_new_partition(const dht::decorated_key& dk) {
|
|
_last_ckey->reset();
|
|
_consumer.consume_new_partition(dk);
|
|
}
|
|
void consume(tombstone t) {
|
|
_consumer.consume(t);
|
|
}
|
|
stop_iteration consume(static_row&& sr, tombstone t, bool is_live) {
|
|
return _consumer.consume(std::move(sr), std::move(t), is_live);
|
|
}
|
|
stop_iteration consume(clustering_row&& cr, row_tombstone t, bool is_live) {
|
|
*_last_ckey = cr.key();
|
|
return _consumer.consume(std::move(cr), std::move(t), is_live);
|
|
}
|
|
stop_iteration consume(range_tombstone&& rt) {
|
|
return _consumer.consume(std::move(rt));
|
|
}
|
|
stop_iteration consume_end_of_partition() {
|
|
return _consumer.consume_end_of_partition();
|
|
}
|
|
auto consume_end_of_stream() {
|
|
return _consumer.consume_end_of_stream();
|
|
}
|
|
};
|
|
|
|
/// Consume a page worth of data from the reader.
|
|
///
|
|
/// Uses `compaction_state` for compacting the fragments and `consumer` for
|
|
/// building the results.
|
|
/// Returns a future containing a tuple with the last consumed clustering key,
|
|
/// or std::nullopt if the last row wasn't a clustering row, and whatever the
|
|
/// consumer's `consume_end_of_stream()` method returns.
|
|
template <emit_only_live_rows OnlyLive, typename Consumer>
|
|
requires CompactedFragmentsConsumer<Consumer>
|
|
auto consume_page(flat_mutation_reader& reader,
|
|
lw_shared_ptr<compact_for_query_state<OnlyLive>> compaction_state,
|
|
const query::partition_slice& slice,
|
|
Consumer&& consumer,
|
|
uint64_t row_limit,
|
|
uint32_t partition_limit,
|
|
gc_clock::time_point query_time) {
|
|
return reader.peek().then([=, &reader, consumer = std::move(consumer), &slice] (
|
|
mutation_fragment* next_fragment) mutable {
|
|
const auto next_fragment_kind = next_fragment ? next_fragment->mutation_fragment_kind() : mutation_fragment::kind::partition_end;
|
|
compaction_state->start_new_page(row_limit, partition_limit, query_time, next_fragment_kind, consumer);
|
|
|
|
auto last_ckey = make_lw_shared<std::optional<clustering_key_prefix>>();
|
|
auto reader_consumer = make_stable_flattened_mutations_consumer<compact_for_query<OnlyLive, clustering_position_tracker<Consumer>>>(
|
|
compaction_state,
|
|
clustering_position_tracker(std::move(consumer), last_ckey));
|
|
|
|
return reader.consume(std::move(reader_consumer)).then([last_ckey] (auto&&... results) mutable {
|
|
static_assert(sizeof...(results) <= 1);
|
|
return make_ready_future<std::tuple<std::optional<clustering_key_prefix>, std::decay_t<decltype(results)>...>>(std::tuple(std::move(*last_ckey), std::move(results)...));
|
|
});
|
|
});
|
|
}
|
|
|
|
struct position_view {
|
|
const dht::decorated_key* partition_key;
|
|
const clustering_key_prefix* clustering_key;
|
|
};
|
|
|
|
class querier_base {
|
|
friend class querier_utils;
|
|
|
|
protected:
|
|
schema_ptr _schema;
|
|
reader_permit _permit;
|
|
lw_shared_ptr<const dht::partition_range> _range;
|
|
std::unique_ptr<const query::partition_slice> _slice;
|
|
std::variant<flat_mutation_reader, reader_concurrency_semaphore::inactive_read_handle> _reader;
|
|
dht::partition_ranges_view _query_ranges;
|
|
|
|
public:
|
|
querier_base(reader_permit permit, lw_shared_ptr<const dht::partition_range> range,
|
|
std::unique_ptr<const query::partition_slice> slice, flat_mutation_reader reader, dht::partition_ranges_view query_ranges)
|
|
: _schema(reader.schema())
|
|
, _permit(std::move(permit))
|
|
, _range(std::move(range))
|
|
, _slice(std::move(slice))
|
|
, _reader(std::move(reader))
|
|
, _query_ranges(query_ranges)
|
|
{ }
|
|
|
|
querier_base(schema_ptr schema, reader_permit permit, dht::partition_range range,
|
|
query::partition_slice slice, const mutation_source& ms, const io_priority_class& pc, tracing::trace_state_ptr trace_ptr)
|
|
: _schema(std::move(schema))
|
|
, _permit(std::move(permit))
|
|
, _range(make_lw_shared<const dht::partition_range>(std::move(range)))
|
|
, _slice(std::make_unique<const query::partition_slice>(std::move(slice)))
|
|
, _reader(ms.make_reader(_schema, _permit, *_range, *_slice, pc, std::move(trace_ptr), streamed_mutation::forwarding::no, mutation_reader::forwarding::no))
|
|
, _query_ranges(*_range)
|
|
{ }
|
|
|
|
querier_base(querier_base&&) = default;
|
|
querier_base& operator=(querier_base&&) = default;
|
|
|
|
virtual ~querier_base() = default;
|
|
|
|
const ::schema& schema() const {
|
|
return *_schema;
|
|
}
|
|
|
|
reader_permit& permit() {
|
|
return _permit;
|
|
}
|
|
|
|
bool is_reversed() const {
|
|
return _slice->options.contains(query::partition_slice::option::reversed);
|
|
}
|
|
|
|
virtual position_view current_position() const = 0;
|
|
|
|
dht::partition_ranges_view ranges() const {
|
|
return _query_ranges;
|
|
}
|
|
|
|
size_t memory_usage() const {
|
|
return _permit.consumed_resources().memory;
|
|
}
|
|
|
|
future<> close() noexcept;
|
|
};
|
|
|
|
/// One-stop object for serving queries.
|
|
///
|
|
/// Encapsulates all state and logic for serving all pages for a given range
|
|
/// of a query on a given shard. Can be used with any CompactedMutationsConsumer
|
|
/// certified result-builder.
|
|
/// Intended to be created on the first page of a query then saved and reused on
|
|
/// subsequent pages.
|
|
/// (1) Create with the parameters of your query.
|
|
/// (2) Call consume_page() with your consumer to consume the contents of the
|
|
/// next page.
|
|
/// (3) At the end of the page save the querier if you expect more pages.
|
|
/// The are_limits_reached() method can be used to determine whether the
|
|
/// page was filled or not. Also check your result builder for short reads.
|
|
/// Most result builders have memory-accounters that will stop the read
|
|
/// once some memory limit was reached. This is called a short read as the
|
|
/// read stops before the row and/or partition limits are reached.
|
|
/// (4) At the beginning of the next page validate whether it can be used with
|
|
/// the page's schema and start position. In case a schema or position
|
|
/// mismatch is detected the querier shouldn't be used to produce the next
|
|
/// page. It should be dropped instead and a new one should be created
|
|
/// instead.
|
|
template <emit_only_live_rows OnlyLive>
|
|
class querier : public querier_base {
|
|
lw_shared_ptr<compact_for_query_state<OnlyLive>> _compaction_state;
|
|
std::optional<clustering_key_prefix> _last_ckey;
|
|
|
|
public:
|
|
querier(const mutation_source& ms,
|
|
schema_ptr schema,
|
|
reader_permit permit,
|
|
dht::partition_range range,
|
|
query::partition_slice slice,
|
|
const io_priority_class& pc,
|
|
tracing::trace_state_ptr trace_ptr)
|
|
: querier_base(schema, permit, std::move(range), std::move(slice), ms, pc, std::move(trace_ptr))
|
|
, _compaction_state(make_lw_shared<compact_for_query_state<OnlyLive>>(*schema, gc_clock::time_point{}, *_slice, 0, 0)) {
|
|
}
|
|
|
|
bool are_limits_reached() const {
|
|
return _compaction_state->are_limits_reached();
|
|
}
|
|
|
|
template <typename Consumer>
|
|
requires CompactedFragmentsConsumer<Consumer>
|
|
auto consume_page(Consumer&& consumer,
|
|
uint64_t row_limit,
|
|
uint32_t partition_limit,
|
|
gc_clock::time_point query_time,
|
|
tracing::trace_state_ptr trace_ptr = {}) {
|
|
return ::query::consume_page(std::get<flat_mutation_reader>(_reader), _compaction_state, *_slice, std::move(consumer), row_limit,
|
|
partition_limit, query_time).then([this, trace_ptr = std::move(trace_ptr)] (auto&& results) {
|
|
_last_ckey = std::get<std::optional<clustering_key>>(std::move(results));
|
|
const auto& cstats = _compaction_state->stats();
|
|
tracing::trace(trace_ptr, "Page stats: {} partition(s), {} static row(s) ({} live, {} dead), {} clustering row(s) ({} live, {} dead) and {} range tombstone(s)",
|
|
cstats.partitions,
|
|
cstats.static_rows.total(),
|
|
cstats.static_rows.live,
|
|
cstats.static_rows.dead,
|
|
cstats.clustering_rows.total(),
|
|
cstats.clustering_rows.live,
|
|
cstats.clustering_rows.dead,
|
|
cstats.range_tombstones);
|
|
constexpr auto size = std::tuple_size<std::decay_t<decltype(results)>>::value;
|
|
static_assert(size <= 2);
|
|
if constexpr (size == 1) {
|
|
return make_ready_future<>();
|
|
} else {
|
|
auto result = std::get<1>(std::move(results));
|
|
return make_ready_future<std::decay_t<decltype(result)>>(std::move(result));
|
|
}
|
|
});
|
|
}
|
|
|
|
virtual position_view current_position() const override {
|
|
const dht::decorated_key* dk = _compaction_state->current_partition();
|
|
const clustering_key_prefix* clustering_key = _last_ckey ? &*_last_ckey : nullptr;
|
|
return {dk, clustering_key};
|
|
}
|
|
};
|
|
|
|
using data_querier = querier<emit_only_live_rows::yes>;
|
|
using mutation_querier = querier<emit_only_live_rows::no>;
|
|
|
|
/// Local state of a multishard query.
|
|
///
|
|
/// This querier is not intended to be used directly to read pages. Instead it
|
|
/// is merely a shard local state of a suspended multishard query and is
|
|
/// intended to be used for storing the state of the query on each shard where
|
|
/// it executes. It stores the local reader and the referenced parameters it was
|
|
/// created with (similar to other queriers).
|
|
/// For position validation purposes (at lookup) the reader's position is
|
|
/// considered to be the same as that of the query.
|
|
class shard_mutation_querier : public querier_base {
|
|
std::unique_ptr<const dht::partition_range_vector> _query_ranges;
|
|
dht::decorated_key _nominal_pkey;
|
|
std::optional<clustering_key_prefix> _nominal_ckey;
|
|
|
|
private:
|
|
shard_mutation_querier(
|
|
std::unique_ptr<const dht::partition_range_vector> query_ranges,
|
|
lw_shared_ptr<const dht::partition_range> reader_range,
|
|
std::unique_ptr<const query::partition_slice> reader_slice,
|
|
flat_mutation_reader reader,
|
|
reader_permit permit,
|
|
dht::decorated_key nominal_pkey,
|
|
std::optional<clustering_key_prefix> nominal_ckey)
|
|
: querier_base(permit, std::move(reader_range), std::move(reader_slice), std::move(reader), *query_ranges)
|
|
, _query_ranges(std::move(query_ranges))
|
|
, _nominal_pkey(std::move(nominal_pkey))
|
|
, _nominal_ckey(std::move(nominal_ckey)) {
|
|
}
|
|
|
|
|
|
public:
|
|
shard_mutation_querier(
|
|
const dht::partition_range_vector query_ranges,
|
|
lw_shared_ptr<const dht::partition_range> reader_range,
|
|
std::unique_ptr<const query::partition_slice> reader_slice,
|
|
flat_mutation_reader reader,
|
|
reader_permit permit,
|
|
dht::decorated_key nominal_pkey,
|
|
std::optional<clustering_key_prefix> nominal_ckey)
|
|
: shard_mutation_querier(std::make_unique<const dht::partition_range_vector>(std::move(query_ranges)), std::move(reader_range),
|
|
std::move(reader_slice), std::move(reader), std::move(permit), std::move(nominal_pkey), std::move(nominal_ckey)) {
|
|
}
|
|
|
|
virtual position_view current_position() const override {
|
|
return {&_nominal_pkey, _nominal_ckey ? &*_nominal_ckey : nullptr};
|
|
}
|
|
|
|
lw_shared_ptr<const dht::partition_range> reader_range() && {
|
|
return std::move(_range);
|
|
}
|
|
|
|
std::unique_ptr<const query::partition_slice> reader_slice() && {
|
|
return std::move(_slice);
|
|
}
|
|
|
|
flat_mutation_reader reader() && {
|
|
return std::move(std::get<flat_mutation_reader>(_reader));
|
|
}
|
|
};
|
|
|
|
/// Special-purpose cache for saving queriers between pages.
|
|
///
|
|
/// Queriers are saved at the end of the page and looked up at the beginning of
|
|
/// the next page. The lookup() always removes the querier from the cache, it
|
|
/// has to be inserted again at the end of the page.
|
|
/// Lookup provides the following extra logic, special to queriers:
|
|
/// * It accepts a factory function which is used to create a new querier if
|
|
/// the lookup fails (see below). This allows for simple call sites.
|
|
/// * It does range matching. A query sometimes will result in multiple querier
|
|
/// objects executing on the same node and shard paralelly. To identify the
|
|
/// appropriate querier lookup() will consider - in addition to the lookup
|
|
/// key - the read range.
|
|
/// * It does schema version and position checking. In some case a subsequent
|
|
/// page will have a different schema version or will start from a position
|
|
/// that is before the end position of the previous page. lookup() will
|
|
/// recognize these cases and drop the previous querier and create a new one.
|
|
///
|
|
/// Inserted queriers will have a TTL. When this expires the querier is
|
|
/// evicted. This is to avoid excess and unnecessary resource usage due to
|
|
/// abandoned queriers.
|
|
/// Registers cached readers with the reader concurrency semaphore, as inactive
|
|
/// readers, so the latter can evict them if needed.
|
|
/// Keeps the total memory consumption of cached queriers
|
|
/// below max_queriers_memory_usage by evicting older entries upon inserting
|
|
/// new ones if the the memory consupmtion would go above the limit.
|
|
class querier_cache {
|
|
public:
|
|
static const std::chrono::seconds default_entry_ttl;
|
|
|
|
struct stats {
|
|
// The number of inserts into the cache.
|
|
uint64_t inserts = 0;
|
|
// The number of cache lookups.
|
|
uint64_t lookups = 0;
|
|
// The subset of lookups that missed.
|
|
uint64_t misses = 0;
|
|
// The subset of lookups that hit but the looked up querier had to be
|
|
// dropped due to position mismatch.
|
|
uint64_t drops = 0;
|
|
// The number of queriers evicted due to their TTL expiring.
|
|
uint64_t time_based_evictions = 0;
|
|
// The number of queriers evicted to free up resources to be able to
|
|
// create new readers.
|
|
uint64_t resource_based_evictions = 0;
|
|
// The number of queriers currently in the cache.
|
|
uint64_t population = 0;
|
|
};
|
|
|
|
using index = std::unordered_multimap<utils::UUID, std::unique_ptr<querier_base>>;
|
|
|
|
private:
|
|
index _data_querier_index;
|
|
index _mutation_querier_index;
|
|
index _shard_mutation_querier_index;
|
|
std::chrono::seconds _entry_ttl;
|
|
stats _stats;
|
|
gate _closing_gate;
|
|
|
|
private:
|
|
template <typename Querier>
|
|
void insert_querier(
|
|
utils::UUID key,
|
|
querier_cache::index& index,
|
|
querier_cache::stats& stats,
|
|
Querier&& q,
|
|
std::chrono::seconds ttl,
|
|
tracing::trace_state_ptr trace_state);
|
|
|
|
template <typename Querier>
|
|
std::optional<Querier> lookup_querier(
|
|
querier_cache::index& index,
|
|
utils::UUID key,
|
|
const schema& s,
|
|
dht::partition_ranges_view ranges,
|
|
const query::partition_slice& slice,
|
|
tracing::trace_state_ptr trace_state,
|
|
db::timeout_clock::time_point timeout);
|
|
|
|
public:
|
|
explicit querier_cache(std::chrono::seconds entry_ttl = default_entry_ttl);
|
|
|
|
querier_cache(const querier_cache&) = delete;
|
|
querier_cache& operator=(const querier_cache&) = delete;
|
|
|
|
// this is captured
|
|
querier_cache(querier_cache&&) = delete;
|
|
querier_cache& operator=(querier_cache&&) = delete;
|
|
|
|
void insert(utils::UUID key, data_querier&& q, tracing::trace_state_ptr trace_state);
|
|
|
|
void insert(utils::UUID key, mutation_querier&& q, tracing::trace_state_ptr trace_state);
|
|
|
|
void insert(utils::UUID key, shard_mutation_querier&& q, tracing::trace_state_ptr trace_state);
|
|
|
|
/// Lookup a data querier in the cache.
|
|
///
|
|
/// Queriers are found based on `key` and `range`. There may be multiple
|
|
/// queriers for the same `key` differentiated by their read range. Since
|
|
/// each subsequent page may have a narrower read range then the one before
|
|
/// it ranges cannot be simply matched based on equality. For matching we
|
|
/// use the fact that the coordinator splits the query range into
|
|
/// non-overlapping ranges. Thus both bounds of any range, or in case of
|
|
/// singular ranges only the start bound are guaranteed to be unique.
|
|
///
|
|
/// The found querier is checked for a matching position and schema version.
|
|
/// The start position of the querier is checked against the start position
|
|
/// of the page using the `range' and `slice'.
|
|
std::optional<data_querier> lookup_data_querier(utils::UUID key,
|
|
const schema& s,
|
|
const dht::partition_range& range,
|
|
const query::partition_slice& slice,
|
|
tracing::trace_state_ptr trace_state,
|
|
db::timeout_clock::time_point timeout);
|
|
|
|
/// Lookup a mutation querier in the cache.
|
|
///
|
|
/// See \ref lookup_data_querier().
|
|
std::optional<mutation_querier> lookup_mutation_querier(utils::UUID key,
|
|
const schema& s,
|
|
const dht::partition_range& range,
|
|
const query::partition_slice& slice,
|
|
tracing::trace_state_ptr trace_state,
|
|
db::timeout_clock::time_point timeout);
|
|
|
|
/// Lookup a shard mutation querier in the cache.
|
|
///
|
|
/// See \ref lookup_data_querier().
|
|
std::optional<shard_mutation_querier> lookup_shard_mutation_querier(utils::UUID key,
|
|
const schema& s,
|
|
const dht::partition_range_vector& ranges,
|
|
const query::partition_slice& slice,
|
|
tracing::trace_state_ptr trace_state,
|
|
db::timeout_clock::time_point timeout);
|
|
|
|
/// Change the ttl of cache entries
|
|
///
|
|
/// Applies only to entries inserted after the change.
|
|
void set_entry_ttl(std::chrono::seconds entry_ttl);
|
|
|
|
/// Evict a querier.
|
|
///
|
|
/// Return true if a querier was evicted and false otherwise (if the cache
|
|
/// is empty).
|
|
future<bool> evict_one() noexcept;
|
|
|
|
/// Evict all queriers that belong to a table.
|
|
///
|
|
/// Should be used when dropping a table.
|
|
future<> evict_all_for_table(const utils::UUID& schema_id) noexcept;
|
|
|
|
/// Close all queriers and wait on background work.
|
|
///
|
|
/// Should be used before destroying the querier_cache.
|
|
future<> stop() noexcept;
|
|
|
|
const stats& get_stats() const {
|
|
return _stats;
|
|
}
|
|
};
|
|
|
|
} // namespace query
|