Files
scylladb/memtable.hh
Tomasz Grabiec d20fae96a2 lsa: Make reclaimer run synchronously with allocations
The goal is to make allocation less likely to fail. With async
reclaimer there is an implicit bound on the amount of memory that can
be allocated between deferring points. This bound is difficult to
enforce though. Sync reclaimer lifts this limitation off.

Also, allocations which could not be satisfied before because of
fragmentation now will have higher chances of succeeding, although
depending on how much memory is fragmented, that could involve
evicting a lot of segments from cache, so we should still avoid them.

Downside of sync reclaiming is that now references into regions may be
invalidated not only across deferring points but at any allocation
site. compaction_lock can be used to pin data, preferably just
temporarily.
2015-08-31 21:50:18 +02:00

113 lines
3.5 KiB
C++

/*
* Copyright 2015 Cloudius Systems
*/
#pragma once
#include <map>
#include <memory>
#include "database_fwd.hh"
#include "dht/i_partitioner.hh"
#include "schema.hh"
#include "mutation_reader.hh"
#include "db/commitlog/replay_position.hh"
#include "utils/logalloc.hh"
class frozen_mutation;
namespace bi = boost::intrusive;
class partition_entry {
bi::set_member_hook<> _link;
dht::decorated_key _key;
mutation_partition _p;
public:
friend class memtable;
partition_entry(dht::decorated_key key, mutation_partition p)
: _key(std::move(key))
, _p(std::move(p))
{ }
partition_entry(partition_entry&& o) noexcept;
const dht::decorated_key& key() const { return _key; }
dht::decorated_key& key() { return _key; }
const mutation_partition& partition() const { return _p; }
mutation_partition& partition() { return _p; }
struct compare {
dht::decorated_key::less_comparator _c;
compare(schema_ptr s)
: _c(std::move(s))
{}
bool operator()(const dht::decorated_key& k1, const partition_entry& k2) const {
return _c(k1, k2._key);
}
bool operator()(const partition_entry& k1, const partition_entry& k2) const {
return _c(k1._key, k2._key);
}
bool operator()(const partition_entry& k1, const dht::decorated_key& k2) const {
return _c(k1._key, k2);
}
bool operator()(const partition_entry& k1, const dht::ring_position& k2) const {
return _c(k1._key, k2);
}
bool operator()(const dht::ring_position& k1, const partition_entry& k2) const {
return _c(k1, k2._key);
}
};
};
// Managed by lw_shared_ptr<>.
class memtable final : public enable_lw_shared_from_this<memtable> {
public:
using partitions_type = bi::set<partition_entry,
bi::member_hook<partition_entry, bi::set_member_hook<>, &partition_entry::_link>,
bi::compare<partition_entry::compare>>;
private:
schema_ptr _schema;
mutable logalloc::region _region;
partitions_type partitions;
db::replay_position _replay_position;
void update(const db::replay_position&);
friend class row_cache;
private:
boost::iterator_range<partitions_type::const_iterator> slice(const query::partition_range& r) const;
mutation_partition& find_or_create_partition(const dht::decorated_key& key);
mutation_partition& find_or_create_partition_slow(partition_key_view key);
public:
explicit memtable(schema_ptr schema, logalloc::region_group* dirty_memory_region_group = nullptr);
~memtable();
schema_ptr schema() const { return _schema; }
void apply(const mutation& m, const db::replay_position& = db::replay_position());
void apply(const frozen_mutation& m, const db::replay_position& = db::replay_position());
public:
size_t partition_count() const;
logalloc::occupancy_stats occupancy() const;
// Creates a reader of data in this memtable for given partition range.
//
// Live readers share ownership of the memtable instance, so caller
// doesn't need to ensure that memtable remains live.
//
// The 'range' parameter must be live as long as the reader is being used
mutation_reader make_reader(const query::partition_range& range = query::full_partition_range) const;
mutation_source as_data_source();
bool empty() const { return partitions.empty(); }
const db::replay_position& replay_position() const {
return _replay_position;
}
friend class scanning_reader;
};