Purpose is to reuse code and also make it easier to read. Signed-off-by: Raphael S. Carvalho <raphaelsc@scylladb.com>
346 lines
14 KiB
C++
346 lines
14 KiB
C++
/*
|
|
* Copyright (C) 2015 Cloudius Systems, Ltd.
|
|
*/
|
|
|
|
/*
|
|
* This file is part of Scylla.
|
|
*
|
|
* Scylla is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU Affero General Public License as published by
|
|
* the Free Software Foundation, either version 3 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* Scylla is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with Scylla. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "compaction_manager.hh"
|
|
#include "database.hh"
|
|
#include "core/scollectd.hh"
|
|
|
|
static logging::logger cmlog("compaction_manager");
|
|
|
|
void compaction_manager::task_start(lw_shared_ptr<compaction_manager::task>& task) {
|
|
// NOTE: Compaction code runs in parallel to the rest of the system.
|
|
// When it's time to shutdown, we need to prevent any new compaction
|
|
// from starting and wait for a possible ongoing compaction.
|
|
// That's possible by closing gate, busting semaphore and waiting for
|
|
// the future compaction_done to resolve.
|
|
|
|
task->compaction_done = keep_doing([this, task] {
|
|
return task->compaction_sem.wait().then([this, task] {
|
|
return seastar::with_gate(task->compaction_gate, [this, task] {
|
|
if (_cfs_to_compact.empty() && _cfs_to_cleanup.empty()) {
|
|
return make_ready_future<>();
|
|
}
|
|
|
|
if (!_cfs_to_cleanup.empty()) {
|
|
task->cleanup = true;
|
|
task->compacting_cf = _cfs_to_cleanup.front();
|
|
_cfs_to_cleanup.pop_front();
|
|
} else {
|
|
task->cleanup = false;
|
|
task->compacting_cf = _cfs_to_compact.front();
|
|
_cfs_to_compact.pop_front();
|
|
}
|
|
_stats.pending_tasks--;
|
|
_stats.active_tasks++;
|
|
|
|
column_family& cf = *task->compacting_cf;
|
|
std::unordered_set<unsigned long>& compacting_generations = cf.compacting_generations();
|
|
std::vector<sstables::shared_sstable> candidates; // candidates for compaction
|
|
|
|
candidates.reserve(cf.sstables_count());
|
|
// Filter out sstables that are being compacted.
|
|
for (auto& entry : *cf.get_sstables()) {
|
|
auto& sst = entry.second;
|
|
if (!compacting_generations.count(sst->generation())) {
|
|
candidates.push_back(sst);
|
|
}
|
|
}
|
|
|
|
sstables::compaction_descriptor descriptor;
|
|
// Created to erase generations from cf.compacting_generations() after compaction finishes.
|
|
std::vector<unsigned long> generations_to_compact;
|
|
generations_to_compact.reserve(descriptor.sstables.size());
|
|
|
|
auto insert_generations_to_compact = [&compacting_generations, &generations_to_compact, &descriptor] {
|
|
for (auto& sst : descriptor.sstables) {
|
|
auto generation = sst->generation();
|
|
generations_to_compact.push_back(generation);
|
|
compacting_generations.insert(generation);
|
|
}
|
|
};
|
|
|
|
future<> operation = make_ready_future<>();
|
|
if (task->cleanup) {
|
|
descriptor = sstables::compaction_descriptor(std::move(candidates));
|
|
insert_generations_to_compact();
|
|
operation = cf.cleanup_sstables(std::move(descriptor));
|
|
} else {
|
|
sstables::compaction_strategy cs = cf.get_compaction_strategy();
|
|
descriptor = cs.get_sstables_for_compaction(cf, std::move(candidates));
|
|
insert_generations_to_compact();
|
|
operation = cf.run_compaction(std::move(descriptor));
|
|
}
|
|
|
|
return operation.then([this, task] {
|
|
_stats.completed_tasks++;
|
|
// If compaction completed successfully, let's reset
|
|
// sleep time of compaction_retry.
|
|
task->compaction_retry.reset();
|
|
|
|
// The code below is to re-schedule the operation, so let's skip it
|
|
// if cleaning up.
|
|
if (task->cleanup) {
|
|
task->compacting_cf = nullptr;
|
|
return make_ready_future<>();
|
|
}
|
|
|
|
// Re-schedule compaction for compacting_cf, if needed.
|
|
if (!task->stopping && task->compacting_cf->pending_compactions()) {
|
|
// If there are pending compactions for compacting cf,
|
|
// push it into the back of the queue.
|
|
add_column_family(task->compacting_cf);
|
|
task->compaction_sem.signal();
|
|
} else {
|
|
// If so, cf is no longer queued by compaction manager.
|
|
task->compacting_cf->set_compaction_manager_queued(false);
|
|
}
|
|
task->compacting_cf = nullptr;
|
|
|
|
return make_ready_future<>();
|
|
}).finally([this, &compacting_generations, generations_to_compact = std::move(generations_to_compact)] {
|
|
// Remove compacted generations from the set of compacting generations,
|
|
// stored in column family.
|
|
for (auto generation : generations_to_compact) {
|
|
compacting_generations.erase(generation);
|
|
}
|
|
_stats.active_tasks--;
|
|
});
|
|
});
|
|
}).then_wrapped([this, task] (future<> f) {
|
|
bool retry = false;
|
|
|
|
// seastar::gate_closed_exception is used for regular termination
|
|
// of the fiber.
|
|
try {
|
|
f.get();
|
|
} catch (seastar::gate_closed_exception& e) {
|
|
cmlog.info("compaction task handler stopped due to shutdown");
|
|
throw;
|
|
} catch (std::exception& e) {
|
|
cmlog.error("compaction failed: {}", e.what());
|
|
retry = true;
|
|
} catch (...) {
|
|
cmlog.error("compaction failed: unknown error");
|
|
retry = true;
|
|
}
|
|
|
|
// We shouldn't retry compaction if task was asked to stop.
|
|
if (!task->stopping && retry) {
|
|
cmlog.info("compaction task handler sleeping for {} seconds",
|
|
std::chrono::duration_cast<std::chrono::seconds>(task->compaction_retry.sleep_time()).count());
|
|
return task->compaction_retry.retry().then([this, task] {
|
|
if (!task->compacting_cf) {
|
|
return make_ready_future<>();
|
|
}
|
|
// pushing cf to the back, so if the error is persistent,
|
|
// at least the others get a chance.
|
|
if (task->cleanup) {
|
|
_cfs_to_cleanup.push_back(task->compacting_cf);
|
|
_stats.pending_tasks++;
|
|
} else {
|
|
add_column_family(task->compacting_cf);
|
|
}
|
|
task->compacting_cf = nullptr;
|
|
|
|
// after sleeping, signal semaphore for the next compaction attempt.
|
|
task->compaction_sem.signal();
|
|
return make_ready_future<>();
|
|
});
|
|
}
|
|
return make_ready_future<>();
|
|
});
|
|
}).then_wrapped([] (future<> f) {
|
|
try {
|
|
f.get();
|
|
} catch (seastar::gate_closed_exception& e) {
|
|
// exception logged in keep_doing.
|
|
} catch (...) {
|
|
// this shouldn't happen, let's log it anyway.
|
|
cmlog.error("compaction task: unexpected error");
|
|
}
|
|
});
|
|
}
|
|
|
|
future<> compaction_manager::task_stop(lw_shared_ptr<compaction_manager::task>& task) {
|
|
task->stopping = true;
|
|
return task->compaction_gate.close().then([task] {
|
|
// NOTE: Signalling semaphore because we want task to finish with the
|
|
// gate_closed_exception exception.
|
|
task->compaction_sem.signal();
|
|
return task->compaction_done.then([task] {
|
|
task->compaction_gate = seastar::gate();
|
|
task->stopping = false;
|
|
return make_ready_future<>();
|
|
});
|
|
});
|
|
}
|
|
|
|
void compaction_manager::add_column_family(column_family* cf) {
|
|
_cfs_to_compact.push_back(cf);
|
|
_stats.pending_tasks++;
|
|
}
|
|
|
|
compaction_manager::compaction_manager() = default;
|
|
|
|
compaction_manager::~compaction_manager() {
|
|
// Assert that compaction manager was explicitly stopped, if started.
|
|
// Otherwise, fiber(s) will be alive after the object is destroyed.
|
|
assert(_stopped == true);
|
|
}
|
|
|
|
void compaction_manager::register_collectd_metrics() {
|
|
auto add = [this] (auto type_name, auto name, auto data_type, auto func) {
|
|
_registrations.push_back(
|
|
scollectd::add_polled_metric(scollectd::type_instance_id("compaction_manager",
|
|
scollectd::per_cpu_plugin_instance,
|
|
type_name, name),
|
|
scollectd::make_typed(data_type, func)));
|
|
};
|
|
|
|
add("objects", "compactions", scollectd::data_type::GAUGE, [&] { return _stats.active_tasks; });
|
|
}
|
|
|
|
void compaction_manager::start(int task_nr) {
|
|
_stopped = false;
|
|
_tasks.reserve(task_nr);
|
|
register_collectd_metrics();
|
|
for (int i = 0; i < task_nr; i++) {
|
|
auto task = make_lw_shared<compaction_manager::task>();
|
|
task_start(task);
|
|
_tasks.push_back(task);
|
|
}
|
|
}
|
|
|
|
future<> compaction_manager::stop() {
|
|
if (_stopped) {
|
|
return make_ready_future<>();
|
|
}
|
|
_stopped = true;
|
|
_registrations.clear();
|
|
return do_for_each(_tasks, [this] (auto& task) {
|
|
return this->task_stop(task);
|
|
}).then([this] {
|
|
for (auto& cf : _cfs_to_compact) {
|
|
cf->set_compaction_manager_queued(false);
|
|
}
|
|
_cfs_to_compact.clear();
|
|
_cfs_to_cleanup.clear();
|
|
return make_ready_future<>();
|
|
});
|
|
}
|
|
|
|
void compaction_manager::signal_less_busy_task() {
|
|
auto result = std::min_element(std::begin(_tasks), std::end(_tasks), [] (auto& i, auto& j) {
|
|
return i->compaction_sem.current() < j->compaction_sem.current();
|
|
});
|
|
(*result)->compaction_sem.signal();
|
|
}
|
|
|
|
bool compaction_manager::can_submit() {
|
|
return !_stopped && !_tasks.empty();
|
|
}
|
|
|
|
void compaction_manager::submit(column_family* cf) {
|
|
if (!can_submit()) {
|
|
return;
|
|
}
|
|
// To avoid having two or more entries of the same cf stored in the queue.
|
|
if (cf->compaction_manager_queued()) {
|
|
return;
|
|
}
|
|
cf->set_compaction_manager_queued(true);
|
|
add_column_family(cf);
|
|
signal_less_busy_task();
|
|
}
|
|
|
|
void compaction_manager::submit_cleanup_job(column_family* cf) {
|
|
if (!can_submit()) {
|
|
return;
|
|
}
|
|
// To avoid having two or more entries of the same cf stored in the queue.
|
|
if (std::find(_cfs_to_cleanup.begin(), _cfs_to_cleanup.end(), cf) != _cfs_to_cleanup.end()) {
|
|
return;
|
|
}
|
|
_cfs_to_cleanup.push_back(cf);
|
|
_stats.pending_tasks++;
|
|
signal_less_busy_task();
|
|
}
|
|
|
|
future<> compaction_manager::remove(column_family* cf) {
|
|
// Remove every reference to cf from _cfs_to_compact.
|
|
_cfs_to_compact.erase(
|
|
std::remove_if(_cfs_to_compact.begin(), _cfs_to_compact.end(), [cf] (column_family* entry) {
|
|
return cf == entry;
|
|
}),
|
|
_cfs_to_compact.end());
|
|
// Remove every reference to cf from _cfs_to_cleanup.
|
|
_cfs_to_cleanup.erase(
|
|
std::remove_if(_cfs_to_cleanup.begin(), _cfs_to_cleanup.end(), [cf] (column_family* entry) {
|
|
return cf == entry;
|
|
}),
|
|
_cfs_to_cleanup.end());
|
|
_stats.pending_tasks = _cfs_to_compact.size();
|
|
cf->set_compaction_manager_queued(false);
|
|
// We need to guarantee that a task being stopped will not re-queue the
|
|
// column family being removed.
|
|
auto tasks_to_stop = make_lw_shared<std::vector<lw_shared_ptr<task>>>();
|
|
for (auto& task : _tasks) {
|
|
if (task->compacting_cf == cf) {
|
|
tasks_to_stop->push_back(task);
|
|
task->stopping = true;
|
|
}
|
|
}
|
|
|
|
// Wait for the termination of an ongoing compaction on cf, if any.
|
|
return do_for_each(*tasks_to_stop, [this, cf] (auto& task) {
|
|
if (task->compacting_cf != cf) {
|
|
// There is no need to stop a task that is no longer compacting
|
|
// the column family being removed.
|
|
task->stopping = false;
|
|
return make_ready_future<>();
|
|
}
|
|
return this->task_stop(task).then([this, &task] {
|
|
// assert that task finished successfully.
|
|
assert(task->compacting_cf == nullptr);
|
|
this->task_start(task);
|
|
return make_ready_future<>();
|
|
});
|
|
}).then([tasks_to_stop] {});
|
|
}
|
|
|
|
void compaction_manager::stop_compaction(sstring type) {
|
|
// TODO: this method only works for compaction of type compaction and cleanup.
|
|
// Other types are: validation, scrub, index_build.
|
|
sstables::compaction_type target_type;
|
|
if (type == "COMPACTION") {
|
|
target_type = sstables::compaction_type::Compaction;
|
|
} else if (type == "CLEANUP") {
|
|
target_type = sstables::compaction_type::Cleanup;
|
|
} else {
|
|
throw std::runtime_error(sprint("Compaction of type %s cannot be stopped by compaction manager", type.c_str()));
|
|
}
|
|
for (auto& info : _compactions) {
|
|
if (target_type == info->type) {
|
|
info->stop();
|
|
}
|
|
}
|
|
}
|