Files
scylladb/service/topology_coordinator.hh
Andrzej Jackowski 1ff605005e topology_coordinator: add service_level_controller reference
This adds a reference to sl_controller so that, later in this patch
series, topology_coordinator can manage creating `sl:driver` once
group0 is fully operational.

Refs: scylladb/scylladb#24411
2025-10-08 08:24:28 +02:00

103 lines
3.2 KiB
C++

/*
* Copyright (C) 2024-present ScyllaDB
*/
/*
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
*/
#pragma once
#include <chrono>
#include <seastar/core/future.hh>
#include <seastar/core/sharded.hh>
#include <stdexcept>
#include "utils/log.hh"
#include "raft/raft.hh"
#include "service/endpoint_lifecycle_subscriber.hh"
#include "service/topology_state_machine.hh"
#include "db/view/view_building_state.hh"
#include "service/qos/service_level_controller.hh"
namespace db {
class system_keyspace;
class system_distributed_keyspace;
}
namespace gms {
class gossiper;
class feature_service;
}
namespace netw {
class messaging_service;
}
namespace locator {
class shared_token_metadata;
}
namespace replica {
class database;
}
namespace raft {
class server;
}
namespace cdc {
class generation_service;
}
namespace service {
class raft_group0;
class tablet_allocator;
extern logging::logger rtlogger;
struct wait_for_ip_timeout : public std::runtime_error {
wait_for_ip_timeout(raft::server_id id, long timeout) :
std::runtime_error::runtime_error(format("failed to obtain an IP for {} in {}s", id, timeout)) {}
};
// The topology coordinator takes guard before operation start, but it releases it during various
// RPC commands that it sends to make it possible to submit new requests to the state machine while
// the coordinator drives current topology change. It is safe to do so since only the coordinator is
// ever allowed to change node's state, others may only create requests. To make sure the coordinator did
// not change while the lock was released, and hence the old coordinator does not work on old state, we check
// that the raft term is still the same after the lock is re-acquired. Throw term_changed_error if it did.
struct term_changed_error : public std::runtime_error {
term_changed_error() : std::runtime_error::runtime_error("Raft term changed.") {}
};
// Wait for the node with provided id to appear in the gossiper
future<> wait_for_gossiper(raft::server_id id, const gms::gossiper& g, seastar::abort_source& as);
using raft_topology_cmd_handler_type = noncopyable_function<future<raft_topology_cmd_result>(
raft::term_t, uint64_t, const raft_topology_cmd&)>;
struct topology_coordinator_cmd_rpc_tracker {
raft_topology_cmd::command current;
uint64_t index;
std::set<raft::server_id> active_dst;
};
future<> run_topology_coordinator(
seastar::sharded<db::system_distributed_keyspace>& sys_dist_ks, gms::gossiper& gossiper,
netw::messaging_service& messaging, locator::shared_token_metadata& shared_tm,
db::system_keyspace& sys_ks, replica::database& db, service::raft_group0& group0,
service::topology_state_machine& topo_sm, db::view::view_building_state_machine& vb_sm, seastar::abort_source& as, raft::server& raft,
raft_topology_cmd_handler_type raft_topology_cmd_handler,
tablet_allocator& tablet_allocator,
cdc::generation_service& cdc_gens,
std::chrono::milliseconds ring_delay,
endpoint_lifecycle_notifier& lifecycle_notifier,
gms::feature_service& feature_service,
qos::service_level_controller& sl_controller,
topology_coordinator_cmd_rpc_tracker& topology_cmd_rpc_tracker);
}