mirror of
https://github.com/scylladb/scylladb.git
synced 2026-05-02 06:05:53 +00:00
This adds a reference to sl_controller so that, later in this patch series, topology_coordinator can manage creating `sl:driver` once group0 is fully operational. Refs: scylladb/scylladb#24411
103 lines
3.2 KiB
C++
103 lines
3.2 KiB
C++
/*
|
|
* Copyright (C) 2024-present ScyllaDB
|
|
*/
|
|
|
|
/*
|
|
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
|
|
*/
|
|
|
|
#pragma once
|
|
|
|
#include <chrono>
|
|
|
|
#include <seastar/core/future.hh>
|
|
#include <seastar/core/sharded.hh>
|
|
#include <stdexcept>
|
|
|
|
#include "utils/log.hh"
|
|
#include "raft/raft.hh"
|
|
#include "service/endpoint_lifecycle_subscriber.hh"
|
|
#include "service/topology_state_machine.hh"
|
|
#include "db/view/view_building_state.hh"
|
|
#include "service/qos/service_level_controller.hh"
|
|
|
|
namespace db {
|
|
class system_keyspace;
|
|
class system_distributed_keyspace;
|
|
}
|
|
|
|
namespace gms {
|
|
class gossiper;
|
|
class feature_service;
|
|
}
|
|
|
|
namespace netw {
|
|
class messaging_service;
|
|
}
|
|
|
|
namespace locator {
|
|
class shared_token_metadata;
|
|
}
|
|
|
|
namespace replica {
|
|
class database;
|
|
}
|
|
|
|
namespace raft {
|
|
class server;
|
|
}
|
|
|
|
namespace cdc {
|
|
class generation_service;
|
|
}
|
|
|
|
namespace service {
|
|
|
|
class raft_group0;
|
|
class tablet_allocator;
|
|
|
|
extern logging::logger rtlogger;
|
|
|
|
struct wait_for_ip_timeout : public std::runtime_error {
|
|
wait_for_ip_timeout(raft::server_id id, long timeout) :
|
|
std::runtime_error::runtime_error(format("failed to obtain an IP for {} in {}s", id, timeout)) {}
|
|
};
|
|
|
|
// The topology coordinator takes guard before operation start, but it releases it during various
|
|
// RPC commands that it sends to make it possible to submit new requests to the state machine while
|
|
// the coordinator drives current topology change. It is safe to do so since only the coordinator is
|
|
// ever allowed to change node's state, others may only create requests. To make sure the coordinator did
|
|
// not change while the lock was released, and hence the old coordinator does not work on old state, we check
|
|
// that the raft term is still the same after the lock is re-acquired. Throw term_changed_error if it did.
|
|
struct term_changed_error : public std::runtime_error {
|
|
term_changed_error() : std::runtime_error::runtime_error("Raft term changed.") {}
|
|
};
|
|
|
|
// Wait for the node with provided id to appear in the gossiper
|
|
future<> wait_for_gossiper(raft::server_id id, const gms::gossiper& g, seastar::abort_source& as);
|
|
|
|
using raft_topology_cmd_handler_type = noncopyable_function<future<raft_topology_cmd_result>(
|
|
raft::term_t, uint64_t, const raft_topology_cmd&)>;
|
|
|
|
struct topology_coordinator_cmd_rpc_tracker {
|
|
raft_topology_cmd::command current;
|
|
uint64_t index;
|
|
std::set<raft::server_id> active_dst;
|
|
};
|
|
|
|
future<> run_topology_coordinator(
|
|
seastar::sharded<db::system_distributed_keyspace>& sys_dist_ks, gms::gossiper& gossiper,
|
|
netw::messaging_service& messaging, locator::shared_token_metadata& shared_tm,
|
|
db::system_keyspace& sys_ks, replica::database& db, service::raft_group0& group0,
|
|
service::topology_state_machine& topo_sm, db::view::view_building_state_machine& vb_sm, seastar::abort_source& as, raft::server& raft,
|
|
raft_topology_cmd_handler_type raft_topology_cmd_handler,
|
|
tablet_allocator& tablet_allocator,
|
|
cdc::generation_service& cdc_gens,
|
|
std::chrono::milliseconds ring_delay,
|
|
endpoint_lifecycle_notifier& lifecycle_notifier,
|
|
gms::feature_service& feature_service,
|
|
qos::service_level_controller& sl_controller,
|
|
topology_coordinator_cmd_rpc_tracker& topology_cmd_rpc_tracker);
|
|
|
|
}
|