/* * * Modified by ScyllaDB * Copyright (C) 2015-present ScyllaDB */ /* * SPDX-License-Identifier: (LicenseRef-ScyllaDB-Source-Available-1.0 and Apache-2.0) */ #include "locator/token_metadata.hh" #include "utils/log.hh" #include "message/messaging_service.hh" #include #include "streaming/stream_session.hh" #include "streaming/prepare_message.hh" #include "streaming/stream_result_future.hh" #include "streaming/stream_manager.hh" #include "dht/auto_refreshing_sharder.hh" #include #include #include "streaming/stream_blob.hh" #include "streaming/stream_session_state.hh" #include "streaming/table_check.hh" #include "service/migration_manager.hh" #include "mutation_writer/multishard_writer.hh" #include "sstables/sstable_set.hh" #include "db/view/view_update_checks.hh" #include "replica/database.hh" #include "streaming/stream_mutation_fragments_cmd.hh" #include "consumer.hh" #include "readers/generating.hh" #include "service/topology_guard.hh" #include "utils/assert.hh" #include "utils/error_injection.hh" #include "db/config.hh" #include "idl/streaming.dist.hh" namespace streaming { logging::logger sslog("stream_session"); static sstables::offstrategy is_offstrategy_supported(streaming::stream_reason reason) { static const std::unordered_set operations_supported = { streaming::stream_reason::bootstrap, streaming::stream_reason::replace, streaming::stream_reason::removenode, streaming::stream_reason::decommission, streaming::stream_reason::repair, streaming::stream_reason::rebuild, }; return sstables::offstrategy(operations_supported.contains(reason)); } class offstrategy_trigger { sharded& _db; table_id _id; streaming::plan_id _plan_id; replica::column_family& _cf; lowres_clock::time_point last_update; public: offstrategy_trigger(sharded& db, table_id id, streaming::plan_id plan_id) : _db(db) , _id(id) , _plan_id(plan_id) , _cf(_db.local().find_column_family(_id)) { _cf.enable_off_strategy_trigger(); } void update() { auto now = lowres_clock::now(); // Call update_off_strategy_trigger at most every 30s. In the worst // case, we would shorten the offstrategy trigger timer by 10%. The // reward is that we now batch thousands or even millions of calls to // update_off_strategy_trigger. The update() is called for each and // every mutation fragment we received. So it is worth the // optimization. if (now - last_update > std::chrono::seconds(30)) { sslog.debug("[Stream #{}] Updated offstrategy trigger for ks={}, table={}, table_id={}", _plan_id, _cf.schema()->ks_name(), _cf.schema()->cf_name(), _id); _cf.update_off_strategy_trigger(); last_update = now; } } }; mutation_reader_consumer stream_manager::make_streaming_consumer(uint64_t estimated_partitions, stream_reason reason, service::frozen_topology_guard topo_guard) { return streaming::make_streaming_consumer("streaming", _db, _view_builder, _view_building_worker, estimated_partitions, reason, is_offstrategy_supported(reason), topo_guard); } void stream_manager::init_messaging_service_handler(abort_source& as) { auto& ms = _ms.local(); ser::streaming_rpc_verbs::register_prepare_message(&ms, [this] (const rpc::client_info& cinfo, prepare_message msg, streaming::plan_id plan_id, sstring description, rpc::optional reason_opt, rpc::optional session) { const auto& src_cpu_id = cinfo.retrieve_auxiliary("src_cpu_id"); const auto& from = cinfo.retrieve_auxiliary("host_id"); auto dst_cpu_id = this_shard_id(); auto reason = reason_opt ? *reason_opt : stream_reason::unspecified; auto topo_guard = service::frozen_topology_guard(session.value_or(service::default_session_id)); return container().invoke_on(dst_cpu_id, [msg = std::move(msg), plan_id, description = std::move(description), from, src_cpu_id, reason, topo_guard] (auto& sm) mutable { auto sr = stream_result_future::init_receiving_side(sm, plan_id, description, from); auto session = sm.get_session(plan_id, from, "PREPARE_MESSAGE"); session->init(sr); session->dst_cpu_id = src_cpu_id; session->set_reason(reason); session->set_topo_guard(topo_guard); return session->prepare(std::move(msg.requests), std::move(msg.summaries)); }); }); ser::streaming_rpc_verbs::register_prepare_done_message(&ms, [this] (const rpc::client_info& cinfo, streaming::plan_id plan_id, unsigned dst_cpu_id) { const auto& from = cinfo.retrieve_auxiliary("host_id"); return container().invoke_on(dst_cpu_id, [plan_id, from] (auto& sm) mutable { auto session = sm.get_session(plan_id, from, "PREPARE_DONE_MESSAGE"); session->follower_start_sent(); return make_ready_future<>(); }); }); ms.register_stream_mutation_fragments([this, &as] (const rpc::client_info& cinfo, streaming::plan_id plan_id, table_schema_version schema_id, table_id cf_id, uint64_t estimated_partitions, rpc::optional reason_opt, rpc::source> source, rpc::optional session) { auto from = cinfo.retrieve_auxiliary("host_id"); auto cpu_id = cinfo.retrieve_auxiliary("src_cpu_id"); auto src = cinfo.retrieve_auxiliary("host_id"); auto reason = reason_opt ? *reason_opt: stream_reason::unspecified; service::frozen_topology_guard topo_guard = session.value_or(service::default_session_id); sslog.trace("Got stream_mutation_fragments from {} reason {}, session {}", from, int(reason), session); return _mm.local().get_schema_for_write(schema_id, src, cpu_id, _ms.local(), as).then([this, from, estimated_partitions, plan_id, cf_id, source, reason, topo_guard, &as] (schema_ptr s) mutable { auto permit = _db.local().get_reader_concurrency_semaphore().make_tracking_only_permit(s, "stream-session", db::no_timeout, {}); struct stream_mutation_fragments_cmd_status { bool got_cmd = false; bool got_end_of_stream = false; }; auto cmd_status = make_lw_shared(); auto offstrategy_update = make_lw_shared(_db, cf_id, plan_id); auto guard = service::topology_guard(topo_guard); // Will log a message when streaming is done. Used to synchronize tests. lw_shared_ptr log_done; if (utils::get_local_injector().is_enabled("stream_mutation_fragments")) { log_done = make_lw_shared(seastar::make_shared(seastar::defer([] { sslog.info("stream_mutation_fragments: done"); }))); } auto get_next_mutation_fragment = [guard = std::move(guard), &as, &sm = container(), source, plan_id, from, s, cmd_status, offstrategy_update, permit] () mutable { guard.check(); return source().then([&sm, &guard, &as, plan_id, from, s, cmd_status, offstrategy_update, permit] (std::optional>> opt) mutable { if (opt) { auto cmd = std::get<1>(*opt); if (cmd) { cmd_status->got_cmd = true; switch (*cmd) { case stream_mutation_fragments_cmd::mutation_fragment_data: break; case stream_mutation_fragments_cmd::error: return make_exception_future(std::runtime_error("Sender failed")); case stream_mutation_fragments_cmd::abort: return make_exception_future(rpc::canceled_error()); case stream_mutation_fragments_cmd::end_of_stream: cmd_status->got_end_of_stream = true; return make_ready_future(); default: return make_exception_future(std::runtime_error("Sender sent wrong cmd")); } } frozen_mutation_fragment& fmf = std::get<0>(*opt); auto sz = fmf.representation().size(); auto mf = fmf.unfreeze(*s, permit); sm.local().update_progress(plan_id, from, progress_info::direction::IN, sz); offstrategy_update->update(); return utils::get_local_injector().inject("stream_mutation_fragments", [&guard, &as] (auto& handler) -> future<> { auto& guard_ = guard; auto& as_ = as; sslog.info("stream_mutation_fragments: waiting"); while (!handler.poll_for_message()) { guard_.check(); co_await sleep_abortable(std::chrono::milliseconds(5), as_); } sslog.info("stream_mutation_fragments: released"); }).then([mf = std::move(mf)] () mutable { if (utils::get_local_injector().is_enabled("stream_mutation_fragments_rx_error")) { sslog.info("stream_mutation_fragments_rx_error: throw"); throw std::runtime_error("stream_mutation_fragments_rx_error"); } return mutation_fragment_opt(std::move(mf)); }); } else { // If the sender has sent stream_mutation_fragments_cmd it means it is // a node that understands the new protocol. It must send end_of_stream // before close the stream. if (cmd_status->got_cmd && !cmd_status->got_end_of_stream) { return make_exception_future(std::runtime_error("Sender did not sent end_of_stream")); } return make_ready_future(); } }); }; auto sink = _ms.local().make_sink_for_stream_mutation_fragments(source); try { // Make sure the table with cf_id is still present at this point. // Close the sink in case the table is dropped. auto& table = _db.local().find_column_family(cf_id); utils::get_local_injector().inject("stream_mutation_fragments_table_dropped", [this] () { _db.local().find_column_family(table_id::create_null_id()); }); auto op = table.stream_in_progress(); auto sharder_ptr = std::make_unique(table.shared_from_this()); auto& sharder = *sharder_ptr; auto result_handling_cont = [this, s, plan_id, from, sink, estimated_partitions, log_done, cf_id, sh_ptr = std::move(sharder_ptr)] (future f) mutable -> future<> { auto& db = _db; auto& mm = _mm; int32_t status = 0; uint64_t received_partitions = 0; if (f.failed()) { auto ex = f.get_exception(); auto level = seastar::log_level::error; if (try_catch(ex)) { level = seastar::log_level::debug; } bool aborted = false; if (auto nested = try_catch(ex)) { if (try_catch(nested->inner)) { level = seastar::log_level::info; aborted = true; } } status = -1; // The status code -2 means error and the table is dropped if (db.local_is_initialized() && mm.local_is_initialized() && co_await table_sync_and_check(db.local(), mm.local(), cf_id)) { level = seastar::log_level::debug; status = -2; } sslog.log(level, "[Stream #{}] Failed to handle STREAM_MUTATION_FRAGMENTS (receive and distribute phase) for ks={}, cf={}, peer={}: {}", plan_id, s->ks_name(), s->cf_name(), from, aborted?"Streaming aborted":format("{}",ex)); } else { received_partitions = f.get(); } if (received_partitions) { sslog.info("[Stream #{}] Write to sstable for ks={}, cf={}, estimated_partitions={}, received_partitions={}", plan_id, s->ks_name(), s->cf_name(), estimated_partitions, received_partitions); } if (status == -1) { try { if (!utils::get_local_injector().is_enabled("stream_mutation_fragments_skip_fail_stream_plan")) { co_await fail_stream_plan(plan_id); } } catch (...) { sslog.warn("[Stream #{}] Failed to abort the stream plan: {}", plan_id, std::current_exception()); } } co_await sink(status).finally([sink] () mutable { return sink.close(); }); }; //FIXME: discarded future. (void)do_with(std::move(result_handling_cont), [&] (auto& result_handling) { return mutation_writer::distribute_reader_and_consume_on_shards(s, sharder, make_generating_reader_v1(s, permit, std::move(get_next_mutation_fragment)), make_streaming_consumer(estimated_partitions, reason, topo_guard), std::move(op) ).then_wrapped(std::ref(result_handling)).handle_exception([s, plan_id, from, sink] (std::exception_ptr ep) mutable -> future<> { auto level = seastar::log_level::error; if (try_catch(ep)) { level = seastar::log_level::debug; } sslog.log(level, "[Stream #{}] Failed to handle STREAM_MUTATION_FRAGMENTS (respond phase) for ks={}, cf={}, peer={}: {}", plan_id, s->ks_name(), s->cf_name(), from, ep); co_await sink.close(); }); }); } catch (...) { return sink.close().then([sink, eptr = std::current_exception()] () -> future> { return make_exception_future>(eptr); }); } return make_ready_future>(sink); }); }); ser::streaming_rpc_verbs::register_stream_mutation_done(&ms, [this] (const rpc::client_info& cinfo, streaming::plan_id plan_id, dht::token_range_vector ranges, table_id cf_id, unsigned dst_cpu_id) { const auto& from = cinfo.retrieve_auxiliary("host_id"); return container().invoke_on(dst_cpu_id, [ranges = std::move(ranges), plan_id, cf_id, from] (auto& sm) mutable { auto session = sm.get_session(plan_id, from, "STREAM_MUTATION_DONE", cf_id); session->receive_task_completed(cf_id); }); }); ser::streaming_rpc_verbs::register_complete_message(&ms, [this] (const rpc::client_info& cinfo, streaming::plan_id plan_id, unsigned dst_cpu_id, rpc::optional failed) { const auto& from = cinfo.retrieve_auxiliary("host_id"); if (failed && *failed) { return container().invoke_on(dst_cpu_id, [plan_id, from, dst_cpu_id] (auto& sm) { auto session = sm.get_session(plan_id, from, "COMPLETE_MESSAGE"); sslog.debug("[Stream #{}] COMPLETE_MESSAGE with error flag from {} dst_cpu_id={}", plan_id, from, dst_cpu_id); session->received_failed_complete_message(); return make_ready_future<>(); }); } else { // Be compatible with old version. Do nothing but return a ready future. sslog.debug("[Stream #{}] COMPLETE_MESSAGE from {} dst_cpu_id={}", plan_id, from, dst_cpu_id); return make_ready_future<>(); } }); ms.register_stream_blob([this] (const rpc::client_info& cinfo, streaming::stream_blob_meta meta, rpc::source source) { const auto& from = cinfo.retrieve_auxiliary("host_id"); auto sink = _ms.local().make_sink_for_stream_blob(source); (void)stream_blob_handler(_db.local(), _view_building_worker.local(), _ms.local(), from, meta, sink, source).handle_exception([ms = _ms.local().shared_from_this()] (std::exception_ptr eptr) { sslog.warn("Failed to run stream blob handler: {}", eptr); }); return make_ready_future>(sink); }); } future<> stream_manager::uninit_messaging_service_handler() { auto& ms = _ms.local(); return when_all_succeed( ser::streaming_rpc_verbs::unregister(&ms), ms.unregister_stream_blob(), ms.unregister_stream_mutation_fragments()).discard_result(); } stream_session::stream_session(stream_manager& mgr, locator::host_id peer_) : peer(peer_) , _mgr(mgr) { //this.metrics = StreamingMetrics.get(connecting); } stream_session::~stream_session() = default; future<> stream_session::on_initialization_complete() { // send prepare message set_state(stream_session_state::PREPARING); auto prepare = prepare_message(); std::copy(_requests.begin(), _requests.end(), std::back_inserter(prepare.requests)); for (auto& x : _transfers) { prepare.summaries.emplace_back(x.second.get_summary()); } auto id = this->peer; sslog.debug("[Stream #{}] SEND PREPARE_MESSAGE to {}", plan_id(), id); return ser::streaming_rpc_verbs::send_prepare_message(&manager().ms(), id, std::move(prepare), plan_id(), description(), get_reason(), topo_guard()).then_wrapped([this, id] (auto&& f) { try { auto msg = f.get(); sslog.debug("[Stream #{}] GOT PREPARE_MESSAGE Reply from {}", this->plan_id(), this->peer); this->dst_cpu_id = msg.dst_cpu_id; for (auto& summary : msg.summaries) { this->prepare_receiving(summary); } if (_stream_result) { _stream_result->handle_session_prepared(this->shared_from_this()); } } catch (...) { sslog.warn("[Stream #{}] Fail to send PREPARE_MESSAGE to {}, {}", this->plan_id(), id, std::current_exception()); throw; } return make_ready_future<>(); }).then([this, id] { auto plan_id = this->plan_id(); sslog.debug("[Stream #{}] SEND PREPARE_DONE_MESSAGE to {}", plan_id, id); return ser::streaming_rpc_verbs::send_prepare_done_message(&manager().ms(), id, plan_id, this->dst_cpu_id).then([this] { sslog.debug("[Stream #{}] GOT PREPARE_DONE_MESSAGE Reply from {}", this->plan_id(), this->peer); }).handle_exception([id, plan_id] (auto ep) { sslog.warn("[Stream #{}] Fail to send PREPARE_DONE_MESSAGE to {}, {}", plan_id, id, ep); std::rethrow_exception(ep); }); }).then([this] { sslog.debug("[Stream #{}] Initiator starts to sent", this->plan_id()); this->start_streaming_files(); }); } void stream_session::received_failed_complete_message() { if (utils::get_local_injector().is_enabled("stream_session_ignore_failed_message")) { sslog.info("[Stream #{}] Ignored failed complete message, peer={}", plan_id(), peer); return; } sslog.info("[Stream #{}] Received failed complete message, peer={}", plan_id(), peer); _received_failed_complete_message = true; close_session(stream_session_state::FAILED); } void stream_session::abort() { if (sslog.is_enabled(logging::log_level::debug)) { sslog.debug("[Stream #{}] Aborted stream session={}, peer={}, is_initialized={}", plan_id(), fmt::ptr(this), peer, is_initialized()); } else { sslog.info("[Stream #{}] Aborted stream session, peer={}, is_initialized={}", plan_id(), peer, is_initialized()); } close_session(stream_session_state::FAILED); } void stream_session::on_error() { sslog.warn("[Stream #{}] Streaming error occurred, peer={}", plan_id(), peer); close_session(stream_session_state::FAILED); } // Only follower calls this function upon receiving of prepare_message from initiator future stream_session::prepare(std::vector requests, std::vector summaries) { auto plan_id = this->plan_id(); auto nr_requests = requests.size(); sslog.debug("[Stream #{}] prepare requests nr={}, summaries nr={}", plan_id, nr_requests, summaries.size()); // prepare tasks set_state(stream_session_state::PREPARING); for (auto& request : requests) { // always flush on stream request sslog.debug("[Stream #{}] prepare stream_request={}", plan_id, request); add_transfer_ranges(std::move(request.keyspace), std::move(request.ranges), std::move(request.column_families)); co_await coroutine::maybe_yield(); } for (auto& summary : summaries) { sslog.debug("[Stream #{}] prepare stream_summary={}", plan_id, summary); prepare_receiving(summary); } // Always send a prepare_message back to follower prepare_message prepare; if (nr_requests) { for (auto& x: _transfers) { auto& task = x.second; prepare.summaries.emplace_back(task.get_summary()); } } prepare.dst_cpu_id = this_shard_id(); if (_stream_result) { _stream_result->handle_session_prepared(shared_from_this()); } co_return prepare; } void stream_session::follower_start_sent() { sslog.debug("[Stream #{}] Follower start to sent", this->plan_id()); this->start_streaming_files(); } session_info stream_session::make_session_info() { std::vector receiving_summaries; for (auto& receiver : _receivers) { receiving_summaries.emplace_back(receiver.second.get_summary()); } std::vector transfer_summaries; for (auto& transfer : _transfers) { transfer_summaries.emplace_back(transfer.second.get_summary()); } return session_info(peer, std::move(receiving_summaries), std::move(transfer_summaries), _state); } void stream_session::receive_task_completed(table_id cf_id) { _receivers.erase(cf_id); sslog.debug("[Stream #{}] receive task_completed: cf_id={} done, stream_receive_task.size={} stream_transfer_task.size={}", plan_id(), cf_id, _receivers.size(), _transfers.size()); maybe_completed(); } void stream_session::transfer_task_completed(table_id cf_id) { _transfers.erase(cf_id); sslog.debug("[Stream #{}] transfer task_completed: cf_id={} done, stream_receive_task.size={} stream_transfer_task.size={}", plan_id(), cf_id, _receivers.size(), _transfers.size()); maybe_completed(); } void stream_session::transfer_task_completed_all() { _transfers.clear(); sslog.debug("[Stream #{}] transfer task_completed: all done, stream_receive_task.size={} stream_transfer_task.size={}", plan_id(), _receivers.size(), _transfers.size()); maybe_completed(); } void stream_session::send_failed_complete_message() { if (!is_initialized()) { return; } auto plan_id = this->plan_id(); if (_received_failed_complete_message) { sslog.debug("[Stream #{}] Skip sending failed message back to peer", plan_id); return; } if (!_complete_sent) { _complete_sent = true; } else { return; } sslog.debug("[Stream #{}] SEND COMPLETE_MESSAGE to {}", plan_id, peer); auto session = shared_from_this(); bool failed = true; //FIXME: discarded future. (void)ser::streaming_rpc_verbs::send_complete_message(&manager().ms(), peer, plan_id, dst_cpu_id, failed).then([session, peer = this->peer, plan_id] { sslog.debug("[Stream #{}] GOT COMPLETE_MESSAGE Reply from {}", plan_id, peer); }).handle_exception([session, peer = this->peer, plan_id] (auto ep) { sslog.debug("[Stream #{}] COMPLETE_MESSAGE for {} has failed: {}", plan_id, peer, ep); }); } bool stream_session::maybe_completed() { bool completed = _receivers.empty() && _transfers.empty(); if (completed) { sslog.debug("[Stream #{}] maybe_completed: {} -> COMPLETE: session={}, peer={}", plan_id(), _state, fmt::ptr(this), peer); close_session(stream_session_state::COMPLETE); } return completed; } void stream_session::prepare_receiving(stream_summary& summary) { if (summary.files > 0) { // FIXME: handle when cf_id already exists _receivers.emplace(summary.cf_id, stream_receive_task(shared_from_this(), summary.cf_id, summary.files, summary.total_size)); } } void stream_session::start_streaming_files() { sslog.debug("[Stream #{}] {}: {} transfers to send", plan_id(), __func__, _transfers.size()); if (!_transfers.empty()) { set_state(stream_session_state::STREAMING); } //FIXME: discarded future. (void)do_for_each(_transfers.begin(), _transfers.end(), [this] (auto& item) { sslog.debug("[Stream #{}] Start to send cf_id={}", this->plan_id(), item.first); return item.second.execute(); }).then([this] { this->transfer_task_completed_all(); }).handle_exception([this] (auto ep) { sslog.warn("[Stream #{}] Failed to send: {}", this->plan_id(), ep); this->on_error(); }); } std::vector stream_session::get_column_family_stores(const sstring& keyspace, const std::vector& column_families) { // if columnfamilies are not specified, we add all cf under the keyspace std::vector stores; auto& db = manager().db(); if (column_families.empty()) { db.get_tables_metadata().for_each_table([&] (table_id, lw_shared_ptr tp) { replica::column_family& cf = *tp; auto cf_name = cf.schema()->cf_name(); auto ks_name = cf.schema()->ks_name(); if (ks_name == keyspace) { sslog.debug("Find ks={} cf={}", ks_name, cf_name); stores.push_back(&cf); } }); } else { // TODO: We can move this to database class and use shared_ptr instead for (auto& cf_name : column_families) { try { auto& x = db.find_column_family(keyspace, cf_name); stores.push_back(&x); } catch (replica::no_such_column_family&) { sslog.warn("stream_session: {}.{} does not exist: {}\n", keyspace, cf_name, std::current_exception()); continue; } } } return stores; } void stream_session::add_transfer_ranges(sstring keyspace, dht::token_range_vector ranges, std::vector column_families) { auto cfs = get_column_family_stores(keyspace, column_families); for (auto& cf : cfs) { auto cf_id = cf->schema()->id(); auto it = _transfers.find(cf_id); if (it == _transfers.end()) { stream_transfer_task task(shared_from_this(), cf_id, ranges); auto inserted = _transfers.emplace(cf_id, std::move(task)).second; SCYLLA_ASSERT(inserted); } else { it->second.append_ranges(ranges); } } } future<> stream_session::receiving_failed(table_id cf_id) { return make_ready_future<>(); } void stream_session::close_session(stream_session_state final_state) { sslog.debug("[Stream #{}] close_session session={}, state={}, is_aborted={}", plan_id(), fmt::ptr(this), final_state, _is_aborted); if (!_is_aborted) { _is_aborted = true; set_state(final_state); if (final_state == stream_session_state::FAILED) { for (auto& x : _transfers) { stream_transfer_task& task = x.second; sslog.debug("[Stream #{}] close_session session={}, state={}, abort stream_transfer_task cf_id={}", plan_id(), fmt::ptr(this), final_state, task.cf_id); task.abort(); } for (auto& x : _receivers) { stream_receive_task& task = x.second; sslog.debug("[Stream #{}] close_session session={}, state={}, abort stream_receive_task cf_id={}", plan_id(), fmt::ptr(this), final_state, task.cf_id); //FIXME: discarded future. (void)receiving_failed(x.first); task.abort(); } send_failed_complete_message(); } // Note that we shouldn't block on this close because this method is called on the handler // incoming thread (so we would deadlock). //handler.close(); if (_stream_result) { _stream_result->handle_session_complete(shared_from_this()); } sslog.debug("[Stream #{}] close_session session={}, state={}", plan_id(), fmt::ptr(this), final_state); } } void stream_session::start() { if (_requests.empty() && _transfers.empty()) { sslog.info("[Stream #{}] Session does not have any tasks.", plan_id()); close_session(stream_session_state::COMPLETE); return; } sslog.debug("[Stream #{}] Starting streaming to {}", plan_id(), peer); //FIXME: discarded future. (void)on_initialization_complete().handle_exception([this] (auto ep) { this->on_error(); }); } bool stream_session::is_initialized() const { return bool(_stream_result); } void stream_session::init(shared_ptr stream_result_) { _stream_result = stream_result_; } streaming::plan_id stream_session::plan_id() const { return _stream_result ? _stream_result->plan_id : streaming::plan_id::create_null_id(); } sstring stream_session::description() const { return _stream_result ? _stream_result->description : ""; } future<> stream_session::update_progress() { return manager().get_progress_on_all_shards(plan_id(), peer).then([this] (auto sbytes) { auto bytes_sent = sbytes.bytes_sent; if (bytes_sent > 0) { auto tx = progress_info(this->peer, "txnofile", progress_info::direction::OUT, bytes_sent, bytes_sent); _session_info.update_progress(std::move(tx)); } auto bytes_received = sbytes.bytes_received; if (bytes_received > 0) { auto rx = progress_info(this->peer, "rxnofile", progress_info::direction::IN, bytes_received, bytes_received); _session_info.update_progress(std::move(rx)); } }); } } // namespace streaming