/* * Copyright (C) 2015 ScyllaDB */ /* * This file is part of Scylla. * * Scylla is free software: you can redistribute it and/or modify * it under the terms of the GNU Affero General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * Scylla is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with Scylla. If not, see . */ #include "server.hh" #include #include #include #include #include #include "cql3/statements/batch_statement.hh" #include "types/collection.hh" #include "types/list.hh" #include "types/set.hh" #include "types/map.hh" #include "dht/token-sharding.hh" #include "service/migration_manager.hh" #include "service/storage_service.hh" #include "db/consistency_level_type.hh" #include "db/write_type.hh" #include #include #include "utils/UUID.hh" #include #include #include #include #include #include "enum_set.hh" #include "service/query_state.hh" #include "service/client_state.hh" #include "exceptions/exceptions.hh" #include "connection_notifier.hh" #include "auth/authenticator.hh" #include #include #include #include #include "response.hh" #include "request.hh" #include "types/user.hh" #include "transport/cql_protocol_extension.hh" namespace cql_transport { static logging::logger clogger("cql_server"); struct cql_frame_error : std::exception { const char* what() const throw () override { return "bad cql binary frame"; } }; inline int16_t consistency_to_wire(db::consistency_level c) { switch (c) { case db::consistency_level::ANY: return 0x0000; case db::consistency_level::ONE: return 0x0001; case db::consistency_level::TWO: return 0x0002; case db::consistency_level::THREE: return 0x0003; case db::consistency_level::QUORUM: return 0x0004; case db::consistency_level::ALL: return 0x0005; case db::consistency_level::LOCAL_QUORUM: return 0x0006; case db::consistency_level::EACH_QUORUM: return 0x0007; case db::consistency_level::SERIAL: return 0x0008; case db::consistency_level::LOCAL_SERIAL: return 0x0009; case db::consistency_level::LOCAL_ONE: return 0x000A; default: throw std::runtime_error("Invalid consistency level"); } } sstring to_string(const event::topology_change::change_type t) { using type = event::topology_change::change_type; switch (t) { case type::NEW_NODE: return "NEW_NODE"; case type::REMOVED_NODE: return "REMOVED_NODE"; case type::MOVED_NODE: return "MOVED_NODE"; } throw std::invalid_argument("unknown change type"); } sstring to_string(const event::status_change::status_type t) { using type = event::status_change::status_type; switch (t) { case type::UP: return "UP"; case type::DOWN: return "DOWN"; } throw std::invalid_argument("unknown change type"); } sstring to_string(const event::schema_change::change_type t) { switch (t) { case event::schema_change::change_type::CREATED: return "CREATED"; case event::schema_change::change_type::UPDATED: return "UPDATED"; case event::schema_change::change_type::DROPPED: return "DROPPED"; } assert(false && "unreachable"); } sstring to_string(const event::schema_change::target_type t) { switch (t) { case event::schema_change::target_type::KEYSPACE: return "KEYSPACE"; case event::schema_change::target_type::TABLE: return "TABLE"; case event::schema_change::target_type::TYPE: return "TYPE"; case event::schema_change::target_type::FUNCTION: return "FUNCTION"; case event::schema_change::target_type::AGGREGATE:return "AGGREGATE"; } assert(false && "unreachable"); } event::event_type parse_event_type(const sstring& value) { if (value == "TOPOLOGY_CHANGE") { return event::event_type::TOPOLOGY_CHANGE; } else if (value == "STATUS_CHANGE") { return event::event_type::STATUS_CHANGE; } else if (value == "SCHEMA_CHANGE") { return event::event_type::SCHEMA_CHANGE; } else { throw exceptions::protocol_exception(format("Invalid value '{}' for Event.Type", value)); } } cql_server::cql_server(distributed& qp, auth::service& auth_service, service::migration_notifier& mn, cql_server_config config) : _query_processor(qp) , _config(config) , _max_request_size(config.max_request_size) , _memory_available(config.get_service_memory_limiter_semaphore()) , _notifier(std::make_unique(mn)) , _auth_service(auth_service) { namespace sm = seastar::metrics; _metrics.add_group("transport", { sm::make_derive("cql-connections", _connects, sm::description("Counts a number of client connections.")), sm::make_gauge("current_connections", _connections, sm::description("Holds a current number of client connections.")), sm::make_derive("requests_served", _requests_served, sm::description("Counts a number of served requests.")), sm::make_gauge("requests_serving", _requests_serving, sm::description("Holds a number of requests that are being processed right now.")), sm::make_gauge("requests_blocked_memory_current", [this] { return _memory_available.waiters(); }, sm::description( seastar::format("Holds the number of requests that are currently blocked due to reaching the memory quota limit ({}B). " "Non-zero value indicates that our bottleneck is memory and more specifically - the memory quota allocated for the \"CQL transport\" component.", _max_request_size))), sm::make_derive("requests_blocked_memory", _requests_blocked_memory, sm::description( seastar::format("Holds an incrementing counter with the requests that ever blocked due to reaching the memory quota limit ({}B). " "The first derivative of this value shows how often we block due to memory exhaustion in the \"CQL transport\" component.", _max_request_size))), sm::make_gauge("requests_memory_available", [this] { return _memory_available.current(); }, sm::description( seastar::format("Holds the amount of available memory for admitting new requests (max is {}B)." "Zero value indicates that our bottleneck is memory and more specifically - the memory quota allocated for the \"CQL transport\" component.", _max_request_size))), }); } future<> cql_server::stop() { _stopping = true; size_t nr = 0; size_t nr_total = _listeners.size(); clogger.debug("cql_server: abort accept nr_total={}", nr_total); for (auto&& l : _listeners) { l.abort_accept(); clogger.debug("cql_server: abort accept {} out of {} done", ++nr, nr_total); } auto nr_conn = make_lw_shared(0); auto nr_conn_total = _connections_list.size(); clogger.debug("cql_server: shutdown connection nr_total={}", nr_conn_total); return parallel_for_each(_connections_list.begin(), _connections_list.end(), [nr_conn, nr_conn_total] (auto&& c) { return c.shutdown().then([nr_conn, nr_conn_total] { clogger.debug("cql_server: shutdown connection {} out of {} done", ++(*nr_conn), nr_conn_total); }); }).then([this] { return _notifier->stop(); }).then([this] { return std::move(_stopped); }); } future<> cql_server::listen(socket_address addr, std::shared_ptr creds, bool is_shard_aware, bool keepalive) { auto f = make_ready_future>(nullptr); if (creds) { f = creds->build_reloadable_server_credentials([](const std::unordered_set& files, std::exception_ptr ep) { if (ep) { clogger.warn("Exception loading {}: {}", files, ep); } else { clogger.info("Reloaded {}", files); } }); } return f.then([this, addr, is_shard_aware, keepalive](shared_ptr creds) { listen_options lo; lo.reuse_address = true; if (is_shard_aware) { lo.lba = server_socket::load_balancing_algorithm::port; } server_socket ss; try { ss = creds ? seastar::tls::listen(std::move(creds), addr, lo) : seastar::listen(addr, lo); } catch (...) { throw std::runtime_error(format("CQLServer error while listening on {} -> {}", addr, std::current_exception())); } _listeners.emplace_back(std::move(ss)); _stopped = when_all(std::move(_stopped), do_accepts(_listeners.size() - 1, keepalive, addr)).discard_result(); }); } future<> cql_server::do_accepts(int which, bool keepalive, socket_address server_addr) { return repeat([this, which, keepalive, server_addr] { ++_connections_being_accepted; return _listeners[which].accept().then_wrapped([this, which, keepalive, server_addr] (future f_cs_sa) mutable { --_connections_being_accepted; if (_stopping) { f_cs_sa.ignore_ready_future(); maybe_idle(); return stop_iteration::yes; } auto cs_sa = f_cs_sa.get0(); auto fd = std::move(cs_sa.connection); auto addr = std::move(cs_sa.remote_address); fd.set_nodelay(true); fd.set_keepalive(keepalive); auto conn = make_shared(*this, server_addr, std::move(fd), std::move(addr)); ++_connects; ++_connections; // Move the processing into the background. (void)futurize_invoke([this, conn] { return advertise_new_connection(conn); // Notify any listeners about new connection. }).then_wrapped([this, conn] (future<> f) { try { f.get(); } catch (...) { clogger.info("exception while advertising new connection: {}", std::current_exception()); } // Block while monitoring for lifetime/errors. return conn->process().finally([this, conn] { --_connections; return unadvertise_connection(conn); }).handle_exception([] (std::exception_ptr ep) { try { std::rethrow_exception(ep); } catch(std::system_error& serr) { if (serr.code().category() == std::system_category() && serr.code().value() == EPIPE) { // expected if another side closes a connection return; } } catch(...) {}; clogger.info("exception while processing connection: {}", ep); }); }); return stop_iteration::no; }).handle_exception([] (auto ep) { clogger.debug("accept failed: {}", ep); return stop_iteration::no; }); }); } future<> cql_server::advertise_new_connection(shared_ptr conn) { client_data cd = conn->make_client_data(); clogger.trace("Advertising new connection from CQL client {}:{}", cd.ip, cd.port); return notify_new_client(std::move(cd)); } future<> cql_server::unadvertise_connection(shared_ptr conn) { const auto ip = conn->get_client_state().get_client_address().addr(); const auto port = conn->get_client_state().get_client_port(); clogger.trace("Advertising disconnection of CQL client {}:{}", ip, port); return notify_disconnected_client(ip, client_type::cql, port); } unsigned cql_server::connection::frame_size() const { if (_version < 3) { return 8; } else { return 9; } } cql_binary_frame_v3 cql_server::connection::parse_frame(temporary_buffer buf) const { if (buf.size() != frame_size()) { throw cql_frame_error(); } cql_binary_frame_v3 v3; switch (_version) { case 1: case 2: { auto raw = reinterpret_cast(buf.get()); auto cooked = net::ntoh(*raw); v3.version = cooked.version; v3.flags = cooked.flags; v3.opcode = cooked.opcode; v3.stream = cooked.stream; v3.length = cooked.length; break; } case 3: case 4: { v3 = net::ntoh(*reinterpret_cast(buf.get())); break; } default: throw exceptions::protocol_exception(format("Invalid or unsupported protocol version: {:d}", _version)); } if (v3.version != _version) { throw exceptions::protocol_exception(format("Invalid message version. Got {:d} but previous messages on this connection had version {:d}", v3.version, _version)); } return v3; } future> cql_server::connection::read_frame() { using ret_type = std::optional; if (!_version) { // We don't know the frame size before reading the first frame, // so read just one byte, and then read the rest of the frame. return _read_buf.read_exactly(1).then([this] (temporary_buffer buf) { if (buf.empty()) { return make_ready_future(); } _version = buf[0]; init_cql_serialization_format(); if (_version < 1 || _version > current_version) { auto client_version = _version; _version = current_version; throw exceptions::protocol_exception(format("Invalid or unsupported protocol version: {:d}", client_version)); } return _read_buf.read_exactly(frame_size() - 1).then([this] (temporary_buffer tail) { temporary_buffer full(frame_size()); full.get_write()[0] = _version; std::copy(tail.get(), tail.get() + tail.size(), full.get_write() + 1); auto frame = parse_frame(std::move(full)); // This is the very first frame, so reject obviously incorrect frames, to // avoid allocating large amounts of memory for the message body if (frame.length > 100'000) { // The STARTUP message body is a [string map] containing just a few options, // so it should be smaller that 100kB. See #4366. throw exceptions::protocol_exception(format("Initial message size too large ({:d}), rejecting as invalid", frame.length)); } return make_ready_future(frame); }); }); } else { // Not the first frame, so we know the size. return _read_buf.read_exactly(frame_size()).then([this] (temporary_buffer buf) { if (buf.empty()) { return make_ready_future(); } return make_ready_future(parse_frame(std::move(buf))); }); } } future>> cql_server::connection::process_request_one(fragmented_temporary_buffer::istream fbuf, uint8_t op, uint16_t stream, service::client_state& client_state, tracing_request_type tracing_request, service_permit permit) { using auth_state = service::client_state::auth_state; auto cqlop = static_cast(op); tracing::trace_state_props_set trace_props; trace_props.set_if(tracing::tracing::get_local_tracing_instance().slow_query_tracing_enabled()); trace_props.set_if(tracing_request != tracing_request_type::not_requested); tracing::trace_state_ptr trace_state; if (trace_props) { if (cqlop == cql_binary_opcode::QUERY || cqlop == cql_binary_opcode::PREPARE || cqlop == cql_binary_opcode::EXECUTE || cqlop == cql_binary_opcode::BATCH) { trace_props.set_if(tracing_request == tracing_request_type::write_on_close); trace_state = tracing::tracing::get_local_tracing_instance().create_session(tracing::trace_type::QUERY, trace_props); } } tracing::set_request_size(trace_state, fbuf.bytes_left()); auto linearization_buffer = std::make_unique(); auto linearization_buffer_ptr = linearization_buffer.get(); return futurize_invoke([this, cqlop, stream, &fbuf, &client_state, linearization_buffer_ptr, permit = std::move(permit), trace_state] () mutable { // When using authentication, we need to ensure we are doing proper state transitions, // i.e. we cannot simply accept any query/exec ops unless auth is complete switch (client_state.get_auth_state()) { case auth_state::UNINITIALIZED: if (cqlop != cql_binary_opcode::STARTUP && cqlop != cql_binary_opcode::OPTIONS) { throw exceptions::protocol_exception(format("Unexpected message {:d}, expecting STARTUP or OPTIONS", int(cqlop))); } break; case auth_state::AUTHENTICATION: // Support both SASL auth from protocol v2 and the older style Credentials auth from v1 if (cqlop != cql_binary_opcode::AUTH_RESPONSE && cqlop != cql_binary_opcode::CREDENTIALS) { throw exceptions::protocol_exception(format("Unexpected message {:d}, expecting {}", int(cqlop), _version == 1 ? "CREDENTIALS" : "SASL_RESPONSE")); } break; case auth_state::READY: default: if (cqlop == cql_binary_opcode::STARTUP) { throw exceptions::protocol_exception("Unexpected message STARTUP, the connection is already initialized"); } break; } tracing::set_username(trace_state, client_state.user()); auto wrap_in_foreign = [] (future> f) { return f.then([] (std::unique_ptr p) { return make_ready_future>>(make_foreign(std::move(p))); }); }; auto in = request_reader(std::move(fbuf), *linearization_buffer_ptr); switch (cqlop) { case cql_binary_opcode::STARTUP: return wrap_in_foreign(process_startup(stream, std::move(in), client_state, trace_state)); case cql_binary_opcode::AUTH_RESPONSE: return wrap_in_foreign(process_auth_response(stream, std::move(in), client_state, trace_state)); case cql_binary_opcode::OPTIONS: return wrap_in_foreign(process_options(stream, std::move(in), client_state, trace_state)); case cql_binary_opcode::QUERY: return process_query(stream, std::move(in), client_state, std::move(permit), trace_state); case cql_binary_opcode::PREPARE: return wrap_in_foreign(process_prepare(stream, std::move(in), client_state, trace_state)); case cql_binary_opcode::EXECUTE: return process_execute(stream, std::move(in), client_state, std::move(permit), trace_state); case cql_binary_opcode::BATCH: return process_batch(stream, std::move(in), client_state, std::move(permit), trace_state); case cql_binary_opcode::REGISTER: return wrap_in_foreign(process_register(stream, std::move(in), client_state, trace_state)); default: throw exceptions::protocol_exception(format("Unknown opcode {:d}", int(cqlop))); } }).then_wrapped([this, cqlop, stream, &client_state, linearization_buffer = std::move(linearization_buffer), trace_state] (future>> f) -> foreign_ptr> { auto stop_trace = defer([&] { tracing::stop_foreground(trace_state); }); --_server._requests_serving; try { foreign_ptr> response = f.get0(); auto res_op = response->opcode(); // and modify state now that we've generated a response. switch (client_state.get_auth_state()) { case auth_state::UNINITIALIZED: if (cqlop == cql_binary_opcode::STARTUP) { if (res_op == cql_binary_opcode::AUTHENTICATE) { client_state.set_auth_state(auth_state::AUTHENTICATION); } else if (res_op == cql_binary_opcode::READY) { client_state.set_auth_state(auth_state::READY); } } break; case auth_state::AUTHENTICATION: // Support both SASL auth from protocol v2 and the older style Credentials auth from v1 assert(cqlop == cql_binary_opcode::AUTH_RESPONSE || cqlop == cql_binary_opcode::CREDENTIALS); if (res_op == cql_binary_opcode::READY || res_op == cql_binary_opcode::AUTH_SUCCESS) { client_state.set_auth_state(auth_state::READY); } break; default: case auth_state::READY: break; } tracing::set_response_size(trace_state, response->size()); return response; } catch (const exceptions::unavailable_exception& ex) { return make_unavailable_error(stream, ex.code(), ex.what(), ex.consistency, ex.required, ex.alive, trace_state); } catch (const exceptions::read_timeout_exception& ex) { return make_read_timeout_error(stream, ex.code(), ex.what(), ex.consistency, ex.received, ex.block_for, ex.data_present, trace_state); } catch (const exceptions::read_failure_exception& ex) { return make_read_failure_error(stream, ex.code(), ex.what(), ex.consistency, ex.received, ex.failures, ex.block_for, ex.data_present, trace_state); } catch (const exceptions::mutation_write_timeout_exception& ex) { return make_mutation_write_timeout_error(stream, ex.code(), ex.what(), ex.consistency, ex.received, ex.block_for, ex.type, trace_state); } catch (const exceptions::mutation_write_failure_exception& ex) { return make_mutation_write_failure_error(stream, ex.code(), ex.what(), ex.consistency, ex.received, ex.failures, ex.block_for, ex.type, trace_state); } catch (const exceptions::already_exists_exception& ex) { return make_already_exists_error(stream, ex.code(), ex.what(), ex.ks_name, ex.cf_name, trace_state); } catch (const exceptions::prepared_query_not_found_exception& ex) { return make_unprepared_error(stream, ex.code(), ex.what(), ex.id, trace_state); } catch (const exceptions::cassandra_exception& ex) { return make_error(stream, ex.code(), ex.what(), trace_state); } catch (std::exception& ex) { return make_error(stream, exceptions::exception_code::SERVER_ERROR, ex.what(), trace_state); } catch (...) { return make_error(stream, exceptions::exception_code::SERVER_ERROR, "unknown error", trace_state); } }); } cql_server::connection::connection(cql_server& server, socket_address server_addr, connected_socket&& fd, socket_address addr) : _server(server) , _server_addr(server_addr) , _fd(std::move(fd)) , _read_buf(_fd.input()) , _write_buf(_fd.output()) , _client_state(service::client_state::external_tag{}, server._auth_service, addr) { ++_server._total_connections; ++_server._current_connections; _server._connections_list.push_back(*this); } cql_server::connection::~connection() { --_server._current_connections; _server._connections_list.erase(_server._connections_list.iterator_to(*this)); _server.maybe_idle(); } future<> cql_server::connection::process() { return with_gate(_pending_requests_gate, [this] { return do_until([this] { return _read_buf.eof(); }, [this] { return process_request(); }).then_wrapped([this] (future<> f) { try { f.get(); } catch (const exceptions::cassandra_exception& ex) { write_response(make_error(0, ex.code(), ex.what(), tracing::trace_state_ptr())); } catch (std::exception& ex) { write_response(make_error(0, exceptions::exception_code::SERVER_ERROR, ex.what(), tracing::trace_state_ptr())); } catch (...) { write_response(make_error(0, exceptions::exception_code::SERVER_ERROR, "unknown error", tracing::trace_state_ptr())); } }); }).finally([this] { return _pending_requests_gate.close().then([this] { _server._notifier->unregister_connection(this); return _ready_to_respond.finally([this] { return _write_buf.close(); }); }); }); } future<> cql_server::connection::shutdown() { try { _fd.shutdown_input(); _fd.shutdown_output(); } catch (...) { } return make_ready_future<>(); } client_data cql_server::connection::make_client_data() const { client_data cd; cd.ip = _client_state.get_client_address().addr(); cd.port = _client_state.get_client_port(); cd.ct = client_type::cql; cd.shard_id = this_shard_id(); cd.protocol_version = _version; if (const auto user_ptr = _client_state.user(); user_ptr) { cd.username = user_ptr->name; } return cd; } thread_local cql_server::connection::execution_stage_type cql_server::connection::_process_request_stage{"transport", &connection::process_request_one}; future<> cql_server::connection::process_request() { return read_frame().then_wrapped([this] (future>&& v) { auto maybe_frame = std::get<0>(v.get()); if (!maybe_frame) { // eof return make_ready_future<>(); } auto& f = *maybe_frame; tracing_request_type tracing_requested = tracing_request_type::not_requested; if (f.flags & cql_frame_flags::tracing) { // If tracing is requested for a specific CQL command - flush // tracing info right after the command is over. tracing_requested = tracing_request_type::write_on_close; } else if (tracing::tracing::get_local_tracing_instance().trace_next_query()) { tracing_requested = tracing_request_type::no_write_on_close; } auto op = f.opcode; auto stream = f.stream; auto mem_estimate = f.length * 2 + 8000; // Allow for extra copies and bookkeeping if (mem_estimate > _server._max_request_size) { throw exceptions::invalid_request_exception(format("request size too large (frame size {:d}; estimate {:d}; allowed {:d}", f.length, mem_estimate, _server._max_request_size)); } auto fut = get_units(_server._memory_available, mem_estimate); if (_server._memory_available.waiters()) { ++_server._requests_blocked_memory; } return fut.then([this, length = f.length, flags = f.flags, op, stream, tracing_requested] (semaphore_units<> mem_permit) { return this->read_and_decompress_frame(length, flags).then([this, op, stream, tracing_requested, mem_permit = make_service_permit(std::move(mem_permit))] (fragmented_temporary_buffer buf) mutable { ++_server._requests_served; ++_server._requests_serving; _pending_requests_gate.enter(); auto leave = defer([this] { _pending_requests_gate.leave(); }); // Replacing the immediately-invoked lambda below with just its body costs 5-10 usec extra per invocation. // Cause not understood. auto istream = buf.get_istream(); (void)_process_request_stage(this, istream, op, stream, seastar::ref(_client_state), tracing_requested, mem_permit) .then_wrapped([this, buf = std::move(buf), mem_permit, leave = std::move(leave)] (future>> response_f) mutable { try { write_response(std::move(response_f.get0()), std::move(mem_permit), _compression); _ready_to_respond = _ready_to_respond.finally([leave = std::move(leave)] {}); } catch (...) { clogger.error("request processing failed: {}", std::current_exception()); } }); return make_ready_future<>(); }); }); }); } static inline bytes_view to_bytes_view(temporary_buffer& b) { using byte = bytes_view::value_type; return bytes_view(reinterpret_cast(b.get()), b.size()); } namespace compression_buffers { // Reusable buffers for compression and decompression. Cleared every // clear_buffers_trigger uses. static constexpr size_t clear_buffers_trigger = 100'000; static thread_local size_t buffer_use_count = 0; static thread_local utils::reusable_buffer input_buffer; static thread_local utils::reusable_buffer output_buffer; void on_compression_buffer_use() { if (++buffer_use_count == clear_buffers_trigger) { input_buffer.clear(); output_buffer.clear(); buffer_use_count = 0; } } } future cql_server::connection::read_and_decompress_frame(size_t length, uint8_t flags) { using namespace compression_buffers; if (flags & cql_frame_flags::compression) { if (_compression == cql_compression::lz4) { if (length < 4) { throw std::runtime_error("Truncated frame"); } return _buffer_reader.read_exactly(_read_buf, length).then([this] (fragmented_temporary_buffer buf) { auto linearization_buffer = bytes_ostream(); int32_t uncomp_len = request_reader(buf.get_istream(), linearization_buffer).read_int(); if (uncomp_len < 0) { throw std::runtime_error("CQL frame uncompressed length is negative: " + std::to_string(uncomp_len)); } buf.remove_prefix(4); auto in = input_buffer.get_linearized_view(fragmented_temporary_buffer::view(buf)); auto uncomp = output_buffer.make_fragmented_temporary_buffer(uncomp_len, fragmented_temporary_buffer::default_fragment_size, [&] (bytes_mutable_view out) { auto ret = LZ4_decompress_safe(reinterpret_cast(in.data()), reinterpret_cast(out.data()), in.size(), out.size()); if (ret < 0) { throw std::runtime_error("CQL frame LZ4 uncompression failure"); } return out.size(); }); on_compression_buffer_use(); return uncomp; }); } else if (_compression == cql_compression::snappy) { return _buffer_reader.read_exactly(_read_buf, length).then([this] (fragmented_temporary_buffer buf) { auto in = input_buffer.get_linearized_view(fragmented_temporary_buffer::view(buf)); size_t uncomp_len; if (snappy_uncompressed_length(reinterpret_cast(in.data()), in.size(), &uncomp_len) != SNAPPY_OK) { throw std::runtime_error("CQL frame Snappy uncompressed size is unknown"); } auto uncomp = output_buffer.make_fragmented_temporary_buffer(uncomp_len, fragmented_temporary_buffer::default_fragment_size, [&] (bytes_mutable_view out) { size_t output_len = out.size(); if (snappy_uncompress(reinterpret_cast(in.data()), in.size(), reinterpret_cast(out.data()), &output_len) != SNAPPY_OK) { throw std::runtime_error("CQL frame Snappy uncompression failure"); } return output_len; }); on_compression_buffer_use(); return uncomp; }); } else { throw exceptions::protocol_exception(format("Unknown compression algorithm")); } } return _buffer_reader.read_exactly(_read_buf, length); } future> cql_server::connection::process_startup(uint16_t stream, request_reader in, service::client_state& client_state, tracing::trace_state_ptr trace_state) { auto options = in.read_string_map(); auto compression_opt = options.find("COMPRESSION"); if (compression_opt != options.end()) { auto compression = compression_opt->second; std::transform(compression.begin(), compression.end(), compression.begin(), ::tolower); if (compression == "lz4") { _compression = cql_compression::lz4; } else if (compression == "snappy") { _compression = cql_compression::snappy; } else { throw exceptions::protocol_exception(format("Unknown compression algorithm: {}", compression)); } } cql_protocol_extension_enum_set cql_proto_exts; for (cql_protocol_extension ext : supported_cql_protocol_extensions()) { if (options.find(protocol_extension_name(ext)) != options.cend()) { cql_proto_exts.set(ext); } } _client_state.set_protocol_extensions(std::move(cql_proto_exts)); auto& a = client_state.get_auth_service()->underlying_authenticator(); if (a.require_authentication()) { return make_ready_future>(make_autheticate(stream, a.qualified_java_name(), trace_state)); } return make_ready_future>(make_ready(stream, trace_state)); } future> cql_server::connection::process_auth_response(uint16_t stream, request_reader in, service::client_state& client_state, tracing::trace_state_ptr trace_state) { auto sasl_challenge = client_state.get_auth_service()->underlying_authenticator().new_sasl_challenge(); auto buf = in.read_raw_bytes_view(in.bytes_left()); auto challenge = sasl_challenge->evaluate_response(buf); if (sasl_challenge->is_complete()) { return sasl_challenge->get_authenticated_user().then([this, sasl_challenge, stream, &client_state, challenge = std::move(challenge), trace_state](auth::authenticated_user user) mutable { client_state.set_login(std::move(user)); auto f = client_state.check_user_can_login(); return f.then([this, stream, &client_state, challenge = std::move(challenge), trace_state]() mutable { return make_ready_future>(make_auth_success(stream, std::move(challenge), trace_state)); }); }); } return make_ready_future>(make_auth_challenge(stream, std::move(challenge), trace_state)); } future> cql_server::connection::process_options(uint16_t stream, request_reader in, service::client_state& client_state, tracing::trace_state_ptr trace_state) { return make_ready_future>(make_supported(stream, std::move(trace_state))); } void cql_server::connection::init_cql_serialization_format() { _cql_serialization_format = cql_serialization_format(_version); } std::unique_ptr make_result(int16_t stream, messages::result_message& msg, const tracing::trace_state_ptr& tr_state, cql_protocol_version_type version, bool skip_metadata = false); template future>> cql_server::connection::process_on_shard(unsigned shard, uint16_t stream, fragmented_temporary_buffer::istream is, service::client_state& cs, service_permit permit, tracing::trace_state_ptr trace_state, Process process_fn) { return _server.container().invoke_on(shard, _server._config.bounce_request_smp_service_group, [this, is = std::move(is), cs = cs.move_to_other_shard(), stream, permit = std::move(permit), process_fn, gt = tracing::global_trace_state_ptr(std::move(trace_state))] (cql_server& server) { service::client_state client_state = cs.get(); return do_with(bytes_ostream(), std::move(client_state), [this, &server, is = std::move(is), stream, process_fn, trace_state = tracing::trace_state_ptr(gt)] (bytes_ostream& linearization_buffer, service::client_state& client_state) mutable { request_reader in(is, linearization_buffer); return process_fn(client_state, server._query_processor, in, stream, _version, _cql_serialization_format, server.timeout_config(), /* FIXME */empty_service_permit(), std::move(trace_state), false).then([] (auto msg) { // result here has to be foreign ptr return std::get>>(std::move(msg)); }); }); }); } template future>> cql_server::connection::process(uint16_t stream, request_reader in, service::client_state& client_state, service_permit permit, tracing::trace_state_ptr trace_state, Process process_fn) { fragmented_temporary_buffer::istream is = in.get_stream(); return process_fn(client_state, _server._query_processor, in, stream, _version, _cql_serialization_format, _server.timeout_config(), permit, trace_state, true) .then([stream, &client_state, this, is, permit, process_fn, trace_state] (std::variant>, unsigned> msg) mutable { unsigned* shard = std::get_if(&msg); if (shard) { return process_on_shard(*shard, stream, is, client_state, std::move(permit), trace_state, process_fn); } return make_ready_future>>(std::get>>(std::move(msg))); }); } static future>, unsigned>> process_query_internal(service::client_state& client_state, distributed& qp, request_reader in, uint16_t stream, cql_protocol_version_type version, cql_serialization_format serialization_format, const ::timeout_config& timeout_config, service_permit permit, tracing::trace_state_ptr trace_state, bool init_trace) { auto query = in.read_long_string_view(); auto q_state = std::make_unique(client_state, trace_state, std::move(permit)); auto& query_state = q_state->query_state; q_state->options = in.read_options(version, serialization_format, timeout_config, qp.local().get_cql_config()); auto& options = *q_state->options; auto skip_metadata = options.skip_metadata(); if (init_trace) { tracing::set_page_size(trace_state, options.get_page_size()); tracing::set_consistency_level(trace_state, options.get_consistency()); tracing::set_optional_serial_consistency_level(trace_state, options.get_serial_consistency()); tracing::add_query(trace_state, query); tracing::set_user_timestamp(trace_state, options.get_specific_options().timestamp); tracing::begin(trace_state, "Execute CQL3 query", client_state.get_client_address()); } return qp.local().execute_direct(query, query_state, options).then([q_state = std::move(q_state), stream, skip_metadata, version] (auto msg) { if (msg->move_to_shard()) { return std::variant>, unsigned>(*msg->move_to_shard()); } else { tracing::trace(q_state->query_state.get_trace_state(), "Done processing - preparing a result"); return std::variant>, unsigned>(make_foreign(make_result(stream, *msg, q_state->query_state.get_trace_state(), version, skip_metadata))); } }); } future>> cql_server::connection::process_query(uint16_t stream, request_reader in, service::client_state& client_state, service_permit permit, tracing::trace_state_ptr trace_state) { return process(stream, in, client_state, std::move(permit), std::move(trace_state), process_query_internal); } future> cql_server::connection::process_prepare(uint16_t stream, request_reader in, service::client_state& client_state, tracing::trace_state_ptr trace_state) { auto query = sstring(in.read_long_string_view()); tracing::add_query(trace_state, query); tracing::begin(trace_state, "Preparing CQL3 query", client_state.get_client_address()); auto cpu_id = this_shard_id(); auto cpus = boost::irange(0u, smp::count); return parallel_for_each(cpus.begin(), cpus.end(), [this, query, cpu_id, &client_state] (unsigned int c) mutable { if (c != cpu_id) { return smp::submit_to(c, [this, query, &client_state] () mutable { return _server._query_processor.local().prepare(std::move(query), client_state, false).discard_result(); }); } else { return make_ready_future<>(); } }).then([this, query, stream, &client_state, trace_state] () mutable { tracing::trace(trace_state, "Done preparing on remote shards"); return _server._query_processor.local().prepare(std::move(query), client_state, false).then([this, stream, &client_state, trace_state] (auto msg) { tracing::trace(trace_state, "Done preparing on a local shard - preparing a result. ID is [{}]", seastar::value_of([&msg] { return messages::result_message::prepared::cql::get_id(msg); })); return make_result(stream, *msg, trace_state, _version); }); }); } static future>, unsigned>> process_execute_internal(service::client_state& client_state, distributed& qp, request_reader in, uint16_t stream, cql_protocol_version_type version, cql_serialization_format serialization_format, const ::timeout_config& timeout_config, service_permit permit, tracing::trace_state_ptr trace_state, bool init_trace) { cql3::prepared_cache_key_type cache_key(in.read_short_bytes()); auto& id = cql3::prepared_cache_key_type::cql_id(cache_key); bool needs_authorization = false; // First, try to lookup in the cache of already authorized statements. If the corresponding entry is not found there // look for the prepared statement and then authorize it. auto prepared = qp.local().get_prepared(client_state.user(), cache_key); if (!prepared) { needs_authorization = true; prepared = qp.local().get_prepared(cache_key); } if (!prepared) { throw exceptions::prepared_query_not_found_exception(id); } auto q_state = std::make_unique(client_state, trace_state, std::move(permit)); auto& query_state = q_state->query_state; if (version == 1) { std::vector values; in.read_value_view_list(version, values); auto consistency = in.read_consistency(); q_state->options = std::make_unique(qp.local().get_cql_config(), consistency, timeout_config, std::nullopt, values, false, cql3::query_options::specific_options::DEFAULT, serialization_format); } else { q_state->options = in.read_options(version, serialization_format, timeout_config, qp.local().get_cql_config()); } auto& options = *q_state->options; auto skip_metadata = options.skip_metadata(); if (init_trace) { tracing::set_page_size(trace_state, options.get_page_size()); tracing::set_consistency_level(trace_state, options.get_consistency()); tracing::set_optional_serial_consistency_level(trace_state, options.get_serial_consistency()); tracing::add_query(trace_state, prepared->statement->raw_cql_statement); tracing::add_prepared_statement(trace_state, prepared); tracing::begin(trace_state, seastar::value_of([&id] { return seastar::format("Execute CQL3 prepared query [{}]", id); }), client_state.get_client_address()); } auto stmt = prepared->statement; tracing::trace(query_state.get_trace_state(), "Checking bounds"); if (stmt->get_bound_terms() != options.get_values_count()) { const auto msg = format("Invalid amount of bind variables: expected {:d} received {:d}", stmt->get_bound_terms(), options.get_values_count()); tracing::trace(query_state.get_trace_state(), msg); throw exceptions::invalid_request_exception(msg); } options.prepare(prepared->bound_names); if (init_trace) { tracing::add_prepared_query_options(trace_state, options); } tracing::trace(trace_state, "Processing a statement"); return qp.local().execute_prepared(std::move(prepared), std::move(cache_key), query_state, options, needs_authorization) .then([trace_state = query_state.get_trace_state(), skip_metadata, q_state = std::move(q_state), stream, version] (auto msg) { if (msg->move_to_shard()) { return std::variant>, unsigned>(*msg->move_to_shard()); } else { tracing::trace(q_state->query_state.get_trace_state(), "Done processing - preparing a result"); return std::variant>, unsigned>(make_foreign(make_result(stream, *msg, q_state->query_state.get_trace_state(), version, skip_metadata))); } }); } future>> cql_server::connection::process_execute(uint16_t stream, request_reader in, service::client_state& client_state, service_permit permit, tracing::trace_state_ptr trace_state) { return process(stream, in, client_state, std::move(permit), std::move(trace_state), process_execute_internal); } static future>, unsigned>> process_batch_internal(service::client_state& client_state, distributed& qp, request_reader in, uint16_t stream, cql_protocol_version_type version, cql_serialization_format serialization_format, const ::timeout_config& timeout_config, service_permit permit, tracing::trace_state_ptr trace_state, bool init_trace) { if (version == 1) { throw exceptions::protocol_exception("BATCH messages are not support in version 1 of the protocol"); } const auto type = in.read_byte(); const unsigned n = in.read_short(); std::vector modifications; std::vector> values; std::unordered_map pending_authorization_entries; modifications.reserve(n); values.reserve(n); if (init_trace) { tracing::begin(trace_state, "Execute batch of CQL3 queries", client_state.get_client_address()); } for ([[gnu::unused]] auto i : boost::irange(0u, n)) { const auto kind = in.read_byte(); std::unique_ptr stmt_ptr; cql3::statements::prepared_statement::checked_weak_ptr ps; bool needs_authorization(kind == 0); switch (kind) { case 0: { auto query = in.read_long_string_view(); stmt_ptr = qp.local().get_statement(query, client_state); ps = stmt_ptr->checked_weak_from_this(); if (init_trace) { tracing::add_query(trace_state, query); } break; } case 1: { cql3::prepared_cache_key_type cache_key(in.read_short_bytes()); auto& id = cql3::prepared_cache_key_type::cql_id(cache_key); // First, try to lookup in the cache of already authorized statements. If the corresponding entry is not found there // look for the prepared statement and then authorize it. ps = qp.local().get_prepared(client_state.user(), cache_key); if (!ps) { ps = qp.local().get_prepared(cache_key); if (!ps) { throw exceptions::prepared_query_not_found_exception(id); } // authorize a particular prepared statement only once needs_authorization = pending_authorization_entries.emplace(std::move(cache_key), ps->checked_weak_from_this()).second; } if (init_trace) { tracing::add_query(trace_state, ps->statement->raw_cql_statement); } break; } default: throw exceptions::protocol_exception( "Invalid query kind in BATCH messages. Must be 0 or 1 but got " + std::to_string(int(kind))); } if (dynamic_cast(ps->statement.get()) == nullptr) { throw exceptions::invalid_request_exception("Invalid statement in batch: only UPDATE, INSERT and DELETE statements are allowed."); } ::shared_ptr modif_statement_ptr = static_pointer_cast(ps->statement); if (init_trace) { tracing::add_table_name(trace_state, modif_statement_ptr->keyspace(), modif_statement_ptr->column_family()); tracing::add_prepared_statement(trace_state, ps); } modifications.emplace_back(std::move(modif_statement_ptr), needs_authorization); std::vector tmp; in.read_value_view_list(version, tmp); auto stmt = ps->statement; if (stmt->get_bound_terms() != tmp.size()) { throw exceptions::invalid_request_exception(format("There were {:d} markers(?) in CQL but {:d} bound variables", stmt->get_bound_terms(), tmp.size())); } values.emplace_back(std::move(tmp)); } auto q_state = std::make_unique(client_state, trace_state, std::move(permit)); auto& query_state = q_state->query_state; // #563. CQL v2 encodes query_options in v1 format for batch requests. q_state->options = std::make_unique(cql3::query_options::make_batch_options(std::move(*in.read_options(version < 3 ? 1 : version, serialization_format, timeout_config, qp.local().get_cql_config())), std::move(values))); auto& options = *q_state->options; if (init_trace) { tracing::set_consistency_level(trace_state, options.get_consistency()); tracing::set_optional_serial_consistency_level(trace_state, options.get_serial_consistency()); tracing::add_prepared_query_options(trace_state, options); tracing::trace(trace_state, "Creating a batch statement"); } auto batch = ::make_shared(cql3::statements::batch_statement::type(type), std::move(modifications), cql3::attributes::none(), qp.local().get_cql_stats()); return qp.local().execute_batch(batch, query_state, options, std::move(pending_authorization_entries)) .then([stream, batch, q_state = std::move(q_state), trace_state = query_state.get_trace_state(), version] (auto msg) { if (msg->move_to_shard()) { return std::variant>, unsigned>(*msg->move_to_shard()); } else { tracing::trace(q_state->query_state.get_trace_state(), "Done processing - preparing a result"); return std::variant>, unsigned>(make_foreign(make_result(stream, *msg, trace_state, version))); } }); } future>> cql_server::connection::process_batch(uint16_t stream, request_reader in, service::client_state& client_state, service_permit permit, tracing::trace_state_ptr trace_state) { return process(stream, in, client_state, permit, std::move(trace_state), process_batch_internal); } future> cql_server::connection::process_register(uint16_t stream, request_reader in, service::client_state& client_state, tracing::trace_state_ptr trace_state) { std::vector event_types; in.read_string_list(event_types); for (auto&& event_type : event_types) { auto et = parse_event_type(event_type); _server._notifier->register_event(et, this); } return make_ready_future>(make_ready(stream, std::move(trace_state))); } std::unique_ptr cql_server::connection::make_unavailable_error(int16_t stream, exceptions::exception_code err, sstring msg, db::consistency_level cl, int32_t required, int32_t alive, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_consistency(cl); response->write_int(required); response->write_int(alive); return response; } std::unique_ptr cql_server::connection::make_read_timeout_error(int16_t stream, exceptions::exception_code err, sstring msg, db::consistency_level cl, int32_t received, int32_t blockfor, bool data_present, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_consistency(cl); response->write_int(received); response->write_int(blockfor); response->write_byte(data_present); return response; } std::unique_ptr cql_server::connection::make_read_failure_error(int16_t stream, exceptions::exception_code err, sstring msg, db::consistency_level cl, int32_t received, int32_t numfailures, int32_t blockfor, bool data_present, const tracing::trace_state_ptr& tr_state) const { if (_version < 4) { return make_read_timeout_error(stream, err, std::move(msg), cl, received, blockfor, data_present, tr_state); } auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_consistency(cl); response->write_int(received); response->write_int(blockfor); response->write_int(numfailures); response->write_byte(data_present); return response; } std::unique_ptr cql_server::connection::make_mutation_write_timeout_error(int16_t stream, exceptions::exception_code err, sstring msg, db::consistency_level cl, int32_t received, int32_t blockfor, db::write_type type, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_consistency(cl); response->write_int(received); response->write_int(blockfor); response->write_string(format("{}", type)); return response; } std::unique_ptr cql_server::connection::make_mutation_write_failure_error(int16_t stream, exceptions::exception_code err, sstring msg, db::consistency_level cl, int32_t received, int32_t numfailures, int32_t blockfor, db::write_type type, const tracing::trace_state_ptr& tr_state) const { if (_version < 4) { return make_mutation_write_timeout_error(stream, err, std::move(msg), cl, received, blockfor, type, tr_state); } auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_consistency(cl); response->write_int(received); response->write_int(blockfor); response->write_int(numfailures); response->write_string(format("{}", type)); return response; } std::unique_ptr cql_server::connection::make_already_exists_error(int16_t stream, exceptions::exception_code err, sstring msg, sstring ks_name, sstring cf_name, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_string(ks_name); response->write_string(cf_name); return response; } std::unique_ptr cql_server::connection::make_unprepared_error(int16_t stream, exceptions::exception_code err, sstring msg, bytes id, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); response->write_short_bytes(id); return response; } std::unique_ptr cql_server::connection::make_error(int16_t stream, exceptions::exception_code err, sstring msg, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::ERROR, tr_state); response->write_int(static_cast(err)); response->write_string(msg); return response; } std::unique_ptr cql_server::connection::make_ready(int16_t stream, const tracing::trace_state_ptr& tr_state) const { return std::make_unique(stream, cql_binary_opcode::READY, tr_state); } std::unique_ptr cql_server::connection::make_autheticate(int16_t stream, std::string_view clz, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::AUTHENTICATE, tr_state); response->write_string(clz); return response; } std::unique_ptr cql_server::connection::make_auth_success(int16_t stream, bytes b, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::AUTH_SUCCESS, tr_state); response->write_bytes(std::move(b)); return response; } std::unique_ptr cql_server::connection::make_auth_challenge(int16_t stream, bytes b, const tracing::trace_state_ptr& tr_state) const { auto response = std::make_unique(stream, cql_binary_opcode::AUTH_CHALLENGE, tr_state); response->write_bytes(std::move(b)); return response; } std::unique_ptr cql_server::connection::make_supported(int16_t stream, const tracing::trace_state_ptr& tr_state) const { std::multimap opts; opts.insert({"CQL_VERSION", cql3::query_processor::CQL_VERSION}); opts.insert({"COMPRESSION", "lz4"}); opts.insert({"COMPRESSION", "snappy"}); if (_server._config.allow_shard_aware_drivers) { opts.insert({"SCYLLA_SHARD", format("{:d}", this_shard_id())}); opts.insert({"SCYLLA_NR_SHARDS", format("{:d}", smp::count)}); opts.insert({"SCYLLA_SHARDING_ALGORITHM", dht::cpu_sharding_algorithm_name()}); if (_server._config.shard_aware_transport_port) { opts.insert({"SCYLLA_SHARD_AWARE_PORT", format("{:d}", *_server._config.shard_aware_transport_port)}); } if (_server._config.shard_aware_transport_port_ssl) { opts.insert({"SCYLLA_SHARD_AWARE_PORT_SSL", format("{:d}", *_server._config.shard_aware_transport_port_ssl)}); } opts.insert({"SCYLLA_SHARDING_IGNORE_MSB", format("{:d}", _server._config.sharding_ignore_msb)}); opts.insert({"SCYLLA_PARTITIONER", _server._config.partitioner_name}); } for (cql_protocol_extension ext : supported_cql_protocol_extensions()) { const sstring ext_key_name = protocol_extension_name(ext); std::vector params = additional_options_for_proto_ext(ext); if (params.empty()) { opts.emplace(ext_key_name, ""); } else { for (sstring val : params) { opts.emplace(ext_key_name, std::move(val)); } } } auto response = std::make_unique(stream, cql_binary_opcode::SUPPORTED, tr_state); response->write_string_multimap(std::move(opts)); return response; } class cql_server::fmt_visitor : public messages::result_message::visitor_base { private: uint8_t _version; cql_server::response& _response; bool _skip_metadata; public: fmt_visitor(uint8_t version, cql_server::response& response, bool skip_metadata) : _version{version} , _response{response} , _skip_metadata{skip_metadata} { } virtual void visit(const messages::result_message::void_message&) override { _response.write_int(0x0001); } virtual void visit(const messages::result_message::set_keyspace& m) override { _response.write_int(0x0003); _response.write_string(m.get_keyspace()); } virtual void visit(const messages::result_message::prepared::cql& m) override { _response.write_int(0x0004); _response.write_short_bytes(m.get_id()); _response.write(m.metadata(), _version); if (_version > 1) { _response.write(*m.result_metadata()); } } virtual void visit(const messages::result_message::schema_change& m) override { auto change = m.get_change(); switch (change->type) { case event::event_type::SCHEMA_CHANGE: { auto sc = static_pointer_cast(change); _response.write_int(0x0005); _response.serialize(*sc, _version); break; } default: assert(0); } } virtual void visit(const messages::result_message::rows& m) override { _response.write_int(0x0002); auto& rs = m.rs(); _response.write(rs.get_metadata(), _skip_metadata); auto row_count_plhldr = _response.write_int_placeholder(); class visitor { cql_server::response& _response; int32_t _row_count = 0; public: visitor(cql_server::response& r) : _response(r) { } void start_row() { _row_count++; } void accept_value(std::optional cell) { _response.write_value(cell); } void end_row() { } int32_t row_count() const { return _row_count; } }; auto v = visitor(_response); rs.visit(v); row_count_plhldr.write(v.row_count()); } }; std::unique_ptr make_result(int16_t stream, messages::result_message& msg, const tracing::trace_state_ptr& tr_state, cql_protocol_version_type version, bool skip_metadata) { auto response = std::make_unique(stream, cql_binary_opcode::RESULT, tr_state); if (__builtin_expect(!msg.warnings().empty() && version > 3, false)) { response->set_frame_flag(cql_frame_flags::warning); response->write_string_list(msg.warnings()); } cql_server::fmt_visitor fmt{version, *response, skip_metadata}; msg.accept(fmt); return response; } std::unique_ptr cql_server::connection::make_topology_change_event(const event::topology_change& event) const { auto response = std::make_unique(-1, cql_binary_opcode::EVENT, tracing::trace_state_ptr()); response->write_string("TOPOLOGY_CHANGE"); response->write_string(to_string(event.change)); response->write_inet(event.node); return response; } std::unique_ptr cql_server::connection::make_status_change_event(const event::status_change& event) const { auto response = std::make_unique(-1, cql_binary_opcode::EVENT, tracing::trace_state_ptr()); response->write_string("STATUS_CHANGE"); response->write_string(to_string(event.status)); response->write_inet(event.node); return response; } std::unique_ptr cql_server::connection::make_schema_change_event(const event::schema_change& event) const { auto response = std::make_unique(-1, cql_binary_opcode::EVENT, tracing::trace_state_ptr()); response->write_string("SCHEMA_CHANGE"); response->serialize(event, _version); return response; } void cql_server::connection::write_response(foreign_ptr>&& response, service_permit permit, cql_compression compression) { _ready_to_respond = _ready_to_respond.then([this, compression, response = std::move(response), permit = std::move(permit)] () mutable { auto message = response->make_message(_version, compression); message.on_delete([response = std::move(response)] { }); return _write_buf.write(std::move(message)).then([this] { return _write_buf.flush(); }); }); } scattered_message cql_server::response::make_message(uint8_t version, cql_compression compression) { if (compression != cql_compression::none) { compress(compression); } scattered_message msg; auto frame = make_frame(version, _body.size()); msg.append(std::move(frame)); for (auto&& fragment : _body.fragments()) { msg.append_static(reinterpret_cast(fragment.data()), fragment.size()); } return msg; } void cql_server::response::compress(cql_compression compression) { switch (compression) { case cql_compression::lz4: compress_lz4(); break; case cql_compression::snappy: compress_snappy(); break; default: throw std::invalid_argument("Invalid CQL compression algorithm"); } set_frame_flag(cql_frame_flags::compression); } void cql_server::response::compress_lz4() { using namespace compression_buffers; auto view = input_buffer.get_linearized_view(_body); const char* input = reinterpret_cast(view.data()); size_t input_len = view.size(); size_t output_len = LZ4_COMPRESSBOUND(input_len) + 4; _body = output_buffer.make_buffer(output_len, [&] (bytes_mutable_view output_view) { char* output = reinterpret_cast(output_view.data()); output[0] = (input_len >> 24) & 0xFF; output[1] = (input_len >> 16) & 0xFF; output[2] = (input_len >> 8) & 0xFF; output[3] = input_len & 0xFF; #ifdef HAVE_LZ4_COMPRESS_DEFAULT auto ret = LZ4_compress_default(input, output + 4, input_len, LZ4_compressBound(input_len)); #else auto ret = LZ4_compress(input, output + 4, input_len); #endif if (ret == 0) { throw std::runtime_error("CQL frame LZ4 compression failure"); } return ret + 4; }); on_compression_buffer_use(); } void cql_server::response::compress_snappy() { using namespace compression_buffers; auto view = input_buffer.get_linearized_view(_body); const char* input = reinterpret_cast(view.data()); size_t input_len = view.size(); size_t output_len = snappy_max_compressed_length(input_len); _body = output_buffer.make_buffer(output_len, [&] (bytes_mutable_view output_view) { char* output = reinterpret_cast(output_view.data()); if (snappy_compress(input, input_len, output, &output_len) != SNAPPY_OK) { throw std::runtime_error("CQL frame Snappy compression failure"); } return output_len; }); on_compression_buffer_use(); } void cql_server::response::serialize(const event::schema_change& event, uint8_t version) { if (version >= 3) { write_string(to_string(event.change)); write_string(to_string(event.target)); write_string(event.keyspace); switch (event.target) { case event::schema_change::target_type::KEYSPACE: break; case event::schema_change::target_type::TYPE: case event::schema_change::target_type::TABLE: write_string(event.arguments[0]); break; case event::schema_change::target_type::FUNCTION: case event::schema_change::target_type::AGGREGATE: write_string(event.arguments[0]); write_string_list(std::vector(event.arguments.begin() + 1, event.arguments.end())); break; } } else { switch (event.target) { // FIXME: Should we handle FUNCTION and AGGREGATE the same way as type? // FIXME: How do we get here? Can a client using v2 know about UDF? case event::schema_change::target_type::TYPE: case event::schema_change::target_type::FUNCTION: case event::schema_change::target_type::AGGREGATE: // The v1/v2 protocol is unable to represent these changes. Tell the // client that the keyspace was updated instead. write_string(to_string(event::schema_change::change_type::UPDATED)); write_string(event.keyspace); write_string(""); break; case event::schema_change::target_type::TABLE: case event::schema_change::target_type::KEYSPACE: write_string(to_string(event.change)); write_string(event.keyspace); if (event.target == event::schema_change::target_type::TABLE) { write_string(event.arguments[0]); } else { write_string(""); } } } } void cql_server::response::write_byte(uint8_t b) { auto s = reinterpret_cast(&b); _body.write(bytes_view(s, sizeof(b))); } void cql_server::response::write_int(int32_t n) { auto u = htonl(n); auto *s = reinterpret_cast(&u); _body.write(bytes_view(s, sizeof(u))); } cql_server::response::placeholder cql_server::response::write_int_placeholder() { return placeholder(_body.write_place_holder(sizeof(int32_t))); } void cql_server::response::write_long(int64_t n) { auto u = htonq(n); auto *s = reinterpret_cast(&u); _body.write(bytes_view(s, sizeof(u))); } void cql_server::response::write_short(uint16_t n) { auto u = htons(n); auto *s = reinterpret_cast(&u); _body.write(bytes_view(s, sizeof(u))); } template inline T cast_if_fits(size_t v) { size_t max = std::numeric_limits::max(); if (v > max) { throw std::runtime_error(format("Value too large, {:d} > {:d}", v, max)); } return static_cast(v); } void cql_server::response::write_string(std::string_view s) { write_short(cast_if_fits(s.size())); _body.write(bytes_view(reinterpret_cast(s.data()), s.size())); } void cql_server::response::write_bytes_as_string(bytes_view s) { write_short(cast_if_fits(s.size())); _body.write(s); } void cql_server::response::write_long_string(const sstring& s) { write_int(cast_if_fits(s.size())); _body.write(bytes_view(reinterpret_cast(s.data()), s.size())); } void cql_server::response::write_string_list(std::vector string_list) { write_short(cast_if_fits(string_list.size())); for (auto&& s : string_list) { write_string(s); } } void cql_server::response::write_bytes(bytes b) { write_int(cast_if_fits(b.size())); _body.write(b); } void cql_server::response::write_short_bytes(bytes b) { write_short(cast_if_fits(b.size())); _body.write(b); } void cql_server::response::write_inet(socket_address inet) { auto addr = inet.addr(); write_byte(uint8_t(addr.size())); auto * p = static_cast(addr.data()); _body.write(bytes_view(p, addr.size())); write_int(inet.port()); } void cql_server::response::write_consistency(db::consistency_level c) { write_short(consistency_to_wire(c)); } void cql_server::response::write_string_map(std::map string_map) { write_short(cast_if_fits(string_map.size())); for (auto&& s : string_map) { write_string(s.first); write_string(s.second); } } void cql_server::response::write_string_multimap(std::multimap string_map) { std::vector keys; for (auto it = string_map.begin(), end = string_map.end(); it != end; it = string_map.upper_bound(it->first)) { keys.push_back(it->first); } write_short(cast_if_fits(keys.size())); for (auto&& key : keys) { std::vector values; auto range = string_map.equal_range(key); for (auto it = range.first; it != range.second; ++it) { values.push_back(it->second); } write_string(key); write_string_list(values); } } void cql_server::response::write_value(bytes_opt value) { if (!value) { write_int(-1); return; } write_int(value->size()); _body.write(*value); } void cql_server::response::write_value(std::optional value) { if (!value) { write_int(-1); return; } write_int(value->size_bytes()); using boost::range::for_each; for_each(*value, [&] (bytes_view fragment) { _body.write(fragment); }); } class type_codec { private: enum class type_id : int16_t { CUSTOM = 0x0000, ASCII = 0x0001, BIGINT = 0x0002, BLOB = 0x0003, BOOLEAN = 0x0004, COUNTER = 0x0005, DECIMAL = 0x0006, DOUBLE = 0x0007, FLOAT = 0x0008, INT = 0x0009, TIMESTAMP = 0x000B, UUID = 0x000C, VARCHAR = 0x000D, VARINT = 0x000E, TIMEUUID = 0x000F, INET = 0x0010, DATE = 0x0011, TIME = 0x0012, SMALLINT = 0x0013, TINYINT = 0x0014, DURATION = 0x0015, LIST = 0x0020, MAP = 0x0021, SET = 0x0022, UDT = 0x0030, TUPLE = 0x0031, }; using type_id_to_type_type = std::unordered_map; static thread_local const type_id_to_type_type type_id_to_type; public: static void encode(cql_server::response& r, data_type type) { type = type->underlying_type(); // For compatibility sake, we still return DateType as the timestamp type in resultSet metadata (#5723) if (type == date_type) { type = timestamp_type; } auto i = type_id_to_type.find(type); if (i != type_id_to_type.end()) { r.write_short(static_cast::type>(i->second)); return; } if (type->is_reversed()) { fail(unimplemented::cause::REVERSED); } if (type->is_user_type()) { r.write_short(uint16_t(type_id::UDT)); auto udt = static_pointer_cast(type); r.write_string(udt->_keyspace); r.write_bytes_as_string(udt->_name); r.write_short(udt->size()); for (auto&& i : boost::irange(0, udt->size())) { r.write_bytes_as_string(udt->field_name(i)); encode(r, udt->field_type(i)); } return; } if (type->is_tuple()) { r.write_short(uint16_t(type_id::TUPLE)); auto ttype = static_pointer_cast(type); r.write_short(ttype->size()); for (auto&& t : ttype->all_types()) { encode(r, t); } return; } if (type->is_collection()) { auto&& ctype = static_cast(type.get()); if (ctype->get_kind() == abstract_type::kind::map) { r.write_short(uint16_t(type_id::MAP)); auto&& mtype = static_cast(ctype); encode(r, mtype->get_keys_type()); encode(r, mtype->get_values_type()); } else if (ctype->get_kind() == abstract_type::kind::set) { r.write_short(uint16_t(type_id::SET)); auto&& stype = static_cast(ctype); encode(r, stype->get_elements_type()); } else if (ctype->get_kind() == abstract_type::kind::list) { r.write_short(uint16_t(type_id::LIST)); auto&& ltype = static_cast(ctype); encode(r, ltype->get_elements_type()); } else { abort(); } return; } abort(); } }; thread_local const type_codec::type_id_to_type_type type_codec::type_id_to_type { { ascii_type, type_id::ASCII }, { long_type, type_id::BIGINT }, { bytes_type, type_id::BLOB }, { boolean_type, type_id::BOOLEAN }, { counter_type, type_id::COUNTER }, { decimal_type, type_id::DECIMAL }, { double_type, type_id::DOUBLE }, { float_type, type_id::FLOAT }, { int32_type, type_id::INT }, { byte_type, type_id::TINYINT }, { duration_type, type_id::DURATION }, { short_type, type_id::SMALLINT }, { timestamp_type, type_id::TIMESTAMP }, { uuid_type, type_id::UUID }, { utf8_type, type_id::VARCHAR }, { varint_type, type_id::VARINT }, { timeuuid_type, type_id::TIMEUUID }, { simple_date_type, type_id::DATE }, { time_type, type_id::TIME }, { inet_addr_type, type_id::INET }, }; void cql_server::response::write(const cql3::metadata& m, bool no_metadata) { auto flags = m.flags(); bool global_tables_spec = m.flags().contains(); bool has_more_pages = m.flags().contains(); if (no_metadata) { flags.set(); } write_int(flags.mask()); write_int(m.column_count()); if (has_more_pages) { write_value(m.paging_state()->serialize()); } if (no_metadata) { return; } auto names_i = m.get_names().begin(); if (global_tables_spec) { auto first_spec = *names_i; write_string(first_spec->ks_name); write_string(first_spec->cf_name); } for (uint32_t i = 0; i < m.column_count(); ++i, ++names_i) { lw_shared_ptr name = *names_i; if (!global_tables_spec) { write_string(name->ks_name); write_string(name->cf_name); } write_string(name->name->text()); type_codec::encode(*this, name->type); }; } void cql_server::response::write(const cql3::prepared_metadata& m, uint8_t version) { bool global_tables_spec = m.flags().contains(); write_int(m.flags().mask()); write_int(m.names().size()); if (version >= 4) { if (!global_tables_spec) { write_int(0); } else { write_int(m.partition_key_bind_indices().size()); for (uint16_t bind_index : m.partition_key_bind_indices()) { write_short(bind_index); } } } if (global_tables_spec) { write_string(m.names()[0]->ks_name); write_string(m.names()[0]->cf_name); } for (auto const& name : m.names()) { if (!global_tables_spec) { write_string(name->ks_name); write_string(name->cf_name); } write_string(name->name->text()); type_codec::encode(*this, name->type); } } }