/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* * Copyright 2015 Cloudius Systems * * Modified by Cloudius Systems */ #include "service/migration_manager.hh" #include "service/migration_listener.hh" #include "message/messaging_service.hh" #include "service/storage_service.hh" #include "service/migration_task.hh" #include "utils/runtime.hh" #include "gms/gossiper.hh" namespace service { static logging::logger logger("migration_manager"); distributed _the_migration_manager; using namespace std::chrono_literals; const std::chrono::milliseconds migration_manager::migration_delay = 60000ms; migration_manager::migration_manager() : _listeners{} { } future<> migration_manager::stop() { return make_ready_future<>(); } void migration_manager::register_listener(migration_listener* listener) { _listeners.emplace_back(listener); } void migration_manager::unregister_listener(migration_listener* listener) { _listeners.erase(std::remove(_listeners.begin(), _listeners.end(), listener), _listeners.end()); } future<> migration_manager::schedule_schema_pull(const gms::inet_address& endpoint, const gms::endpoint_state& state) { const auto& value = state.get_application_state(gms::application_state::SCHEMA); if (endpoint != utils::fb_utilities::get_broadcast_address() && value) { return maybe_schedule_schema_pull(utils::UUID{value->value}, endpoint); } return make_ready_future<>(); } /** * If versions differ this node sends request with local migration list to the endpoint * and expecting to receive a list of migrations to apply locally. */ future<> migration_manager::maybe_schedule_schema_pull(const utils::UUID& their_version, const gms::inet_address& endpoint) { auto& proxy = get_local_storage_proxy(); auto& db = proxy.get_db().local(); if (db.get_version() == their_version || !should_pull_schema_from(endpoint)) { logger.debug("Not pulling schema because versions match or shouldPullSchemaFrom returned false"); return make_ready_future<>(); } if (db.get_version() == database::empty_version || runtime::get_uptime() < migration_delay) { // If we think we may be bootstrapping or have recently started, submit MigrationTask immediately logger.debug("Submitting migration task for {}", endpoint); return submit_migration_task(endpoint); } else { // Include a delay to make sure we have a chance to apply any changes being // pushed out simultaneously. See CASSANDRA-5025 return sleep(migration_delay).then([this, &proxy, endpoint] { // grab the latest version of the schema since it may have changed again since the initial scheduling auto& gossiper = gms::get_local_gossiper(); auto ep_state = gossiper.get_endpoint_state_for_endpoint(endpoint); if (!ep_state) { logger.debug("epState vanished for {}, not submitting migration task", endpoint); return make_ready_future<>(); } const auto& value = ep_state->get_application_state(gms::application_state::SCHEMA); utils::UUID current_version{value->value}; auto& db = proxy.get_db().local(); if (db.get_version() == current_version) { logger.debug("not submitting migration task for {} because our versions match", endpoint); return make_ready_future<>(); } logger.debug("submitting migration task for {}", endpoint); return submit_migration_task(endpoint); }); } } future<> migration_manager::submit_migration_task(const gms::inet_address& endpoint) { return service::migration_task::run_may_throw(get_local_storage_proxy(), endpoint); } bool migration_manager::should_pull_schema_from(const gms::inet_address& endpoint) { /* * Don't request schema from nodes with a differnt or unknonw major version (may have incompatible schema) * Don't request schema from fat clients */ auto& ms = net::get_local_messaging_service(); return ms.knows_version(endpoint) && ms.get_raw_version(endpoint) == net::messaging_service::current_version && !gms::get_local_gossiper().is_gossip_only_member(endpoint); } future<> migration_manager::notify_create_keyspace(const lw_shared_ptr& ksm) { return get_migration_manager().invoke_on_all([name = ksm->name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_create_keyspace(name); } }); } future<> migration_manager::notify_create_column_family(schema_ptr cfm) { return get_migration_manager().invoke_on_all([ks_name = cfm->ks_name(), cf_name = cfm->cf_name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_create_column_family(ks_name, cf_name); } }); } #if 0 public void notifyCreateUserType(UserType ut) { for (IMigrationListener listener : listeners) listener.onCreateUserType(ut.keyspace, ut.getNameAsString()); } public void notifyCreateFunction(UDFunction udf) { for (IMigrationListener listener : listeners) listener.onCreateFunction(udf.name().keyspace, udf.name().name); } public void notifyCreateAggregate(UDAggregate udf) { for (IMigrationListener listener : listeners) listener.onCreateAggregate(udf.name().keyspace, udf.name().name); } #endif future<> migration_manager::notify_update_keyspace(const lw_shared_ptr& ksm) { return get_migration_manager().invoke_on_all([name = ksm->name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_update_keyspace(name); } }); } future<> migration_manager::notify_update_column_family(schema_ptr cfm) { return get_migration_manager().invoke_on_all([ks_name = cfm->ks_name(), cf_name = cfm->cf_name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_update_column_family(ks_name, cf_name); } }); } #if 0 public void notifyUpdateUserType(UserType ut) { for (IMigrationListener listener : listeners) listener.onUpdateUserType(ut.keyspace, ut.getNameAsString()); } public void notifyUpdateFunction(UDFunction udf) { for (IMigrationListener listener : listeners) listener.onUpdateFunction(udf.name().keyspace, udf.name().name); } public void notifyUpdateAggregate(UDAggregate udf) { for (IMigrationListener listener : listeners) listener.onUpdateAggregate(udf.name().keyspace, udf.name().name); } #endif future<> migration_manager::notify_drop_keyspace(const lw_shared_ptr& ksm) { return get_migration_manager().invoke_on_all([name = ksm->name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_drop_keyspace(name); } }); } future<> migration_manager::notify_drop_column_family(schema_ptr cfm) { return get_migration_manager().invoke_on_all([ks_name = cfm->ks_name(), cf_name = cfm->cf_name()] (auto&& mm) { for (auto&& listener : mm._listeners) { listener->on_drop_column_family(ks_name, cf_name); } }); } #if 0 public void notifyDropUserType(UserType ut) { for (IMigrationListener listener : listeners) listener.onDropUserType(ut.keyspace, ut.getNameAsString()); } public void notifyDropFunction(UDFunction udf) { for (IMigrationListener listener : listeners) listener.onDropFunction(udf.name().keyspace, udf.name().name); } public void notifyDropAggregate(UDAggregate udf) { for (IMigrationListener listener : listeners) listener.onDropAggregate(udf.name().keyspace, udf.name().name); } #endif future<>migration_manager::announce_new_keyspace(lw_shared_ptr ksm, bool announce_locally) { return announce_new_keyspace(ksm, db_clock::now_in_usecs(), announce_locally); } future<> migration_manager::announce_new_keyspace(lw_shared_ptr ksm, api::timestamp_type timestamp, bool announce_locally) { #if 0 ksm.validate(); #endif auto& proxy = get_local_storage_proxy(); if (proxy.get_db().local().has_keyspace(ksm->name())) { throw exceptions::already_exists_exception{ksm->name()}; } logger.info("Create new Keyspace: {}", ksm->name()); auto mutations = db::schema_tables::make_create_keyspace_mutations(ksm, timestamp); return announce(std::move(mutations), announce_locally); } future<> migration_manager::announce_new_column_family(schema_ptr cfm, bool announce_locally) { #if 0 cfm.validate(); #endif try { auto& db = get_local_storage_proxy().get_db().local(); auto&& keyspace = db.find_keyspace(cfm->ks_name()); if (db.has_schema(cfm->ks_name(), cfm->cf_name())) { throw exceptions::already_exists_exception(cfm->ks_name(), cfm->cf_name()); } logger.info("Create new table: {}", cfm->cf_name()); auto mutations = db::schema_tables::make_create_table_mutations(keyspace.metadata(), cfm, db_clock::now_in_usecs()); return announce(std::move(mutations), announce_locally); } catch (const no_such_keyspace& e) { throw exceptions::configuration_exception(sprint("Cannot add table '%s' to non existing keyspace '%s'.", cfm->cf_name(), cfm->ks_name())); } } future<> migration_manager::announce_column_family_update(schema_ptr cfm, bool from_thrift, bool announce_locally) { warn(unimplemented::cause::MIGRATIONS); return make_ready_future<>(); #if 0 cfm.validate(); CFMetaData oldCfm = Schema.instance.getCFMetaData(cfm.ksName, cfm.cfName); if (oldCfm == null) throw new ConfigurationException(String.format("Cannot update non existing table '%s' in keyspace '%s'.", cfm.cfName, cfm.ksName)); KSMetaData ksm = Schema.instance.getKSMetaData(cfm.ksName); oldCfm.validateCompatility(cfm); logger.info(String.format("Update table '%s/%s' From %s To %s", cfm.ksName, cfm.cfName, oldCfm, cfm)); announce(LegacySchemaTables.makeUpdateTableMutation(ksm, oldCfm, cfm, FBUtilities.timestampMicros(), fromThrift), announceLocally); #endif } #if 0 public static void announceNewType(UserType newType, boolean announceLocally) { KSMetaData ksm = Schema.instance.getKSMetaData(newType.keyspace); announce(LegacySchemaTables.makeCreateTypeMutation(ksm, newType, FBUtilities.timestampMicros()), announceLocally); } public static void announceNewFunction(UDFunction udf, boolean announceLocally) { logger.info(String.format("Create scalar function '%s'", udf.name())); KSMetaData ksm = Schema.instance.getKSMetaData(udf.name().keyspace); announce(LegacySchemaTables.makeCreateFunctionMutation(ksm, udf, FBUtilities.timestampMicros()), announceLocally); } public static void announceNewAggregate(UDAggregate udf, boolean announceLocally) { logger.info(String.format("Create aggregate function '%s'", udf.name())); KSMetaData ksm = Schema.instance.getKSMetaData(udf.name().keyspace); announce(LegacySchemaTables.makeCreateAggregateMutation(ksm, udf, FBUtilities.timestampMicros()), announceLocally); } public static void announceKeyspaceUpdate(KSMetaData ksm) throws ConfigurationException { announceKeyspaceUpdate(ksm, false); } public static void announceKeyspaceUpdate(KSMetaData ksm, boolean announceLocally) throws ConfigurationException { ksm.validate(); KSMetaData oldKsm = Schema.instance.getKSMetaData(ksm.name); if (oldKsm == null) throw new ConfigurationException(String.format("Cannot update non existing keyspace '%s'.", ksm.name)); logger.info(String.format("Update Keyspace '%s' From %s To %s", ksm.name, oldKsm, ksm)); announce(LegacySchemaTables.makeCreateKeyspaceMutation(ksm, FBUtilities.timestampMicros()), announceLocally); } public static void announceColumnFamilyUpdate(CFMetaData cfm, boolean fromThrift) throws ConfigurationException { announceColumnFamilyUpdate(cfm, fromThrift, false); } public static void announceColumnFamilyUpdate(CFMetaData cfm, boolean fromThrift, boolean announceLocally) throws ConfigurationException { cfm.validate(); CFMetaData oldCfm = Schema.instance.getCFMetaData(cfm.ksName, cfm.cfName); if (oldCfm == null) throw new ConfigurationException(String.format("Cannot update non existing table '%s' in keyspace '%s'.", cfm.cfName, cfm.ksName)); KSMetaData ksm = Schema.instance.getKSMetaData(cfm.ksName); oldCfm.validateCompatility(cfm); logger.info(String.format("Update table '%s/%s' From %s To %s", cfm.ksName, cfm.cfName, oldCfm, cfm)); announce(LegacySchemaTables.makeUpdateTableMutation(ksm, oldCfm, cfm, FBUtilities.timestampMicros(), fromThrift), announceLocally); } public static void announceTypeUpdate(UserType updatedType, boolean announceLocally) { announceNewType(updatedType, announceLocally); } #endif future<> migration_manager::announce_keyspace_drop(const sstring& ks_name, bool announce_locally) { try { auto& db = get_local_storage_proxy().get_db().local(); /*auto&& keyspace = */db.find_keyspace(ks_name); #if 0 logger.info(String.format("Drop Keyspace '%s'", oldKsm.name)); announce(LegacySchemaTables.makeDropKeyspaceMutation(oldKsm, FBUtilities.timestampMicros()), announceLocally); #endif // FIXME throw std::runtime_error("not implemented"); } catch (const no_such_keyspace& e) { throw exceptions::configuration_exception(sprint("Cannot drop non existing keyspace '%s'.", ks_name)); } } future<> migration_manager::announce_column_family_drop(const sstring& ks_name, const sstring& cf_name, bool announce_locally) { try { auto& db = get_local_storage_proxy().get_db().local(); /*auto&& cfm = */db.find_schema(ks_name, cf_name); /*auto&& ksm = */db.find_keyspace(ks_name); #if 0 logger.info(String.format("Drop table '%s/%s'", oldCfm.ksName, oldCfm.cfName)); announce(LegacySchemaTables.makeDropTableMutation(ksm, oldCfm, FBUtilities.timestampMicros()), announceLocally); #endif // FIXME throw std::runtime_error("not implemented"); } catch (const no_such_column_family& e) { throw exceptions::configuration_exception(sprint("Cannot drop non existing table '%s' in keyspace '%s'.", cf_name, ks_name)); } } #if 0 public static void announceTypeDrop(UserType droppedType) { announceTypeDrop(droppedType, false); } public static void announceTypeDrop(UserType droppedType, boolean announceLocally) { KSMetaData ksm = Schema.instance.getKSMetaData(droppedType.keyspace); announce(LegacySchemaTables.dropTypeFromSchemaMutation(ksm, droppedType, FBUtilities.timestampMicros()), announceLocally); } public static void announceFunctionDrop(UDFunction udf, boolean announceLocally) { logger.info(String.format("Drop scalar function overload '%s' args '%s'", udf.name(), udf.argTypes())); KSMetaData ksm = Schema.instance.getKSMetaData(udf.name().keyspace); announce(LegacySchemaTables.makeDropFunctionMutation(ksm, udf, FBUtilities.timestampMicros()), announceLocally); } public static void announceAggregateDrop(UDAggregate udf, boolean announceLocally) { logger.info(String.format("Drop aggregate function overload '%s' args '%s'", udf.name(), udf.argTypes())); KSMetaData ksm = Schema.instance.getKSMetaData(udf.name().keyspace); announce(LegacySchemaTables.makeDropAggregateMutation(ksm, udf, FBUtilities.timestampMicros()), announceLocally); } #endif /** * actively announce a new version to active hosts via rpc * @param schema The schema mutation to be applied */ future<> migration_manager::announce(mutation schema, bool announce_locally) { std::vector mutations; mutations.emplace_back(std::move(schema)); return announce(std::move(mutations), announce_locally); } future<> migration_manager::announce(std::vector mutations, bool announce_locally) { if (announce_locally) { return db::schema_tables::merge_schema(get_local_storage_proxy(), std::move(mutations), false); } else { return announce(std::move(mutations)); } } future<> migration_manager::push_schema_mutation(const gms::inet_address& endpoint, const std::vector& schema) { net::messaging_service::shard_id id{endpoint, 0}; auto fm = std::vector(schema.begin(), schema.end()); return net::get_local_messaging_service().send_definitions_update(id, std::move(fm)); } // Returns a future on the local application of the schema future<> migration_manager::announce(std::vector schema) { return gms::get_live_members().then([schema = std::move(schema)](std::set live_members) { return do_with(std::move(schema), [live_members] (auto&& schema) { return parallel_for_each(live_members.begin(), live_members.end(), [&schema] (auto& endpoint) { // only push schema to nodes with known and equal versions if (endpoint != utils::fb_utilities::get_broadcast_address() && net::get_local_messaging_service().knows_version(endpoint) && net::get_local_messaging_service().get_raw_version(endpoint) == net::messaging_service::current_version) { return push_schema_mutation(endpoint, schema); } else { return make_ready_future<>(); } }).then_wrapped([] (future<> f) { try { f.get(); } catch (...) { return make_exception_future<>(std::current_exception()); } return make_ready_future<>(); }).then([&schema] { return db::schema_tables::merge_schema(get_local_storage_proxy(), std::move(schema)); }); }); }); } /** * Announce my version passively over gossip. * Used to notify nodes as they arrive in the cluster. * * @param version The schema version to announce */ future<> migration_manager::passive_announce(utils::UUID version) { return gms::get_gossiper().invoke_on(0, [version] (auto&& gossiper) { auto& ss = service::get_local_storage_service(); gossiper.add_local_application_state(gms::application_state::SCHEMA, ss.value_factory.schema(version)); logger.debug("Gossiping my schema version {}", version); return make_ready_future<>(); }); } #if 0 /** * Clear all locally stored schema information and reset schema to initial state. * Called by user (via JMX) who wants to get rid of schema disagreement. * * @throws IOException if schema tables truncation fails */ public static void resetLocalSchema() throws IOException { logger.info("Starting local schema reset..."); logger.debug("Truncating schema tables..."); LegacySchemaTables.truncateSchemaTables(); logger.debug("Clearing local schema keyspace definitions..."); Schema.instance.clear(); Set liveEndpoints = Gossiper.instance.getLiveMembers(); liveEndpoints.remove(FBUtilities.getBroadcastAddress()); // force migration if there are nodes around for (InetAddress node : liveEndpoints) { if (shouldPullSchemaFrom(node)) { logger.debug("Requesting schema from {}", node); FBUtilities.waitOnFuture(submitMigrationTask(node)); break; } } logger.info("Local schema reset is complete."); } public static class MigrationsSerializer implements IVersionedSerializer> { public static MigrationsSerializer instance = new MigrationsSerializer(); public void serialize(Collection schema, DataOutputPlus out, int version) throws IOException { out.writeInt(schema.size()); for (Mutation mutation : schema) Mutation.serializer.serialize(mutation, out, version); } public Collection deserialize(DataInput in, int version) throws IOException { int count = in.readInt(); Collection schema = new ArrayList<>(count); for (int i = 0; i < count; i++) schema.add(Mutation.serializer.deserialize(in, version)); return schema; } public long serializedSize(Collection schema, int version) { int size = TypeSizes.NATIVE.sizeof(schema.size()); for (Mutation mutation : schema) size += Mutation.serializer.serializedSize(mutation, version); return size; } } #endif }