/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* * Modified by ScyllaDB * Copyright (C) 2015 ScyllaDB */ /* * This file is part of Scylla. * * Scylla is free software: you can redistribute it and/or modify * it under the terms of the GNU Affero General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * * Scylla is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with Scylla. If not, see . */ #include "db/schema_tables.hh" #include "service/migration_manager.hh" #include "partition_slice_builder.hh" #include "dht/i_partitioner.hh" #include "system_keyspace.hh" #include "query_context.hh" #include "query-result-set.hh" #include "query-result-writer.hh" #include "schema_builder.hh" #include "map_difference.hh" #include "utils/UUID_gen.hh" #include "core/do_with.hh" #include "core/thread.hh" #include "json.hh" #include "log.hh" #include "frozen_schema.hh" #include "schema_registry.hh" #include "db/marshal/type_parser.hh" #include "db/config.hh" #include "md5_hasher.hh" #include #include #include "compaction_strategy.hh" #include "utils/joinpoint.hh" using namespace db::system_keyspace; using namespace std::chrono_literals; /** system.schema_* tables used to store keyspace/table/type attributes prior to C* 3.0 */ namespace db { namespace schema_tables { logging::logger logger("schema_tables"); struct qualified_name { sstring keyspace_name; sstring table_name; qualified_name(sstring keyspace_name, sstring table_name) : keyspace_name(std::move(keyspace_name)) , table_name(std::move(table_name)) { } qualified_name(const schema_ptr& s) : keyspace_name(s->ks_name()) , table_name(s->cf_name()) { } bool operator<(const qualified_name& o) const { return keyspace_name < o.keyspace_name || (keyspace_name == o.keyspace_name && table_name < o.table_name); } bool operator==(const qualified_name& o) const { return keyspace_name == o.keyspace_name && table_name == o.table_name; } }; static future read_table_mutations(distributed& proxy, const qualified_name& table); static void merge_tables(distributed& proxy, std::map&& before, std::map&& after); static void merge_types(distributed& proxy, schema_result&& before, schema_result&& after); std::vector ALL { KEYSPACES, COLUMNFAMILIES, COLUMNS, TRIGGERS, USERTYPES, /* not present in 2.1.8: FUNCTIONS, AGGREGATES */ }; using days = std::chrono::duration>; /* static */ schema_ptr keyspaces() { static thread_local auto keyspaces = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, KEYSPACES), NAME, KEYSPACES, // partition key {{"keyspace_name", utf8_type}}, // clustering key {}, // regular columns { {"durable_writes", boolean_type}, {"strategy_class", utf8_type}, {"strategy_options", utf8_type}, }, // static columns {}, // regular column name type utf8_type, // comment "keyspace definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::yes); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return keyspaces; } /* static */ schema_ptr columnfamilies() { static thread_local auto columnfamilies = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, COLUMNFAMILIES), NAME, COLUMNFAMILIES, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"columnfamily_name", utf8_type}}, // regular columns { {"bloom_filter_fp_chance", double_type}, {"caching", utf8_type}, {"cf_id", uuid_type}, {"comment", utf8_type}, {"compaction_strategy_class", utf8_type}, {"compaction_strategy_options", utf8_type}, {"comparator", utf8_type}, {"compression_parameters", utf8_type}, {"default_time_to_live", int32_type}, {"default_validator", utf8_type}, {"dropped_columns", map_type_impl::get_instance(utf8_type, long_type, true)}, {"gc_grace_seconds", int32_type}, {"is_dense", boolean_type}, {"key_validator", utf8_type}, {"local_read_repair_chance", double_type}, {"max_compaction_threshold", int32_type}, {"max_index_interval", int32_type}, {"memtable_flush_period_in_ms", int32_type}, {"min_compaction_threshold", int32_type}, {"min_index_interval", int32_type}, {"read_repair_chance", double_type}, {"speculative_retry", utf8_type}, {"subcomparator", utf8_type}, {"type", utf8_type}, // The following 4 columns are only present up until 2.1.8 tables {"key_aliases", utf8_type}, {"value_alias", utf8_type}, {"column_aliases", utf8_type}, {"index_interval", int32_type}, }, // static columns {}, // regular column name type utf8_type, // comment "table definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return columnfamilies; } /* static */ schema_ptr columns() { static thread_local auto columns = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, COLUMNS), NAME, COLUMNS, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"columnfamily_name", utf8_type}, {"column_name", utf8_type}}, // regular columns { {"component_index", int32_type}, {"index_name", utf8_type}, {"index_options", utf8_type}, {"index_type", utf8_type}, {"type", utf8_type}, {"validator", utf8_type}, }, // static columns {}, // regular column name type utf8_type, // comment "column definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return columns; } /* static */ schema_ptr triggers() { static thread_local auto triggers = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, TRIGGERS), NAME, TRIGGERS, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"columnfamily_name", utf8_type}, {"trigger_name", utf8_type}}, // regular columns { {"trigger_options", map_type_impl::get_instance(utf8_type, utf8_type, true)}, }, // static columns {}, // regular column name type utf8_type, // comment "trigger definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return triggers; } /* static */ schema_ptr usertypes() { static thread_local auto usertypes = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, USERTYPES), NAME, USERTYPES, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"type_name", utf8_type}}, // regular columns { {"field_names", list_type_impl::get_instance(utf8_type, true)}, {"field_types", list_type_impl::get_instance(utf8_type, true)}, }, // static columns {}, // regular column name type utf8_type, // comment "user defined type definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return usertypes; } /* static */ schema_ptr functions() { static thread_local auto functions = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, FUNCTIONS), NAME, FUNCTIONS, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"function_name", utf8_type}, {"signature", bytes_type}}, // regular columns { {"argument_names", list_type_impl::get_instance(utf8_type, true)}, {"argument_types", list_type_impl::get_instance(utf8_type, true)}, {"body", utf8_type}, {"is_deterministic", boolean_type}, {"language", utf8_type}, {"return_type", utf8_type}, }, // static columns {}, // regular column name type utf8_type, // comment "user defined type definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return functions; } /* static */ schema_ptr aggregates() { static thread_local auto aggregates = [] { schema_builder builder(make_lw_shared(schema(generate_legacy_id(NAME, AGGREGATES), NAME, AGGREGATES, // partition key {{"keyspace_name", utf8_type}}, // clustering key {{"aggregate_name", utf8_type}, {"signature", bytes_type}}, // regular columns { {"argument_types", list_type_impl::get_instance(utf8_type, true)}, {"final_func", utf8_type}, {"intercond", bytes_type}, {"return_type", utf8_type}, {"state_func", utf8_type}, {"state_type", utf8_type}, }, // static columns {}, // regular column name type utf8_type, // comment "user defined aggregate definitions" ))); builder.set_gc_grace_seconds(std::chrono::duration_cast(days(7)).count()); builder.with(schema_builder::compact_storage::no); builder.with_version(generate_schema_version(builder.uuid())); return builder.build(); }(); return aggregates; } /** add entries to system.schema_* for the hardcoded system definitions */ future<> save_system_keyspace_schema() { auto& ks = db::qctx->db().find_keyspace(db::system_keyspace::NAME); auto ksm = ks.metadata(); // delete old, possibly obsolete entries in schema tables return parallel_for_each(ALL, [ksm] (sstring cf) { auto deletion_timestamp = schema_creation_timestamp() - 1; return db::execute_cql(sprint("DELETE FROM system.%%s USING TIMESTAMP %s WHERE keyspace_name = ?", deletion_timestamp), cf, ksm->name()).discard_result(); }).then([ksm] { auto mvec = make_create_keyspace_mutations(ksm, schema_creation_timestamp(), true); return qctx->proxy().mutate_locally(std::move(mvec)); }); } #if 0 public static void truncateSchemaTables() { for (String table : ALL) getSchemaCFS(table).truncateBlocking(); } private static void flushSchemaTables() { for (String table : ALL) SystemKeyspace.forceBlockingFlush(table); } #endif /** * Read schema from system keyspace and calculate MD5 digest of every row, resulting digest * will be converted into UUID which would act as content-based version of the schema. */ future calculate_schema_digest(distributed& proxy) { auto map = [&proxy] (sstring table) { return db::system_keyspace::query_mutations(proxy, table).then([&proxy, table] (auto rs) { auto s = proxy.local().get_db().local().find_schema(system_keyspace::NAME, table); std::vector mutations; for (auto&& p : rs->partitions()) { auto mut = p.mut().unfreeze(s); auto partition_key = value_cast(utf8_type->deserialize(mut.key().get_component(*s, 0))); if (partition_key == system_keyspace::NAME) { continue; } mutations.emplace_back(std::move(mut)); } return mutations; }); }; auto reduce = [] (auto& hash, auto&& mutations) { for (const mutation& m : mutations) { feed_hash_for_schema_digest(hash, m); } }; return do_with(md5_hasher(), [map, reduce] (auto& hash) { return do_for_each(ALL.begin(), ALL.end(), [&hash, map, reduce] (auto& table) { return map(table).then([&hash, reduce] (auto&& mutations) { reduce(hash, mutations); }); }).then([&hash] { return make_ready_future(utils::UUID_gen::get_name_UUID(hash.finalize())); }); }); } future> convert_schema_to_mutations(distributed& proxy) { auto map = [&proxy] (sstring table) { return db::system_keyspace::query_mutations(proxy, table).then([&proxy, table] (auto rs) { auto s = proxy.local().get_db().local().find_schema(system_keyspace::NAME, table); std::vector results; for (auto&& p : rs->partitions()) { auto mut = p.mut().unfreeze(s); auto partition_key = value_cast(utf8_type->deserialize(mut.key().get_component(*s, 0))); if (partition_key == system_keyspace::NAME) { continue; } results.emplace_back(std::move(p.mut())); } return results; }); }; auto reduce = [] (auto&& result, auto&& mutations) { std::move(mutations.begin(), mutations.end(), std::back_inserter(result)); return std::move(result); }; return map_reduce(ALL.begin(), ALL.end(), map, std::vector{}, reduce); } future read_schema_for_keyspaces(distributed& proxy, const sstring& schema_table_name, const std::set& keyspace_names) { auto schema = proxy.local().get_db().local().find_schema(system_keyspace::NAME, schema_table_name); auto map = [&proxy, schema_table_name] (sstring keyspace_name) { return read_schema_partition_for_keyspace(proxy, schema_table_name, keyspace_name); }; auto insert = [] (schema_result&& result, auto&& schema_entity) { if (!schema_entity.second->empty()) { result.insert(std::move(schema_entity)); } return std::move(result); }; return map_reduce(keyspace_names.begin(), keyspace_names.end(), map, schema_result{}, insert); } static future query_partition_mutation(service::storage_proxy& proxy, schema_ptr s, lw_shared_ptr cmd, partition_key pkey) { auto dk = dht::global_partitioner().decorate_key(*s, pkey); return do_with(query::partition_range::make_singular(dk), [&proxy, dk, s = std::move(s), cmd = std::move(cmd)] (auto& range) { return proxy.query_mutations_locally(s, std::move(cmd), range) .then([dk = std::move(dk), s](foreign_ptr> res) { auto&& partitions = res->partitions(); if (partitions.size() == 0) { return mutation(std::move(dk), s); } else if (partitions.size() == 1) { return partitions[0].mut().unfreeze(s); } else { assert(false && "Results must have at most one partition"); } }); }); } future read_schema_partition_for_keyspace(distributed& proxy, const sstring& schema_table_name, const sstring& keyspace_name) { auto schema = proxy.local().get_db().local().find_schema(system_keyspace::NAME, schema_table_name); auto keyspace_key = dht::global_partitioner().decorate_key(*schema, partition_key::from_singular(*schema, keyspace_name)); return db::system_keyspace::query(proxy, schema_table_name, keyspace_key).then([keyspace_name] (auto&& rs) { return schema_result_value_type{keyspace_name, std::move(rs)}; }); } future read_schema_partition_for_table(distributed& proxy, const sstring& schema_table_name, const sstring& keyspace_name, const sstring& table_name) { auto schema = proxy.local().get_db().local().find_schema(system_keyspace::NAME, schema_table_name); auto keyspace_key = partition_key::from_singular(*schema, keyspace_name); auto clustering_range = query::clustering_range(clustering_key_prefix::from_clustering_prefix( *schema, exploded_clustering_prefix({utf8_type->decompose(table_name)}))); auto slice = partition_slice_builder(*schema) .with_range(std::move(clustering_range)) .build(); auto cmd = make_lw_shared(schema->id(), schema->version(), std::move(slice), query::max_rows); return query_partition_mutation(proxy.local(), std::move(schema), std::move(cmd), std::move(keyspace_key)); } static semaphore the_merge_lock; future<> merge_lock() { // ref: #1088 // to avoid deadlocks, we don't want long-standing calls to the shard 0 // as they can cause a deadlock: // // fiber1 fiber2 // merge_lock() (succeeds) // merge_lock() (waits) // invoke_on_all() (waits on merge_lock to relinquish smp::submit_to slot) // // so we issue the lock calls with a timeout; the slot will be relinquished, and invoke_on_all() // can complete return repeat([] () mutable { return smp::submit_to(0, [] { return the_merge_lock.try_wait(); }).then([] (bool result) { if (result) { return make_ready_future(stop_iteration::yes); } else { static thread_local auto rand_engine = std::default_random_engine(); auto dist = std::uniform_int_distribution(0, 100); auto to = std::chrono::microseconds(dist(rand_engine)); return sleep(to).then([] { return make_ready_future(stop_iteration::no); }); } }); }); } future<> merge_unlock() { return smp::submit_to(0, [] { the_merge_lock.signal(); }); } /** * Merge remote schema in form of mutations with local and mutate ks/cf metadata objects * (which also involves fs operations on add/drop ks/cf) * * @param mutations the schema changes to apply * * @throws ConfigurationException If one of metadata attributes has invalid value * @throws IOException If data was corrupted during transportation or failed to apply fs operations */ future<> merge_schema(distributed& proxy, std::vector mutations) { return merge_lock().then([&proxy, mutations = std::move(mutations)] () mutable { return do_merge_schema(proxy, std::move(mutations), true).then([&proxy] { return update_schema_version_and_announce(proxy); }); }).finally([] { return merge_unlock(); }); } future<> merge_schema(distributed& proxy, std::vector mutations, bool do_flush) { return merge_lock().then([&proxy, mutations = std::move(mutations), do_flush] () mutable { return do_merge_schema(proxy, std::move(mutations), do_flush); }).finally([] { return merge_unlock(); }); } // Returns names of live table definitions of given keyspace future> static read_table_names_of_keyspace(distributed& proxy, const sstring& keyspace_name) { auto s = columnfamilies(); auto pkey = dht::global_partitioner().decorate_key(*s, partition_key::from_singular(*s, keyspace_name)); return db::system_keyspace::query(proxy, COLUMNFAMILIES, pkey).then([] (auto&& rs) { std::vector result; for (const query::result_set_row& row : rs->rows()) { result.emplace_back(row.get_nonnull("columnfamily_name")); } return result; }); } // Call inside a seastar thread static std::map read_tables_for_keyspaces(distributed& proxy, const std::set& keyspace_names) { std::map result; for (auto&& keyspace_name : keyspace_names) { for (auto&& table_name : read_table_names_of_keyspace(proxy, keyspace_name).get0()) { auto qn = qualified_name(keyspace_name, table_name); result.emplace(qn, read_table_mutations(proxy, qn).get0()); } } return result; } future<> do_merge_schema(distributed& proxy, std::vector mutations, bool do_flush) { return seastar::async([&proxy, mutations = std::move(mutations), do_flush] () mutable { schema_ptr s = keyspaces(); // compare before/after schemas of the affected keyspaces only std::set keyspaces; std::set column_families; for (auto&& mutation : mutations) { keyspaces.emplace(value_cast(utf8_type->deserialize(mutation.key().get_component(*s, 0)))); column_families.emplace(mutation.column_family_id()); } // current state of the schema auto&& old_keyspaces = read_schema_for_keyspaces(proxy, KEYSPACES, keyspaces).get0(); auto&& old_column_families = read_tables_for_keyspaces(proxy, keyspaces); auto&& old_types = read_schema_for_keyspaces(proxy, USERTYPES, keyspaces).get0(); #if 0 // not in 2.1.8 /*auto& old_functions = */read_schema_for_keyspaces(proxy, FUNCTIONS, keyspaces).get0(); /*auto& old_aggregates = */read_schema_for_keyspaces(proxy, AGGREGATES, keyspaces).get0(); #endif proxy.local().mutate_locally(std::move(mutations)).get0(); if (do_flush) { proxy.local().get_db().invoke_on_all([s, cfs = std::move(column_families)] (database& db) { return parallel_for_each(cfs.begin(), cfs.end(), [&db] (auto& id) { auto& cf = db.find_column_family(id); return cf.flush(); }); }).get(); } // with new data applied auto&& new_keyspaces = read_schema_for_keyspaces(proxy, KEYSPACES, keyspaces).get0(); auto&& new_column_families = read_tables_for_keyspaces(proxy, keyspaces); auto&& new_types = read_schema_for_keyspaces(proxy, USERTYPES, keyspaces).get0(); #if 0 // not in 2.1.8 /*auto& new_functions = */read_schema_for_keyspaces(proxy, FUNCTIONS, keyspaces).get0(); /*auto& new_aggregates = */read_schema_for_keyspaces(proxy, AGGREGATES, keyspaces).get0(); #endif std::set keyspaces_to_drop = merge_keyspaces(proxy, std::move(old_keyspaces), std::move(new_keyspaces)).get0(); merge_tables(proxy, std::move(old_column_families), std::move(new_column_families)); merge_types(proxy, std::move(old_types), std::move(new_types)); #if 0 mergeFunctions(oldFunctions, newFunctions); mergeAggregates(oldAggregates, newAggregates); #endif proxy.local().get_db().invoke_on_all([keyspaces_to_drop = std::move(keyspaces_to_drop)] (database& db) { // it is safe to drop a keyspace only when all nested ColumnFamilies where deleted return do_for_each(keyspaces_to_drop, [&db] (auto keyspace_to_drop) { db.drop_keyspace(keyspace_to_drop); return service::get_local_migration_manager().notify_drop_keyspace(keyspace_to_drop); }); }).get0(); }); } future> merge_keyspaces(distributed& proxy, schema_result&& before, schema_result&& after) { std::vector created; std::vector altered; std::set dropped; /* * - we don't care about entriesOnlyOnLeft() or entriesInCommon(), because only the changes are of interest to us * - of all entriesOnlyOnRight(), we only care about ones that have live columns; it's possible to have a ColumnFamily * there that only has the top-level deletion, if: * a) a pushed DROP KEYSPACE change for a keyspace hadn't ever made it to this node in the first place * b) a pulled dropped keyspace that got dropped before it could find a way to this node * - of entriesDiffering(), we don't care about the scenario where both pre and post-values have zero live columns: * that means that a keyspace had been recreated and dropped, and the recreated keyspace had never found a way * to this node */ auto diff = difference(before, after, indirect_equal_to>()); for (auto&& key : diff.entries_only_on_left) { logger.info("Dropping keyspace {}", key); dropped.emplace(key); } for (auto&& key : diff.entries_only_on_right) { auto&& value = after[key]; logger.info("Creating keyspace {}", key); created.emplace_back(schema_result_value_type{key, std::move(value)}); } for (auto&& key : diff.entries_differing) { logger.info("Altering keyspace {}", key); altered.emplace_back(key); } return do_with(std::move(created), [&proxy, altered = std::move(altered)] (auto& created) mutable { return do_with(std::move(altered), [&proxy, &created](auto& altered) { return proxy.local().get_db().invoke_on_all([&created, &altered] (database& db) { return do_for_each(created, [&db](auto&& val) { auto ksm = create_keyspace_from_schema_partition(val); return db.create_keyspace(ksm).then([ksm] { return service::get_local_migration_manager().notify_create_keyspace(ksm); }); }).then([&altered, &db]() { return do_for_each(altered, [&db](auto& name) { return db.update_keyspace(name); }); }); }); }); }).then([dropped = std::move(dropped)] () { return make_ready_future>(dropped); }); } static future<> update_column_family(database& db, schema_ptr new_schema) { column_family& cfm = db.find_column_family(new_schema->id()); keyspace& ks = db.find_keyspace(new_schema->ks_name()); bool columns_changed = !cfm.schema()->equal_columns(*new_schema); auto s = local_schema_registry().learn(new_schema); s->registry_entry()->mark_synced(); cfm.set_schema(std::move(s)); ks.metadata()->add_or_update_column_family(new_schema); return service::get_local_migration_manager().notify_update_column_family(cfm.schema(), columns_changed); } // see the comments for merge_keyspaces() static void merge_tables(distributed& proxy, std::map&& before, std::map&& after) { struct dropped_table { global_schema_ptr schema; utils::joinpoint jp{[] { return make_ready_future(db_clock::now()); }}; }; std::vector created; std::vector altered; std::vector dropped; auto diff = difference(before, after); for (auto&& key : diff.entries_only_on_left) { auto&& s = proxy.local().get_db().local().find_schema(key.keyspace_name, key.table_name); logger.info("Dropping {}.{} id={} version={}", s->ks_name(), s->cf_name(), s->id(), s->version()); dropped.emplace_back(dropped_table{s}); } for (auto&& key : diff.entries_only_on_right) { auto s = create_table_from_mutations(after.at(key)); logger.info("Creating {}.{} id={} version={}", s->ks_name(), s->cf_name(), s->id(), s->version()); created.emplace_back(s); } for (auto&& key : diff.entries_differing) { auto s = create_table_from_mutations(after.at(key)); logger.info("Altering {}.{} id={} version={}", s->ks_name(), s->cf_name(), s->id(), s->version()); altered.emplace_back(s); } proxy.local().get_db().invoke_on_all([&created, &dropped, &altered] (database& db) { return seastar::async([&] { for (auto&& gs : created) { schema_ptr s = gs.get(); auto& ks = db.find_keyspace(s->ks_name()); auto cfg = ks.make_column_family_config(*s, db.get_config()); db.add_column_family(s, cfg); auto& cf = db.find_column_family(s); cf.mark_ready_for_writes(); ks.make_directory_for_column_family(s->cf_name(), s->id()).get(); service::get_local_migration_manager().notify_create_column_family(s).get(); } for (auto&& gs : altered) { update_column_family(db, gs.get()).get(); } parallel_for_each(dropped.begin(), dropped.end(), [&db](dropped_table& dt) { schema_ptr s = dt.schema.get(); return db.drop_column_family(s->ks_name(), s->cf_name(), [&dt] { return dt.jp.value(); }).then([s] { return service::get_local_migration_manager().notify_drop_column_family(s); }); }).get(); }); }).get(); } static inline void collect_types(std::set& keys, schema_result& result, std::vector& to) { for (auto&& key : keys) { auto&& value = result[key]; auto types = create_types_from_schema_partition(schema_result_value_type{key, std::move(value)}); std::move(types.begin(), types.end(), std::back_inserter(to)); } } static inline void ensure_type_is_unused(distributed& proxy, user_type type) { // We don't want to drop a type unless it's not used anymore (mainly because // if someone drops a type and recreates one with the same name but different // definition with the previous name still in use, things can get messy). // We have two places to check: 1) other user type that can nest the one // we drop and 2) existing tables referencing the type (maybe in a nested // way). auto&& keyspace = type->_keyspace; auto&& name = type->_name; auto&& db = proxy.local().get_db().local(); auto&& ks = db.find_keyspace(type->_keyspace); for (auto&& ut : ks.metadata()->user_types()->get_all_types() | boost::adaptors::map_values) { if (ut->_keyspace == keyspace && ut->_name == name) { continue; } if (ut->references_user_type(keyspace, name)) { throw exceptions::invalid_request_exception(sprint("Cannot drop user type %s.%s as it is still used by user type %s", keyspace, type->get_name_as_string(), ut->get_name_as_string())); } } for (auto&& cfm : ks.metadata()->cf_meta_data() | boost::adaptors::map_values) { for (auto&& col : cfm->all_columns() | boost::adaptors::map_values) { if (col->type->references_user_type(keyspace, name)) { throw exceptions::invalid_request_exception(sprint("Cannot drop user type %s.%s as it is still used by table %s.%s", keyspace, type->get_name_as_string(), cfm->ks_name(), cfm->cf_name())); } } } } // see the comments for merge_keyspaces() static void merge_types(distributed& proxy, schema_result&& before, schema_result&& after) { std::vector created, altered, dropped; auto diff = difference(before, after, indirect_equal_to>()); collect_types(diff.entries_only_on_left, before, dropped); // Keyspaces with no more types collect_types(diff.entries_only_on_right, after, created); // New keyspaces with types for (auto&& key : diff.entries_differing) { // The user types of this keyspace differ, so diff the current types with the updated ones auto current_types = proxy.local().get_db().local().find_keyspace(key).metadata()->user_types()->get_all_types(); decltype(current_types) updated_types; auto ts = create_types_from_schema_partition(schema_result_value_type{key, std::move(after[key])}); updated_types.reserve(ts.size()); for (auto&& type : ts) { updated_types[type->_name] = std::move(type); } auto delta = difference(current_types, updated_types, indirect_equal_to()); for (auto&& key : delta.entries_only_on_left) { dropped.emplace_back(current_types[key]); } for (auto&& key : delta.entries_only_on_right) { created.emplace_back(std::move(updated_types[key])); } for (auto&& key : delta.entries_differing) { altered.emplace_back(std::move(updated_types[key])); } } for (auto&& ut : dropped) { ensure_type_is_unused(proxy, ut); } proxy.local().get_db().invoke_on_all([&created, &dropped, &altered] (database& db) { return seastar::async([&] { for (auto&& type : created) { auto user_type = dynamic_pointer_cast(parse_type(type->name())); db.find_keyspace(user_type->_keyspace).add_user_type(user_type); service::get_local_migration_manager().notify_create_user_type(user_type).get(); } for (auto&& type : dropped) { auto user_type = dynamic_pointer_cast(parse_type(type->name())); db.find_keyspace(user_type->_keyspace).remove_user_type(user_type); service::get_local_migration_manager().notify_drop_user_type(user_type).get(); } for (auto&& type : altered) { auto user_type = dynamic_pointer_cast(parse_type(type->name())); db.find_keyspace(user_type->_keyspace).add_user_type(user_type); service::get_local_migration_manager().notify_update_user_type(user_type).get(); } }); }).get(); } #if 0 // see the comments for mergeKeyspaces() private static void mergeFunctions(Map before, Map after) { List created = new ArrayList<>(); List altered = new ArrayList<>(); List dropped = new ArrayList<>(); MapDifference diff = Maps.difference(before, after); // New keyspace with functions for (Map.Entry entry : diff.entriesOnlyOnRight().entrySet()) if (entry.getValue().hasColumns()) created.addAll(createFunctionsFromFunctionsPartition(new Row(entry.getKey(), entry.getValue())).values()); for (Map.Entry> entry : diff.entriesDiffering().entrySet()) { ColumnFamily pre = entry.getValue().leftValue(); ColumnFamily post = entry.getValue().rightValue(); if (pre.hasColumns() && post.hasColumns()) { MapDifference delta = Maps.difference(createFunctionsFromFunctionsPartition(new Row(entry.getKey(), pre)), createFunctionsFromFunctionsPartition(new Row(entry.getKey(), post))); dropped.addAll(delta.entriesOnlyOnLeft().values()); created.addAll(delta.entriesOnlyOnRight().values()); Iterables.addAll(altered, Iterables.transform(delta.entriesDiffering().values(), new Function, UDFunction>() { public UDFunction apply(MapDifference.ValueDifference pair) { return pair.rightValue(); } })); } else if (pre.hasColumns()) { dropped.addAll(createFunctionsFromFunctionsPartition(new Row(entry.getKey(), pre)).values()); } else if (post.hasColumns()) { created.addAll(createFunctionsFromFunctionsPartition(new Row(entry.getKey(), post)).values()); } } for (UDFunction udf : created) Schema.instance.addFunction(udf); for (UDFunction udf : altered) Schema.instance.updateFunction(udf); for (UDFunction udf : dropped) Schema.instance.dropFunction(udf); } // see the comments for mergeKeyspaces() private static void mergeAggregates(Map before, Map after) { List created = new ArrayList<>(); List altered = new ArrayList<>(); List dropped = new ArrayList<>(); MapDifference diff = Maps.difference(before, after); // New keyspace with functions for (Map.Entry entry : diff.entriesOnlyOnRight().entrySet()) if (entry.getValue().hasColumns()) created.addAll(createAggregatesFromAggregatesPartition(new Row(entry.getKey(), entry.getValue())).values()); for (Map.Entry> entry : diff.entriesDiffering().entrySet()) { ColumnFamily pre = entry.getValue().leftValue(); ColumnFamily post = entry.getValue().rightValue(); if (pre.hasColumns() && post.hasColumns()) { MapDifference delta = Maps.difference(createAggregatesFromAggregatesPartition(new Row(entry.getKey(), pre)), createAggregatesFromAggregatesPartition(new Row(entry.getKey(), post))); dropped.addAll(delta.entriesOnlyOnLeft().values()); created.addAll(delta.entriesOnlyOnRight().values()); Iterables.addAll(altered, Iterables.transform(delta.entriesDiffering().values(), new Function, UDAggregate>() { public UDAggregate apply(MapDifference.ValueDifference pair) { return pair.rightValue(); } })); } else if (pre.hasColumns()) { dropped.addAll(createAggregatesFromAggregatesPartition(new Row(entry.getKey(), pre)).values()); } else if (post.hasColumns()) { created.addAll(createAggregatesFromAggregatesPartition(new Row(entry.getKey(), post)).values()); } } for (UDAggregate udf : created) Schema.instance.addAggregate(udf); for (UDAggregate udf : altered) Schema.instance.updateAggregate(udf); for (UDAggregate udf : dropped) Schema.instance.dropAggregate(udf); } #endif /* * Keyspace metadata serialization/deserialization. */ std::vector make_create_keyspace_mutations(lw_shared_ptr keyspace, api::timestamp_type timestamp, bool with_tables_and_types_and_functions) { std::vector mutations; schema_ptr s = keyspaces(); auto pkey = partition_key::from_singular(*s, keyspace->name()); mutation m(pkey, s); exploded_clustering_prefix ckey; m.set_cell(ckey, "durable_writes", keyspace->durable_writes(), timestamp); m.set_cell(ckey, "strategy_class", keyspace->strategy_name(), timestamp); auto raw = json::to_json(keyspace->strategy_options()); m.set_cell(ckey, "strategy_options", raw, timestamp); mutations.emplace_back(std::move(m)); if (with_tables_and_types_and_functions) { for (auto&& kv : keyspace->user_types()->get_all_types()) { add_type_to_schema_mutation(kv.second, timestamp, mutations); } for (auto&& kv : keyspace->cf_meta_data()) { add_table_to_schema_mutation(kv.second, timestamp, true, mutations); } } return mutations; } std::vector make_drop_keyspace_mutations(lw_shared_ptr keyspace, api::timestamp_type timestamp) { std::vector mutations; for (auto&& schema_table : all_tables()) { auto pkey = partition_key::from_exploded(*schema_table, {utf8_type->decompose(keyspace->name())}); mutation m{pkey, schema_table}; m.partition().apply(tombstone{timestamp, gc_clock::now()}); mutations.emplace_back(std::move(m)); } auto&& schema = db::system_keyspace::built_indexes(); auto pkey = partition_key::from_exploded(*schema, {utf8_type->decompose(keyspace->name())}); mutation m{pkey, schema}; m.partition().apply(tombstone{timestamp, gc_clock::now()}); mutations.emplace_back(std::move(m)); return mutations; } /** * Deserialize only Keyspace attributes without nested tables or types * * @param partition Keyspace attributes in serialized form */ lw_shared_ptr create_keyspace_from_schema_partition(const schema_result_value_type& result) { auto&& rs = result.second; if (rs->empty()) { throw std::runtime_error("query result has no rows"); } auto&& row = rs->row(0); auto keyspace_name = row.get_nonnull("keyspace_name"); auto strategy_name = row.get_nonnull("strategy_class"); auto raw = row.get_nonnull("strategy_options"); std::map strategy_options = json::to_map(raw); bool durable_writes = row.get_nonnull("durable_writes"); return make_lw_shared(keyspace_name, strategy_name, strategy_options, durable_writes); } std::vector create_types_from_schema_partition(const schema_result_value_type& result) { std::vector user_types; user_types.reserve(result.second->rows().size()); for (auto&& row : result.second->rows()) { auto name = to_bytes(row.get_nonnull("type_name")); auto columns = row.get_nonnull("field_names"); std::vector field_names; for (auto&& value : columns) { field_names.emplace_back(to_bytes(value_cast(value))); } auto types = row.get_nonnull("field_types"); std::vector field_types; for (auto&& value : types) { field_types.emplace_back(parse_type(value_cast(value))); } user_types.emplace_back(user_type_impl::get_instance(result.first, name, field_names, field_types)); } return user_types; } /* * User type metadata serialization/deserialization */ template static atomic_cell_or_collection make_list_mutation(const std::vector& values, const column_definition* column, api::timestamp_type timestamp, std::function::value_type)> to_data_value) { assert(column); list_type_impl::mutation m; m.cells.reserve(values.size()); m.tomb.timestamp = timestamp - 1; m.tomb.deletion_time = gc_clock::now(); auto values_type = static_pointer_cast(column->type); for (auto&& value : values) { auto dv = to_data_value(value); auto uuid = utils::UUID_gen::get_time_UUID_bytes(); m.cells.emplace_back( bytes(reinterpret_cast(uuid.data()), uuid.size()), atomic_cell::make_live(timestamp, values_type->get_elements_type()->decompose(std::move(dv)))); } return atomic_cell_or_collection::from_collection_mutation(values_type->serialize_mutation_form(std::move(m))); } void add_type_to_schema_mutation(user_type type, api::timestamp_type timestamp, std::vector& mutations) { schema_ptr s = usertypes(); auto pkey = partition_key::from_singular(*s, type->_keyspace); auto ckey = clustering_key::from_singular(*s, type->get_name_as_string()); mutation m{pkey, s}; auto field_names_column = s->get_column_definition("field_names"); auto field_names = make_list_mutation(type->field_names(), field_names_column, timestamp, [](auto&& name) { return utf8_type->deserialize(name); }); m.set_clustered_cell(ckey, *field_names_column, std::move(field_names)); auto field_types_column = s->get_column_definition("field_types"); auto field_types = make_list_mutation(type->field_types(), field_types_column, timestamp, [](auto&& type) { return data_value(type->name()); }); m.set_clustered_cell(ckey, *field_types_column, std::move(field_types)); mutations.emplace_back(std::move(m)); } std::vector make_create_type_mutations(lw_shared_ptr keyspace, user_type type, api::timestamp_type timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). auto mutations = make_create_keyspace_mutations(keyspace, timestamp, false); add_type_to_schema_mutation(type, timestamp, mutations); return mutations; } std::vector make_drop_type_mutations(lw_shared_ptr keyspace, user_type type, api::timestamp_type timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). auto mutations = make_create_keyspace_mutations(keyspace, timestamp, false); schema_ptr s = usertypes(); auto pkey = partition_key::from_singular(*s, type->_keyspace); auto ckey = clustering_key::from_singular(*s, type->get_name_as_string()); mutation m{pkey, s}; m.partition().apply_delete(*s, ckey, tombstone(timestamp, gc_clock::now())); mutations.emplace_back(std::move(m)); return mutations; } /* * Table metadata serialization/deserialization. */ std::vector make_create_table_mutations(lw_shared_ptr keyspace, schema_ptr table, api::timestamp_type timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). auto mutations = make_create_keyspace_mutations(keyspace, timestamp, false); add_table_to_schema_mutation(table, timestamp, true, mutations); return mutations; } schema_mutations make_table_mutations(schema_ptr table, api::timestamp_type timestamp, bool with_columns_and_triggers) { // When adding new schema properties, don't set cells for default values so that // both old and new nodes will see the same version during rolling upgrades. // For property that can be null (and can be changed), we insert tombstones, to make sure // we don't keep a property the user has removed schema_ptr s = columnfamilies(); auto pkey = partition_key::from_singular(*s, table->ks_name()); mutation m{pkey, s}; auto ckey = clustering_key::from_singular(*s, table->cf_name()); m.set_clustered_cell(ckey, "cf_id", table->id(), timestamp); m.set_clustered_cell(ckey, "type", cf_type_to_sstring(table->type()), timestamp); if (table->is_super()) { warn(unimplemented::cause::SUPER); #if 0 // We need to continue saving the comparator and subcomparator separatly, otherwise // we won't know at deserialization if the subcomparator should be taken into account // TODO: we should implement an on-start migration if we want to get rid of that. adder.add("comparator", table.comparator.subtype(0).toString()); adder.add("subcomparator", table.comparator.subtype(1).toString()); #endif } else { m.set_clustered_cell(ckey, "comparator", cell_comparator::to_sstring(*table), timestamp); } m.set_clustered_cell(ckey, "bloom_filter_fp_chance", table->bloom_filter_fp_chance(), timestamp); m.set_clustered_cell(ckey, "caching", table->caching_options().to_sstring(), timestamp); m.set_clustered_cell(ckey, "comment", table->comment(), timestamp); m.set_clustered_cell(ckey, "compaction_strategy_class", sstables::compaction_strategy::name(table->compaction_strategy()), timestamp); m.set_clustered_cell(ckey, "compaction_strategy_options", json::to_json(table->compaction_strategy_options()), timestamp); const auto& compression_options = table->get_compressor_params(); m.set_clustered_cell(ckey, "compression_parameters", json::to_json(compression_options.get_options()), timestamp); m.set_clustered_cell(ckey, "default_time_to_live", table->default_time_to_live().count(), timestamp); m.set_clustered_cell(ckey, "default_validator", table->default_validator()->name(), timestamp); m.set_clustered_cell(ckey, "gc_grace_seconds", table->gc_grace_seconds().count(), timestamp); m.set_clustered_cell(ckey, "key_validator", table->thrift_key_validator(), timestamp); m.set_clustered_cell(ckey, "local_read_repair_chance", table->dc_local_read_repair_chance(), timestamp); m.set_clustered_cell(ckey, "min_compaction_threshold", table->min_compaction_threshold(), timestamp); m.set_clustered_cell(ckey, "max_compaction_threshold", table->max_compaction_threshold(), timestamp); m.set_clustered_cell(ckey, "min_index_interval", table->min_index_interval(), timestamp); m.set_clustered_cell(ckey, "max_index_interval", table->max_index_interval(), timestamp); m.set_clustered_cell(ckey, "memtable_flush_period_in_ms", table->memtable_flush_period(), timestamp); m.set_clustered_cell(ckey, "read_repair_chance", table->read_repair_chance(), timestamp); m.set_clustered_cell(ckey, "speculative_retry", table->speculative_retry().to_sstring(), timestamp); auto alias = [] (schema::const_iterator_range_type range) -> sstring { sstring alias("["); for (auto& c: range) { alias += "\"" + c.name_as_text() + "\","; } if (alias.back() == ',') { alias.back() = ']'; } else { alias += "]"; } return alias; }; m.set_clustered_cell(ckey, "key_aliases", alias(table->partition_key_columns()), timestamp); m.set_clustered_cell(ckey, "column_aliases", alias(table->clustering_key_columns()), timestamp); if (table->compact_columns_count() == 1) { m.set_clustered_cell(ckey, "value_alias", table->compact_column().name_as_text(), timestamp); } // null if none map_type_impl::mutation dropped_columns; auto dropped_columns_column = s->get_column_definition("dropped_columns"); assert(dropped_columns_column); auto dropped_columns_type = static_pointer_cast(dropped_columns_column->type); for (auto&& entry : table->dropped_columns()) { dropped_columns.cells.emplace_back(dropped_columns_type->get_keys_type()->decompose(data_value(entry.first)), atomic_cell::make_live(timestamp, dropped_columns_type->get_values_type()->decompose(entry.second))); } m.set_clustered_cell(ckey, *dropped_columns_column, atomic_cell_or_collection::from_collection_mutation(dropped_columns_type->serialize_mutation_form(std::move(dropped_columns)))); m.set_clustered_cell(ckey, "is_dense", table->is_dense(), timestamp); mutation columns_mutation(pkey, columns()); if (with_columns_and_triggers) { for (auto&& column : table->all_columns_in_select_order()) { add_column_to_schema_mutation(table, column, timestamp, columns_mutation); } #if 0 for (TriggerDefinition trigger : table.getTriggers().values()) addTriggerToSchemaMutation(table, trigger, timestamp, mutation); #endif } return schema_mutations{std::move(m), std::move(columns_mutation)}; } void add_table_to_schema_mutation(schema_ptr table, api::timestamp_type timestamp, bool with_columns_and_triggers, std::vector& mutations) { make_table_mutations(table, timestamp, with_columns_and_triggers).copy_to(mutations); } std::vector make_update_table_mutations(lw_shared_ptr keyspace, schema_ptr old_table, schema_ptr new_table, api::timestamp_type timestamp, bool from_thrift) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). auto mutations = make_create_keyspace_mutations(keyspace, timestamp, false); add_table_to_schema_mutation(new_table, timestamp, false, mutations); mutation columns_mutation(partition_key::from_singular(*columns(), old_table->ks_name()), columns()); auto diff = difference(old_table->all_columns(), new_table->all_columns()); // columns that are no longer needed for (auto&& name : diff.entries_only_on_left) { // Thrift only knows about the REGULAR ColumnDefinition type, so don't consider other type // are being deleted just because they are not here. const column_definition& column = *old_table->all_columns().at(name); if (from_thrift && !column.is_regular()) { continue; } drop_column_from_schema_mutation(old_table, column, timestamp, mutations); } // newly added columns and old columns with updated attributes for (auto&& name : boost::range::join(diff.entries_differing, diff.entries_only_on_right)) { const column_definition& column = *new_table->all_columns().at(name); add_column_to_schema_mutation(new_table, column, timestamp, columns_mutation); } mutations.emplace_back(std::move(columns_mutation)); warn(unimplemented::cause::TRIGGERS); #if 0 MapDifference triggerDiff = Maps.difference(oldTable.getTriggers(), newTable.getTriggers()); // dropped triggers for (TriggerDefinition trigger : triggerDiff.entriesOnlyOnLeft().values()) dropTriggerFromSchemaMutation(oldTable, trigger, timestamp, mutation); // newly created triggers for (TriggerDefinition trigger : triggerDiff.entriesOnlyOnRight().values()) addTriggerToSchemaMutation(newTable, trigger, timestamp, mutation); #endif return mutations; } std::vector make_drop_table_mutations(lw_shared_ptr keyspace, schema_ptr table, api::timestamp_type timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). auto mutations = make_create_keyspace_mutations(keyspace, timestamp, false); schema_ptr s = columnfamilies(); auto pkey = partition_key::from_singular(*s, table->ks_name()); auto ckey = clustering_key::from_singular(*s, table->cf_name()); mutation m{pkey, s}; m.partition().apply_delete(*s, ckey, tombstone(timestamp, gc_clock::now())); mutations.emplace_back(m); for (auto& column : table->all_columns_in_select_order()) { drop_column_from_schema_mutation(table, column, timestamp, mutations); } #if 0 for (TriggerDefinition trigger : table.getTriggers().values()) dropTriggerFromSchemaMutation(table, trigger, timestamp, mutation); // TODO: get rid of in #6717 ColumnFamily indexCells = mutation.addOrGet(SystemKeyspace.BuiltIndexes); for (String indexName : Keyspace.open(keyspace.name).getColumnFamilyStore(table.cfName).getBuiltIndexes()) indexCells.addTombstone(indexCells.getComparator().makeCellName(indexName), ldt, timestamp); #endif return mutations; } static future read_table_mutations(distributed& proxy, const qualified_name& table) { return read_schema_partition_for_table(proxy, COLUMNFAMILIES, table.keyspace_name, table.table_name) .then([&proxy, table] (mutation cf_m) { return read_schema_partition_for_table(proxy, COLUMNS, table.keyspace_name, table.table_name) .then([cf_m = std::move(cf_m)] (mutation col_m) { return schema_mutations{std::move(cf_m), std::move(col_m)}; }); #if 0 // FIXME: Row serializedTriggers = readSchemaPartitionForTable(TRIGGERS, ksName, cfName); try { for (TriggerDefinition trigger : createTriggersFromTriggersPartition(serializedTriggers)) cfm.addTriggerDefinition(trigger); } catch (InvalidRequestException e) { throw new RuntimeException(e); } #endif }); } future create_table_from_name(distributed& proxy, const sstring& keyspace, const sstring& table) { return do_with(qualified_name(keyspace, table), [&proxy] (auto&& qn) { return read_table_mutations(proxy, qn).then([qn] (schema_mutations sm) { if (!sm.live()) { throw std::runtime_error(sprint("%s:%s not found in the schema definitions keyspace.", qn.keyspace_name, qn.table_name)); } return create_table_from_mutations(std::move(sm)); }); }); } /** * Deserialize tables from low-level schema representation, all of them belong to the same keyspace * * @return map containing name of the table and its metadata for faster lookup */ future> create_tables_from_tables_partition(distributed& proxy, const schema_result::mapped_type& result) { auto tables = make_lw_shared>(); return parallel_for_each(result->rows().begin(), result->rows().end(), [&proxy, tables] (auto&& row) { return create_table_from_table_row(proxy, row).then([tables] (schema_ptr&& cfm) { tables->emplace(cfm->cf_name(), std::move(cfm)); }); }).then([tables] { return std::move(*tables); }); } #if 0 public static CFMetaData createTableFromTablePartitionAndColumnsPartition(Row serializedTable, Row serializedColumns) { String query = String.format("SELECT * FROM %s.%s", SystemKeyspace.NAME, COLUMNFAMILIES); return createTableFromTableRowAndColumnsPartition(QueryProcessor.resultify(query, serializedTable).one(), serializedColumns); } #endif /** * Deserialize table metadata from low-level representation * * @return Metadata deserialized from schema */ future create_table_from_table_row(distributed& proxy, const query::result_set_row& row) { auto ks_name = row.get_nonnull("keyspace_name"); auto cf_name = row.get_nonnull("columnfamily_name"); return create_table_from_name(proxy, ks_name, cf_name); } schema_ptr create_table_from_mutations(schema_mutations sm, std::experimental::optional version) { auto table_rs = query::result_set(sm.columnfamilies_mutation()); query::result_set_row table_row = table_rs.row(0); auto ks_name = table_row.get_nonnull("keyspace_name"); auto cf_name = table_row.get_nonnull("columnfamily_name"); auto id = table_row.get_nonnull("cf_id"); schema_builder builder{ks_name, cf_name, id}; #if 0 AbstractType rawComparator = TypeParser.parse(result.getString("comparator")); AbstractType subComparator = result.has("subcomparator") ? TypeParser.parse(result.getString("subcomparator")) : null; #endif cf_type cf = cf_type::standard; if (table_row.has("type")) { cf = sstring_to_cf_type(table_row.get_nonnull("type")); if (cf == cf_type::super) { fail(unimplemented::cause::SUPER); } } #if 0 AbstractType fullRawComparator = CFMetaData.makeRawAbstractType(rawComparator, subComparator); #endif std::vector column_defs = create_columns_from_column_rows( query::result_set(sm.columns_mutation()), ks_name, cf_name,/*, fullRawComparator, */ cf == cf_type::super); bool is_dense; if (table_row.has("is_dense")) { is_dense = table_row.get_nonnull("is_dense"); } else { // FIXME: // is_dense = CFMetaData.calculateIsDense(fullRawComparator, columnDefs); throw std::runtime_error(sprint("%s not implemented", __PRETTY_FUNCTION__)); } auto comparator = table_row.get_nonnull("comparator"); bool is_compound = cell_comparator::check_compound(comparator); builder.set_is_compound(is_compound); cell_comparator::read_collections(builder, comparator); #if 0 CellNameType comparator = CellNames.fromAbstractType(fullRawComparator, isDense); // if we are upgrading, we use id generated from names initially UUID cfId = result.has("cf_id") ? result.getUUID("cf_id") : CFMetaData.generateLegacyCfId(ksName, cfName); CFMetaData cfm = new CFMetaData(ksName, cfName, cfType, comparator, cfId); #endif builder.set_is_dense(is_dense); if (table_row.has("read_repair_chance")) { builder.set_read_repair_chance(table_row.get_nonnull("read_repair_chance")); } if (table_row.has("local_read_repair_chance")) { builder.set_dc_local_read_repair_chance(table_row.get_nonnull("local_read_repair_chance")); } if (table_row.has("gc_grace_seconds")) { builder.set_gc_grace_seconds(table_row.get_nonnull("gc_grace_seconds")); } if (table_row.has("default_validator")) { builder.set_default_validator(parse_type(table_row.get_nonnull("default_validator"))); } if (table_row.has("min_compaction_threshold")) { builder.set_min_compaction_threshold(table_row.get_nonnull("min_compaction_threshold")); } if (table_row.has("max_compaction_threshold")) { builder.set_max_compaction_threshold(table_row.get_nonnull("max_compaction_threshold")); } if (table_row.has("comment")) { builder.set_comment(table_row.get_nonnull("comment")); } if (table_row.has("memtable_flush_period_in_ms")) { builder.set_memtable_flush_period(table_row.get_nonnull("memtable_flush_period_in_ms")); } if (table_row.has("caching")) { builder.set_caching_options(caching_options::from_sstring(table_row.get_nonnull("caching"))); } if (table_row.has("default_time_to_live")) { builder.set_default_time_to_live(gc_clock::duration(table_row.get_nonnull("default_time_to_live"))); } if (table_row.has("speculative_retry")) { builder.set_speculative_retry(table_row.get_nonnull("speculative_retry")); } if (table_row.has("compaction_strategy_class")) { auto strategy = table_row.get_nonnull("compaction_strategy_class"); try { builder.set_compaction_strategy(sstables::compaction_strategy::type(strategy)); } catch (const exceptions::configuration_exception& e) { // If compaction strategy class isn't supported, fallback to size tiered. logger.warn("Falling back to size-tiered compaction strategy after the problem: {}", e.what()); builder.set_compaction_strategy(sstables::compaction_strategy_type::size_tiered); } } if (table_row.has("compaction_strategy_options")) { builder.set_compaction_strategy_options(json::to_map(table_row.get_nonnull("compaction_strategy_options"))); } auto comp_param = table_row.get_nonnull("compression_parameters"); compression_parameters cp(json::to_map(comp_param)); builder.set_compressor_params(cp); if (table_row.has("min_index_interval")) { builder.set_min_index_interval(table_row.get_nonnull("min_index_interval")); } else if (table_row.has("index_interval")) { // compatibility builder.set_min_index_interval(table_row.get_nonnull("index_interval")); } if (table_row.has("max_index_interval")) { builder.set_max_index_interval(table_row.get_nonnull("max_index_interval")); } if (table_row.has("bloom_filter_fp_chance")) { builder.set_bloom_filter_fp_chance(table_row.get_nonnull("bloom_filter_fp_chance")); } else { builder.set_bloom_filter_fp_chance(builder.get_bloom_filter_fp_chance()); } if (table_row.has("dropped_columns")) { auto map = table_row.get_nonnull("dropped_columns"); for (auto&& entry : map) { builder.without_column(value_cast(entry.first), value_cast(entry.second)); }; } for (auto&& cdef : column_defs) { builder.with_column(cdef); } if (version) { builder.with_version(*version); } else { builder.with_version(sm.digest()); } return builder.build(); } #if 0 private static Map convertDroppedColumns(Map raw) { Map converted = Maps.newHashMap(); for (Map.Entry entry : raw.entrySet()) converted.put(new ColumnIdentifier(entry.getKey(), true), entry.getValue()); return converted; } #endif /* * Column metadata serialization/deserialization. */ void add_column_to_schema_mutation(schema_ptr table, const column_definition& column, api::timestamp_type timestamp, mutation& m) { auto ckey = clustering_key::from_exploded(*m.schema(), {utf8_type->decompose(table->cf_name()), column.name()}); m.set_clustered_cell(ckey, "validator", column.type->name(), timestamp); m.set_clustered_cell(ckey, "type", serialize_kind(column.kind), timestamp); if (!column.is_on_all_components()) { m.set_clustered_cell(ckey, "component_index", int32_t(table->position(column)), timestamp); } #if 0 adder.add("index_name", column.getIndexName()); adder.add("index_type", column.getIndexType() == null ? null : column.getIndexType().toString()); adder.add("index_options", json(column.getIndexOptions())); #endif } sstring serialize_kind(column_kind kind) { switch (kind) { case column_kind::partition_key: return "partition_key"; case column_kind::clustering_key: return "clustering_key"; case column_kind::static_column: return "static"; case column_kind::regular_column: return "regular"; case column_kind::compact_column: return "compact_value"; default: throw std::invalid_argument("unknown column kind"); } } column_kind deserialize_kind(sstring kind) { if (kind == "partition_key") { return column_kind::partition_key; } else if (kind == "clustering_key") { return column_kind::clustering_key; } else if (kind == "static") { return column_kind::static_column; } else if (kind == "regular") { return column_kind::regular_column; } else if (kind == "compact_value") { return column_kind::compact_column; } else { throw std::invalid_argument("unknown column kind: " + kind); } } void drop_column_from_schema_mutation(schema_ptr table, const column_definition& column, long timestamp, std::vector& mutations) { schema_ptr s = columns(); auto pkey = partition_key::from_singular(*s, table->ks_name()); auto ckey = clustering_key::from_exploded(*s, {utf8_type->decompose(table->cf_name()), column.name()}); mutation m{pkey, s}; m.partition().apply_delete(*s, ckey, tombstone(timestamp, gc_clock::now())); mutations.emplace_back(m); } std::vector create_columns_from_column_rows(const query::result_set& rows, const sstring& keyspace, const sstring& table, /*, AbstractType rawComparator, */ bool is_super) { std::vector columns; for (auto&& row : rows.rows()) { columns.emplace_back(std::move(create_column_from_column_row(row, keyspace, table, /*, rawComparator, */ is_super))); } return columns; } column_definition create_column_from_column_row(const query::result_set_row& row, sstring keyspace, sstring table, /*, AbstractType rawComparator, */ bool is_super) { auto kind = deserialize_kind(row.get_nonnull("type")); column_id component_index = 0; if (row.has("component_index")) { // FIXME: We need to pass component_index to schema_builder // to ensure columns are instantiated in the correct order. component_index = row.get_nonnull("component_index"); } #if 0 else if (kind == ColumnDefinition.Kind.CLUSTERING_COLUMN && isSuper) componentIndex = 1; // A ColumnDefinition for super columns applies to the column component #endif #if 0 // Note: we save the column name as string, but we should not assume that it is an UTF8 name, we // we need to use the comparator fromString method AbstractType comparator = kind == ColumnDefinition.Kind.REGULAR ? getComponentComparator(rawComparator, componentIndex) : UTF8Type.instance; #endif auto name_opt = row.get("column_name"); sstring name = name_opt ? *name_opt : sstring(); auto validator = parse_type(row.get_nonnull("validator")); #if 0 IndexType indexType = null; if (row.has("index_type")) indexType = IndexType.valueOf(row.getString("index_type")); Map indexOptions = null; if (row.has("index_options")) indexOptions = fromJsonMap(row.getString("index_options")); String indexName = null; if (row.has("index_name")) indexName = row.getString("index_name"); #endif auto c = column_definition{utf8_type->decompose(name), validator, kind, component_index}; return c; } #if 0 private static AbstractType getComponentComparator(AbstractType rawComparator, Integer componentIndex) { return (componentIndex == null || (componentIndex == 0 && !(rawComparator instanceof CompositeType))) ? rawComparator : ((CompositeType)rawComparator).types.get(componentIndex); } /* * Trigger metadata serialization/deserialization. */ private static void addTriggerToSchemaMutation(CFMetaData table, TriggerDefinition trigger, long timestamp, Mutation mutation) { ColumnFamily cells = mutation.addOrGet(Triggers); Composite prefix = Triggers.comparator.make(table.cfName, trigger.name); CFRowAdder adder = new CFRowAdder(cells, prefix, timestamp); adder.addMapEntry("trigger_options", "class", trigger.classOption); } private static void dropTriggerFromSchemaMutation(CFMetaData table, TriggerDefinition trigger, long timestamp, Mutation mutation) { ColumnFamily cells = mutation.addOrGet(Triggers); int ldt = (int) (System.currentTimeMillis() / 1000); Composite prefix = Triggers.comparator.make(table.cfName, trigger.name); cells.addAtom(new RangeTombstone(prefix, prefix.end(), timestamp, ldt)); } /** * Deserialize triggers from storage-level representation. * * @param partition storage-level partition containing the trigger definitions * @return the list of processed TriggerDefinitions */ private static List createTriggersFromTriggersPartition(Row partition) { List triggers = new ArrayList<>(); String query = String.format("SELECT * FROM %s.%s", SystemKeyspace.NAME, TRIGGERS); for (UntypedResultSet.Row row : QueryProcessor.resultify(query, partition)) { String name = row.getString("trigger_name"); String classOption = row.getMap("trigger_options", UTF8Type.instance, UTF8Type.instance).get("class"); triggers.add(new TriggerDefinition(name, classOption)); } return triggers; } /* * UDF metadata serialization/deserialization. */ public static Mutation makeCreateFunctionMutation(KSMetaData keyspace, UDFunction function, long timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). Mutation mutation = makeCreateKeyspaceMutation(keyspace, timestamp, false); addFunctionToSchemaMutation(function, timestamp, mutation); return mutation; } private static void addFunctionToSchemaMutation(UDFunction function, long timestamp, Mutation mutation) { ColumnFamily cells = mutation.addOrGet(Functions); Composite prefix = Functions.comparator.make(function.name().name, UDHelper.calculateSignature(function)); CFRowAdder adder = new CFRowAdder(cells, prefix, timestamp); adder.resetCollection("argument_names"); adder.resetCollection("argument_types"); for (int i = 0; i < function.argNames().size(); i++) { adder.addListEntry("argument_names", function.argNames().get(i).bytes); adder.addListEntry("argument_types", function.argTypes().get(i).toString()); } adder.add("body", function.body()); adder.add("is_deterministic", function.isDeterministic()); adder.add("language", function.language()); adder.add("return_type", function.returnType().toString()); } public static Mutation makeDropFunctionMutation(KSMetaData keyspace, UDFunction function, long timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). Mutation mutation = makeCreateKeyspaceMutation(keyspace, timestamp, false); ColumnFamily cells = mutation.addOrGet(Functions); int ldt = (int) (System.currentTimeMillis() / 1000); Composite prefix = Functions.comparator.make(function.name().name, UDHelper.calculateSignature(function)); cells.addAtom(new RangeTombstone(prefix, prefix.end(), timestamp, ldt)); return mutation; } private static Map createFunctionsFromFunctionsPartition(Row partition) { Map functions = new HashMap<>(); String query = String.format("SELECT * FROM %s.%s", SystemKeyspace.NAME, FUNCTIONS); for (UntypedResultSet.Row row : QueryProcessor.resultify(query, partition)) { UDFunction function = createFunctionFromFunctionRow(row); functions.put(UDHelper.calculateSignature(function), function); } return functions; } private static UDFunction createFunctionFromFunctionRow(UntypedResultSet.Row row) { String ksName = row.getString("keyspace_name"); String functionName = row.getString("function_name"); FunctionName name = new FunctionName(ksName, functionName); List argNames = new ArrayList<>(); if (row.has("argument_names")) for (String arg : row.getList("argument_names", UTF8Type.instance)) argNames.add(new ColumnIdentifier(arg, true)); List> argTypes = new ArrayList<>(); if (row.has("argument_types")) for (String type : row.getList("argument_types", UTF8Type.instance)) argTypes.add(parseType(type)); AbstractType returnType = parseType(row.getString("return_type")); boolean isDeterministic = row.getBoolean("is_deterministic"); String language = row.getString("language"); String body = row.getString("body"); try { return UDFunction.create(name, argNames, argTypes, returnType, language, body, isDeterministic); } catch (InvalidRequestException e) { logger.error(String.format("Cannot load function '%s' from schema: this function won't be available (on this node)", name), e); return UDFunction.createBrokenFunction(name, argNames, argTypes, returnType, language, body, e); } } /* * Aggregate UDF metadata serialization/deserialization. */ public static Mutation makeCreateAggregateMutation(KSMetaData keyspace, UDAggregate aggregate, long timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). Mutation mutation = makeCreateKeyspaceMutation(keyspace, timestamp, false); addAggregateToSchemaMutation(aggregate, timestamp, mutation); return mutation; } private static void addAggregateToSchemaMutation(UDAggregate aggregate, long timestamp, Mutation mutation) { ColumnFamily cells = mutation.addOrGet(Aggregates); Composite prefix = Aggregates.comparator.make(aggregate.name().name, UDHelper.calculateSignature(aggregate)); CFRowAdder adder = new CFRowAdder(cells, prefix, timestamp); adder.resetCollection("argument_types"); adder.add("return_type", aggregate.returnType().toString()); adder.add("state_func", aggregate.stateFunction().name().name); if (aggregate.stateType() != null) adder.add("state_type", aggregate.stateType().toString()); if (aggregate.finalFunction() != null) adder.add("final_func", aggregate.finalFunction().name().name); if (aggregate.initialCondition() != null) adder.add("initcond", aggregate.initialCondition()); for (AbstractType argType : aggregate.argTypes()) adder.addListEntry("argument_types", argType.toString()); } private static Map createAggregatesFromAggregatesPartition(Row partition) { Map aggregates = new HashMap<>(); String query = String.format("SELECT * FROM %s.%s", SystemKeyspace.NAME, AGGREGATES); for (UntypedResultSet.Row row : QueryProcessor.resultify(query, partition)) { UDAggregate aggregate = createAggregateFromAggregateRow(row); aggregates.put(UDHelper.calculateSignature(aggregate), aggregate); } return aggregates; } private static UDAggregate createAggregateFromAggregateRow(UntypedResultSet.Row row) { String ksName = row.getString("keyspace_name"); String functionName = row.getString("aggregate_name"); FunctionName name = new FunctionName(ksName, functionName); List types = row.getList("argument_types", UTF8Type.instance); List> argTypes; if (types == null) { argTypes = Collections.emptyList(); } else { argTypes = new ArrayList<>(types.size()); for (String type : types) argTypes.add(parseType(type)); } AbstractType returnType = parseType(row.getString("return_type")); FunctionName stateFunc = new FunctionName(ksName, row.getString("state_func")); FunctionName finalFunc = row.has("final_func") ? new FunctionName(ksName, row.getString("final_func")) : null; AbstractType stateType = row.has("state_type") ? parseType(row.getString("state_type")) : null; ByteBuffer initcond = row.has("initcond") ? row.getBytes("initcond") : null; try { return UDAggregate.create(name, argTypes, returnType, stateFunc, finalFunc, stateType, initcond); } catch (InvalidRequestException reason) { return UDAggregate.createBroken(name, argTypes, returnType, initcond, reason); } } public static Mutation makeDropAggregateMutation(KSMetaData keyspace, UDAggregate aggregate, long timestamp) { // Include the serialized keyspace in case the target node missed a CREATE KEYSPACE migration (see CASSANDRA-5631). Mutation mutation = makeCreateKeyspaceMutation(keyspace, timestamp, false); ColumnFamily cells = mutation.addOrGet(Aggregates); int ldt = (int) (System.currentTimeMillis() / 1000); Composite prefix = Aggregates.comparator.make(aggregate.name().name, UDHelper.calculateSignature(aggregate)); cells.addAtom(new RangeTombstone(prefix, prefix.end(), timestamp, ldt)); return mutation; } #endif data_type parse_type(sstring str) { return db::marshal::type_parser::parse(str); } std::vector all_tables() { return { keyspaces(), columnfamilies(), columns(), triggers(), usertypes(), /* Not in 2.1.8 functions(), aggregates() */ }; } } // namespace schema_tables } // namespace schema