Files
scylladb/cql3/statements/modification_statement.cc
Tomasz Grabiec ad99e84505 storage_proxy: Take schema_ptr in query()
It will be needed for reconciliation.
2015-07-12 12:54:38 +02:00

530 lines
23 KiB
C++

/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* Copyright 2015 Cloudius Systems
*
* Modified by Cloudius Systems
*/
#include "cql3/statements/modification_statement.hh"
#include "cql3/restrictions/single_column_restriction.hh"
#include "cql3/single_column_relation.hh"
#include "validation.hh"
#include "core/shared_ptr.hh"
#include "query-result-reader.hh"
#include <boost/range/adaptor/transformed.hpp>
#include <boost/range/algorithm_ext/push_back.hpp>
#include <boost/range/adaptor/filtered.hpp>
namespace cql3 {
namespace statements {
thread_local const ::shared_ptr<column_identifier> modification_statement::CAS_RESULT_COLUMN = ::make_shared<column_identifier>("[applied]", false);
std::ostream&
operator<<(std::ostream& out, modification_statement::statement_type t) {
switch (t) {
case modification_statement::statement_type::UPDATE:
out << "UPDATE";
break;
case modification_statement::statement_type::INSERT:
out << "INSERT";
break;
case modification_statement::statement_type::DELETE:
out << "DELETE";
break;
}
return out;
}
future<std::vector<mutation>>
modification_statement::get_mutations(distributed<service::storage_proxy>& proxy, const query_options& options, bool local, int64_t now) {
auto keys = make_lw_shared(build_partition_keys(options));
auto prefix = make_lw_shared(create_exploded_clustering_prefix(options));
return make_update_parameters(proxy, keys, prefix, options, local, now).then(
[this, keys, prefix, now] (auto params_ptr) {
std::vector<mutation> mutations;
mutations.reserve(keys->size());
for (auto key : *keys) {
mutations.emplace_back(std::move(key), s);
auto& m = mutations.back();
this->add_update_for_key(m, *prefix, *params_ptr);
}
return make_ready_future<decltype(mutations)>(std::move(mutations));
});
}
future<std::unique_ptr<update_parameters>>
modification_statement::make_update_parameters(
distributed<service::storage_proxy>& proxy,
lw_shared_ptr<std::vector<partition_key>> keys,
lw_shared_ptr<exploded_clustering_prefix> prefix,
const query_options& options,
bool local,
int64_t now) {
return read_required_rows(proxy, std::move(keys), std::move(prefix), local, options.get_consistency()).then(
[this, &options, now] (auto rows) {
return make_ready_future<std::unique_ptr<update_parameters>>(
std::make_unique<update_parameters>(s, options,
this->get_timestamp(now, options),
this->get_time_to_live(options),
std::move(rows)));
});
}
// Implements ResultVisitor concept from query.hh
class prefetch_data_builder {
update_parameters::prefetch_data& _data;
const query::partition_slice& _ps;
std::experimental::optional<partition_key> _pkey;
public:
prefetch_data_builder(update_parameters::prefetch_data& data, const query::partition_slice& ps)
: _data(data)
, _ps(ps)
{ }
void accept_new_partition(const partition_key& key, uint32_t row_count) {
_pkey = key;
}
void accept_new_partition(uint32_t row_count) {
assert(0);
}
void accept_new_row(const clustering_key& key, const query::result_row_view& static_row,
const query::result_row_view& row) {
update_parameters::prefetch_data::row cells;
auto add_cell = [&cells] (column_id id, std::experimental::optional<collection_mutation::view>&& cell) {
if (cell) {
cells.emplace(id, collection_mutation::one{to_bytes(cell->data)});
}
};
auto static_row_iterator = static_row.iterator();
for (auto&& id : _ps.static_columns) {
add_cell(id, static_row_iterator.next_collection_cell());
}
auto row_iterator = row.iterator();
for (auto&& id : _ps.regular_columns) {
add_cell(id, row_iterator.next_collection_cell());
}
_data.rows.emplace(std::make_pair(*_pkey, key), std::move(cells));
}
void accept_new_row(const query::result_row_view& static_row, const query::result_row_view& row) {
assert(0);
}
void accept_partition_end(const query::result_row_view& static_row) {}
};
future<update_parameters::prefetched_rows_type>
modification_statement::read_required_rows(
distributed<service::storage_proxy>& proxy,
lw_shared_ptr<std::vector<partition_key>> keys,
lw_shared_ptr<exploded_clustering_prefix> prefix,
bool local,
db::consistency_level cl) {
if (!requires_read()) {
return make_ready_future<update_parameters::prefetched_rows_type>(
update_parameters::prefetched_rows_type{});
}
try {
validate_for_read(keyspace(), cl);
} catch (exceptions::invalid_request_exception& e) {
throw exceptions::invalid_request_exception(sprint("Write operation require a read but consistency %s is not supported on reads", cl));
}
static auto is_collection = [] (const column_definition& def) {
return def.type->is_collection();
};
// FIXME: we read all collection columns, but could be enhanced just to read the list(s) being RMWed
std::vector<column_id> static_cols;
boost::range::push_back(static_cols, s->static_columns()
| boost::adaptors::filtered(is_collection) | boost::adaptors::transformed([] (auto&& col) { return col.id; }));
std::vector<column_id> regular_cols;
boost::range::push_back(regular_cols, s->regular_columns()
| boost::adaptors::filtered(is_collection) | boost::adaptors::transformed([] (auto&& col) { return col.id; }));
query::partition_slice ps(
{query::clustering_range(clustering_key_prefix::from_clustering_prefix(*s, *prefix))},
std::move(static_cols),
std::move(regular_cols),
query::partition_slice::option_set::of<
query::partition_slice::option::send_partition_key,
query::partition_slice::option::send_clustering_key>());
std::vector<query::partition_range> pr;
for (auto&& pk : *keys) {
pr.emplace_back(dht::global_partitioner().decorate_key(*s, pk));
}
query::read_command cmd(s->id(), ps, std::numeric_limits<uint32_t>::max());
// FIXME: ignoring "local"
return proxy.local().query(s, make_lw_shared(std::move(cmd)), std::move(pr), cl).then([this, ps] (auto result) {
// FIXME: copying
// FIXME: Use scattered_reader to avoid copying
bytes_ostream buf(result->buf());
query::result_view v(buf.linearize());
auto prefetched_rows = update_parameters::prefetched_rows_type({update_parameters::prefetch_data(s)});
v.consume(ps, prefetch_data_builder(prefetched_rows.value(), ps));
return prefetched_rows;
});
}
const column_definition*
modification_statement::get_first_empty_key() {
for (auto& def : s->clustering_key_columns()) {
if (_processed_keys.find(&def) == _processed_keys.end()) {
return &def;
}
}
return {};
}
exploded_clustering_prefix
modification_statement::create_exploded_clustering_prefix_internal(const query_options& options) {
std::vector<bytes> components;
const column_definition* first_empty_key = nullptr;
for (auto& def : s->clustering_key_columns()) {
auto i = _processed_keys.find(&def);
if (i == _processed_keys.end()) {
first_empty_key = &def;
// Tomek: Origin had "&& s->comparator->is_composite()" in the condition below.
// Comparator is a thrift concept, not CQL concept, and we want to avoid
// using thrift concepts here. I think it's safe to drop this here because the only
// case in which we would get a non-composite comparator here would be if the cell
// name type is SimpleSparse, which means:
// (a) CQL compact table without clustering columns
// (b) thrift static CF with non-composite comparator
// Those tables don't have clustering columns so we wouldn't reach this code, thus
// the check seems redundant.
if (require_full_clustering_key() && !s->is_dense()) {
throw exceptions::invalid_request_exception(sprint("Missing mandatory PRIMARY KEY part %s", def.name_as_text()));
}
} else if (first_empty_key) {
throw exceptions::invalid_request_exception(sprint("Missing PRIMARY KEY part %s since %s is set", first_empty_key->name_as_text(), def.name_as_text()));
} else {
auto values = i->second->values(options);
assert(values.size() == 1);
auto val = values[0];
if (!val) {
throw exceptions::invalid_request_exception(sprint("Invalid null value for clustering key part %s", def.name_as_text()));
}
components.push_back(*val);
}
}
return exploded_clustering_prefix(std::move(components));
}
exploded_clustering_prefix
modification_statement::create_exploded_clustering_prefix(const query_options& options) {
// If the only updated/deleted columns are static, then we don't need clustering columns.
// And in fact, unless it is an INSERT, we reject if clustering columns are provided as that
// suggest something unintended. For instance, given:
// CREATE TABLE t (k int, v int, s int static, PRIMARY KEY (k, v))
// it can make sense to do:
// INSERT INTO t(k, v, s) VALUES (0, 1, 2)
// but both
// UPDATE t SET s = 3 WHERE k = 0 AND v = 1
// DELETE v FROM t WHERE k = 0 AND v = 1
// sounds like you don't really understand what your are doing.
if (_sets_static_columns && !_sets_regular_columns) {
// If we set no non-static columns, then it's fine not to have clustering columns
if (_has_no_clustering_columns) {
return {};
}
// If we do have clustering columns however, then either it's an INSERT and the query is valid
// but we still need to build a proper prefix, or it's not an INSERT, and then we want to reject
// (see above)
if (type != statement_type::INSERT) {
for (auto& def : s->clustering_key_columns()) {
if (_processed_keys.count(&def)) {
throw exceptions::invalid_request_exception(sprint(
"Invalid restriction on clustering column %s since the %s statement modifies only static columns",
def.name_as_text(), type));
}
}
// we should get there as it contradicts _has_no_clustering_columns == false
throw std::logic_error("contradicts _has_no_clustering_columns == false");
}
}
return create_exploded_clustering_prefix_internal(options);
}
std::vector<partition_key>
modification_statement::build_partition_keys(const query_options& options) {
std::vector<partition_key> result;
std::vector<bytes> components;
auto remaining = s->partition_key_size();
for (auto& def : s->partition_key_columns()) {
auto i = _processed_keys.find(&def);
if (i == _processed_keys.end()) {
throw exceptions::invalid_request_exception(sprint("Missing mandatory PRIMARY KEY part %s", def.name_as_text()));
}
auto values = i->second->values(options);
if (remaining == 1) {
if (values.size() == 1) {
auto val = values[0];
if (!val) {
throw exceptions::invalid_request_exception(sprint("Invalid null value for partition key part %s", def.name_as_text()));
}
components.push_back(*val);
auto key = partition_key::from_exploded(*s, components);
validation::validate_cql_key(s, key);
result.emplace_back(std::move(key));
} else {
for (auto&& val : values) {
if (!val) {
throw exceptions::invalid_request_exception(sprint("Invalid null value for partition key part %s", def.name_as_text()));
}
std::vector<bytes> full_components;
full_components.reserve(components.size() + 1);
auto i = std::copy(components.begin(), components.end(), std::back_inserter(full_components));
*i = *val;
auto key = partition_key::from_exploded(*s, full_components);
validation::validate_cql_key(s, key);
result.emplace_back(std::move(key));
}
}
} else {
if (values.size() != 1) {
throw exceptions::invalid_request_exception("IN is only supported on the last column of the partition key");
}
auto val = values[0];
if (!val) {
throw exceptions::invalid_request_exception(sprint("Invalid null value for partition key part %s", def.name_as_text()));
}
components.push_back(*val);
}
remaining--;
}
return result;
}
future<::shared_ptr<transport::messages::result_message>>
modification_statement::execute(distributed<service::storage_proxy>& proxy, service::query_state& qs, const query_options& options) {
if (has_conditions() && options.get_protocol_version() == 1) {
throw new exceptions::invalid_request_exception("Conditional updates are not supported by the protocol version in use. You need to upgrade to a driver using the native protocol v2.");
}
if (has_conditions()) {
return execute_with_condition(proxy, qs, options);
}
return execute_without_condition(proxy, qs, options).then([] {
return make_ready_future<::shared_ptr<transport::messages::result_message>>(
::shared_ptr<transport::messages::result_message>{});
});
}
future<>
modification_statement::execute_without_condition(distributed<service::storage_proxy>& proxy, service::query_state& qs, const query_options& options) {
auto cl = options.get_consistency();
if (is_counter()) {
db::validate_counter_for_write(s, cl);
} else {
db::validate_for_write(s->ks_name(), cl);
}
return get_mutations(proxy, options, false, options.get_timestamp(qs)).then([cl, &proxy] (auto mutations) {
if (mutations.empty()) {
return now();
}
return proxy.local().mutate_with_triggers(std::move(mutations), cl, false);
});
}
future<::shared_ptr<transport::messages::result_message>>
modification_statement::execute_with_condition(distributed<service::storage_proxy>& proxy, service::query_state& qs, const query_options& options) {
fail(unimplemented::cause::LWT);
#if 0
List<ByteBuffer> keys = buildPartitionKeyNames(options);
// We don't support IN for CAS operation so far
if (keys.size() > 1)
throw new InvalidRequestException("IN on the partition key is not supported with conditional updates");
ByteBuffer key = keys.get(0);
long now = options.getTimestamp(queryState);
Composite prefix = createClusteringPrefix(options);
CQL3CasRequest request = new CQL3CasRequest(cfm, key, false);
addConditions(prefix, request, options);
request.addRowUpdate(prefix, this, options, now);
ColumnFamily result = StorageProxy.cas(keyspace(),
columnFamily(),
key,
request,
options.getSerialConsistency(),
options.getConsistency(),
queryState.getClientState());
return new ResultMessage.Rows(buildCasResultSet(key, result, options));
#endif
}
future<::shared_ptr<transport::messages::result_message>>
modification_statement::execute_internal(distributed<service::storage_proxy>& proxy, service::query_state& qs, const query_options& options) {
if (has_conditions()) {
throw exceptions::unsupported_operation_exception();
}
return get_mutations(proxy, options, true, options.get_timestamp(qs)).then(
[&proxy] (auto mutations) {
return proxy.local().mutate_locally(mutations);
}).then(
[] {
return make_ready_future<::shared_ptr<transport::messages::result_message>>(
::shared_ptr<transport::messages::result_message> {});
});
}
void
modification_statement::add_key_values(const column_definition& def, ::shared_ptr<restrictions::restriction> values) {
if (def.is_clustering_key()) {
_has_no_clustering_columns = false;
}
auto insert_result = _processed_keys.insert({&def, values});
if (!insert_result.second) {
throw exceptions::invalid_request_exception(sprint("Multiple definitions found for PRIMARY KEY part %s", def.name_as_text()));
}
}
void
modification_statement::add_key_value(const column_definition& def, ::shared_ptr<term> value) {
add_key_values(def, ::make_shared<restrictions::single_column_restriction::EQ>(def, value));
}
void
modification_statement::process_where_clause(database& db, std::vector<relation_ptr> where_clause, ::shared_ptr<variable_specifications> names) {
for (auto&& relation : where_clause) {
if (relation->is_multi_column()) {
throw exceptions::invalid_request_exception(sprint("Multi-column relations cannot be used in WHERE clauses for UPDATE and DELETE statements: %s", relation->to_string()));
}
auto rel = dynamic_pointer_cast<single_column_relation>(relation);
if (rel->on_token()) {
throw exceptions::invalid_request_exception(sprint("The token function cannot be used in WHERE clauses for UPDATE and DELETE statements: %s", relation->to_string()));
}
auto id = rel->get_entity()->prepare_column_identifier(s);
auto def = get_column_definition(s, *id);
if (!def) {
throw exceptions::invalid_request_exception(sprint("Unknown key identifier %s", *id));
}
if (def->is_primary_key()) {
if (rel->is_EQ() || (def->is_partition_key() && rel->is_IN())) {
add_key_values(*def, rel->to_restriction(db, s, names));
} else {
throw exceptions::invalid_request_exception(sprint("Invalid operator %s for PRIMARY KEY part %s", rel->get_operator(), def->name_as_text()));
}
} else {
throw exceptions::invalid_request_exception(sprint("Non PRIMARY KEY %s found in where clause", def->name_as_text()));
}
}
}
::shared_ptr<parsed_statement::prepared>
modification_statement::parsed::prepare(database& db) {
auto bound_names = get_bound_variables();
auto statement = prepare(db, bound_names);
return ::make_shared<parsed_statement::prepared>(std::move(statement), *bound_names);
}
::shared_ptr<modification_statement>
modification_statement::parsed::prepare(database& db, ::shared_ptr<variable_specifications> bound_names) {
schema_ptr schema = validation::validate_column_family(db, keyspace(), column_family());
auto prepared_attributes = _attrs->prepare(db, keyspace(), column_family());
prepared_attributes->collect_marker_specification(bound_names);
::shared_ptr<modification_statement> stmt = prepare_internal(db, schema, bound_names, std::move(prepared_attributes));
if (_if_not_exists || _if_exists || !_conditions.empty()) {
if (stmt->is_counter()) {
throw exceptions::invalid_request_exception("Conditional updates are not supported on counter tables");
}
if (_attrs->timestamp) {
throw exceptions::invalid_request_exception("Cannot provide custom timestamp for conditional updates");
}
if (_if_not_exists) {
// To have both 'IF NOT EXISTS' and some other conditions doesn't make sense.
// So far this is enforced by the parser, but let's assert it for sanity if ever the parse changes.
assert(_conditions.empty());
assert(!_if_exists);
stmt->set_if_not_exist_condition();
} else if (_if_exists) {
assert(_conditions.empty());
assert(!_if_not_exists);
stmt->set_if_exist_condition();
} else {
for (auto&& entry : _conditions) {
auto id = entry.first->prepare_column_identifier(schema);
const column_definition* def = get_column_definition(schema, *id);
if (!def) {
throw exceptions::invalid_request_exception(sprint("Unknown identifier %s", *id));
}
auto condition = entry.second->prepare(db, keyspace(), *def);
condition->collect_marker_specificaton(bound_names);
if (def->is_primary_key()) {
throw exceptions::invalid_request_exception(sprint("PRIMARY KEY column '%s' cannot have IF conditions", *id));
}
stmt->add_condition(condition);
}
}
stmt->validate_where_clause_for_conditions();
}
return stmt;
}
void
modification_statement::validate(distributed<service::storage_proxy>&, const service::client_state& state) {
if (has_conditions() && attrs->is_timestamp_set()) {
throw exceptions::invalid_request_exception("Cannot provide custom timestamp for conditional updates");
}
if (is_counter() && attrs->is_timestamp_set()) {
throw exceptions::invalid_request_exception("Cannot provide custom timestamp for counter updates");
}
if (is_counter() && attrs->is_time_to_live_set()) {
throw exceptions::invalid_request_exception("Cannot provide custom TTL for counter updates");
}
}
}
}