seastar::compat::source_location (which should not have been used
outside Seastar) is replaced with std::source_location to avoid
deprecation warnings. The relevant header, which was removed, is no
longer included.
* seastar 8c3fba7a...b5c76d6b (3):
> testing: There can be only one memory_data_sink
> util: Use std::source_location directly
> Merge 'net: support proxy protocol v2' from Avi Kivity
apps: httpd: add --load-balancing-algorithm
apps: httpd: add /shard endpoint
test: socket_test: add proxy protocol v2 test suite
test: socket_test: test load balancer with proxy protocol
net: posix_connected_socket: specialize for proxied connections
net: posix_server_socket_impl: implement proxy protocol in server sockets
net: posix_server_socket_impl: adjust indentation
net: posix_server_socket_impl: avoid immediately-invoked lambda
net: conntrack: complete handle nested class special member functions
net: posix_server_socket_impl: coroutinize accept()
Closes scylladb/scylladb#27316
6118 lines
317 KiB
C++
6118 lines
317 KiB
C++
/*
|
|
* Copyright (C) 2015-present-2020 ScyllaDB
|
|
*/
|
|
|
|
/*
|
|
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
|
|
*/
|
|
|
|
|
|
#include <boost/test/unit_test.hpp>
|
|
#include <boost/multiprecision/cpp_int.hpp>
|
|
|
|
#include <fmt/ranges.h>
|
|
#include <fmt/std.h>
|
|
|
|
#include <seastar/net/inet_address.hh>
|
|
|
|
#undef SEASTAR_TESTING_MAIN
|
|
#include <seastar/testing/test_case.hh>
|
|
#include <seastar/testing/thread_test_case.hh>
|
|
#include "test/lib/cql_test_env.hh"
|
|
#include "test/lib/cql_assertions.hh"
|
|
#include "test/lib/log.hh"
|
|
#include "test/lib/test_utils.hh"
|
|
|
|
#include <seastar/core/future-util.hh>
|
|
#include <seastar/core/sleep.hh>
|
|
#include "transport/messages/result_message.hh"
|
|
#include "transport/messages/result_message_base.hh"
|
|
#include "types/types.hh"
|
|
#include "utils/assert.hh"
|
|
#include "utils/big_decimal.hh"
|
|
#include "types/map.hh"
|
|
#include "types/list.hh"
|
|
#include "types/set.hh"
|
|
#include "types/vector.hh"
|
|
#include "db/config.hh"
|
|
#include "db/extensions.hh"
|
|
#include "cql3/cql_config.hh"
|
|
#include "test/lib/exception_utils.hh"
|
|
#include "service/qos/qos_common.hh"
|
|
#include "utils/rjson.hh"
|
|
#include "schema/schema_builder.hh"
|
|
#include "service/migration_manager.hh"
|
|
#include <boost/regex.hpp>
|
|
#include "service/qos/qos_common.hh"
|
|
#include "utils/UUID_gen.hh"
|
|
#include "tombstone_gc_extension.hh"
|
|
#include "db/tags/extension.hh"
|
|
#include "cdc/cdc_extension.hh"
|
|
#include "db/paxos_grace_seconds_extension.hh"
|
|
#include "db/per_partition_rate_limit_extension.hh"
|
|
#include "replica/schema_describe_helper.hh"
|
|
#include "sstables/sstables.hh"
|
|
#include "replica/distributed_loader.hh"
|
|
#include "compaction/compaction_manager.hh"
|
|
|
|
|
|
BOOST_AUTO_TEST_SUITE(cql_query_test)
|
|
|
|
using namespace std::literals::chrono_literals;
|
|
using namespace tests;
|
|
|
|
SEASTAR_TEST_CASE(test_create_keyspace_statement) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks2 with replication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };").get();
|
|
BOOST_REQUIRE(e.local_db().has_keyspace("ks2"));
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_create_table_statement) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table users (user_name varchar PRIMARY KEY, birth_year bigint);").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "users"));
|
|
e.execute_cql("create table cf (id int primary key, m map<int, int>, s set<text>, l list<uuid>);").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "cf"));
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_create_table_with_id_statement) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a))").get();
|
|
auto id = e.local_db().find_schema("ks", "tbl")->id();
|
|
e.execute_cql("DROP TABLE tbl").get();
|
|
BOOST_REQUIRE_THROW(e.execute_cql("SELECT * FROM tbl").get(), std::exception);
|
|
e.execute_cql(
|
|
format("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a)) WITH id='{}'", id)).get();
|
|
assert_that(e.execute_cql("SELECT * FROM tbl").get())
|
|
.is_rows().with_size(0);
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql(format("CREATE TABLE tbl2 (a int, b int, PRIMARY KEY (a)) WITH id='{}'", id)).get(),
|
|
exceptions::invalid_request_exception);
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b int, PRIMARY KEY (a)) WITH id='55'").get(),
|
|
exceptions::configuration_exception);
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql("ALTER TABLE tbl WITH id='f2a8c099-e723-48cb-8cd9-53e647a011a3'").get(),
|
|
exceptions::configuration_exception);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_create_twcs_table_no_ttl) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
// Create a TWCS table with no TTL defined
|
|
e.execute_cql("UPDATE system.config SET value='warn' WHERE name='restrict_twcs_without_default_ttl';").get();
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'MINUTES'};").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl"));
|
|
// Ensure ALTER TABLE works
|
|
e.execute_cql("ALTER TABLE tbl WITH default_time_to_live=60;").get();
|
|
// LiveUpdate and enforce TTL to be defined
|
|
e.execute_cql("UPDATE system.config SET value='true' WHERE name='restrict_twcs_without_default_ttl';").get();
|
|
// default_time_to_live option is required
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE TABLE tbl2 (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'MINUTES'};").get(), exceptions::configuration_exception);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'MINUTES'} AND "
|
|
"default_time_to_live=60").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl2"));
|
|
// default_time_to_live option must not be set to 0.
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl WITH default_time_to_live=0;").get(), exceptions::configuration_exception);
|
|
// LiveUpdate and disable the check, then try table creation again
|
|
e.execute_cql("UPDATE system.config SET value='false' WHERE name='restrict_twcs_without_default_ttl';").get();
|
|
e.execute_cql("CREATE TABLE tbl3 (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'MINUTES'};").get();
|
|
// LiveUpdate back, and ensure that unrelated CQL requests are able to get through
|
|
e.execute_cql("UPDATE system.config SET value='true' WHERE name='restrict_twcs_without_default_ttl';").get();
|
|
e.execute_cql("ALTER TABLE tbl3 WITH gc_grace_seconds=0;").get();
|
|
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl3"));
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_twcs_max_window) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
// Hardcode restriction to max 10 windows
|
|
e.execute_cql("UPDATE system.config SET value='10' WHERE name='twcs_max_window_count';").get();
|
|
// Creating a TWCS table with a large number of windows/buckets should fail
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'HOURS'} "
|
|
"AND default_time_to_live=86400;").get(), exceptions::configuration_exception);
|
|
// However the creation of a table within bounds should succeed
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'HOURS'} "
|
|
"AND default_time_to_live=36000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl"));
|
|
// LiveUpdate - Disable check
|
|
e.execute_cql("UPDATE system.config SET value='0' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b int, PRIMARY KEY (a)) WITH "
|
|
"compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '1', 'compaction_window_unit': 'HOURS'} "
|
|
"AND default_time_to_live=864000000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl2"));
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_twcs_restrictions_mixed) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
// Hardcode restriction to max 10 windows
|
|
e.execute_cql("UPDATE system.config SET value='10' WHERE name='twcs_max_window_count';").get();
|
|
|
|
// Scenario 1: STCS->TWCS with no TTL defined
|
|
e.execute_cql("CREATE TABLE tbl (a int PRIMARY KEY, b int);").get();
|
|
e.execute_cql("ALTER TABLE tbl WITH compaction = {'class': 'TimeWindowCompactionStrategy'};").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl"));
|
|
|
|
// Scenario 2: STCS->TWCS with small TTL. Note: TWCS default window size is 1 day (86400s)
|
|
e.execute_cql("CREATE TABLE tbl2 (a int PRIMARY KEY, b int) WITH default_time_to_live = 60;").get();
|
|
e.execute_cql("ALTER TABLE tbl2 WITH compaction = {'class': 'TimeWindowCompactionStrategy'};").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl2"));
|
|
|
|
// Scenario 3: STCS->TWCS with large TTL value
|
|
e.execute_cql("CREATE TABLE tbl3 (a int PRIMARY KEY, b int) WITH default_time_to_live = 8640000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl3"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl3 WITH compaction = {'class': 'TimeWindowCompactionStrategy'};").get(), exceptions::configuration_exception);
|
|
|
|
// Scenario 4: TWCS table with small to large TTL
|
|
e.execute_cql("CREATE TABLE tbl4 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'} AND default_time_to_live = 60;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl4"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl4 WITH default_time_to_live = 86400000;").get(), exceptions::configuration_exception);
|
|
|
|
// Scenario 5: No TTL TWCS to large TTL and then small TTL
|
|
e.execute_cql("CREATE TABLE tbl5 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'}").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl5"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl5 WITH default_time_to_live = 86400000;").get(), exceptions::configuration_exception);
|
|
e.execute_cql("ALTER TABLE tbl5 WITH default_time_to_live = 60;").get();
|
|
|
|
// Scenario 6: twcs_max_window_count LiveUpdate - Decrease TTL
|
|
e.execute_cql("UPDATE system.config SET value='0' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("CREATE TABLE tbl6 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'} AND default_time_to_live = 86400000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl6"));
|
|
e.execute_cql("UPDATE system.config SET value='50' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("ALTER TABLE tbl6 WITH default_time_to_live = 60;").get();
|
|
|
|
// Scenario 7: twcs_max_window_count LiveUpdate - Switch CompactionStrategy
|
|
e.execute_cql("UPDATE system.config SET value='0' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("CREATE TABLE tbl7 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'} AND default_time_to_live = 86400000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl7"));
|
|
e.execute_cql("UPDATE system.config SET value='50' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("ALTER TABLE tbl7 WITH compaction = {'class': 'SizeTieredCompactionStrategy'}").get();
|
|
|
|
// Scenario 8: No TTL TWCS table to STCS
|
|
e.execute_cql("CREATE TABLE tbl8 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'};").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks" ,"tbl8"));
|
|
e.execute_cql("ALTER TABLE tbl8 WITH compaction = {'class': 'SizeTieredCompactionStrategy'};").get();
|
|
|
|
// Scenario 9: Large TTL TWCS table, modify attribute other than compaction and default_time_to_live
|
|
e.execute_cql("UPDATE system.config SET value='0' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("CREATE TABLE tbl9 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy'} AND default_time_to_live = 86400000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl9"));
|
|
e.execute_cql("UPDATE system.config SET value='50' WHERE name='twcs_max_window_count';").get();
|
|
e.execute_cql("ALTER TABLE tbl9 WITH gc_grace_seconds = 0;").get();
|
|
|
|
// Scenario 10: Large TTL STCS table, fail to switch to TWCS with no TTL
|
|
e.execute_cql("CREATE TABLE tbl10 (a int PRIMARY KEY, b int) WITH default_time_to_live = 8640000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl10"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl10 WITH compaction = {'class': 'TimeWindowCompactionStrategy'};").get(), exceptions::configuration_exception);
|
|
e.execute_cql("ALTER TABLE tbl10 WITH compaction = {'class': 'TimeWindowCompactionStrategy'} AND default_time_to_live = 0;").get();
|
|
|
|
// Scenario 11: Ensure default_time_to_live updates reference existing table properties
|
|
e.execute_cql("CREATE TABLE tbl11 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy', 'compaction_window_size': '1', "
|
|
"'compaction_window_unit': 'MINUTES'} AND default_time_to_live=3000;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl11"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl11 WITH default_time_to_live=3600;").get(), exceptions::configuration_exception);
|
|
e.execute_cql("ALTER TABLE tbl11 WITH compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': '2', 'compaction_window_unit': 'MINUTES'};").get();
|
|
e.execute_cql("ALTER TABLE tbl11 WITH default_time_to_live=3600;").get();
|
|
|
|
// Scenario 12: Ensure that window sizes <= 0 are forbidden
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE TABLE tbl12 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy', 'compaction_window_size': '0'};").get(), exceptions::configuration_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE TABLE tbl12 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy', 'compaction_window_size': -65535};").get(), exceptions::configuration_exception);
|
|
e.execute_cql("CREATE TABLE tbl12 (a int PRIMARY KEY, b int) WITH compaction = "
|
|
"{'class': 'TimeWindowCompactionStrategy', 'compaction_window_size': 1};").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tbl12"));
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE tbl12 WITH compaction = {'class': 'TimeWindowCompactionStrategy', "
|
|
"'compaction_window_size': 0};").get(), exceptions::configuration_exception);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_drop_table_with_si_and_mv) {
|
|
cql_test_config cfg;
|
|
cfg.need_remote_proxy = true;
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, c float, PRIMARY KEY (a))").get();
|
|
e.execute_cql("CREATE INDEX idx1 ON tbl (b)").get();
|
|
e.execute_cql("CREATE INDEX idx2 ON tbl (c)").get();
|
|
e.execute_cql("CREATE MATERIALIZED VIEW tbl_view AS SELECT c FROM tbl WHERE c IS NOT NULL PRIMARY KEY (c, a)").get();
|
|
// dropping a table with materialized views is prohibited
|
|
assert_that_failed(e.execute_cql("DROP TABLE tbl"));
|
|
e.execute_cql("DROP MATERIALIZED VIEW tbl_view").get();
|
|
// dropping a table with secondary indexes is fine
|
|
e.execute_cql("DROP TABLE tbl").get();
|
|
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, c float, PRIMARY KEY (a))").get();
|
|
e.execute_cql("CREATE INDEX idx1 ON tbl (b)").get();
|
|
e.execute_cql("CREATE INDEX idx2 ON tbl (c)").get();
|
|
e.execute_cql("CREATE MATERIALIZED VIEW tbl_view AS SELECT c FROM tbl WHERE c IS NOT NULL PRIMARY KEY (c, a)").get();
|
|
// dropping whole keyspace with MV and SI is fine too
|
|
e.execute_cql("DROP KEYSPACE ks").get();
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_list_elements_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql = fmt::format("INSERT INTO tbl (a, b) VALUES(1, ['{}'])", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (a int, b list<date>, PRIMARY KEY (a))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b list<text>, PRIMARY KEY (a))").get();
|
|
auto id = e.prepare("INSERT INTO tbl2 (a, b) VALUES(?, ?)").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_list_type = list_type_impl::get_instance(utf8_type, true);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
auto values = my_list_type->decompose(make_list_value(my_list_type, {value}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_set_elements_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql = fmt::format("INSERT INTO tbl (a, b) VALUES(1, {{'{}'}})", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (a int, b set<date>, PRIMARY KEY (a))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b set<text>, PRIMARY KEY (a))").get();
|
|
auto id = e.prepare("INSERT INTO tbl2 (a, b) VALUES(?, ?)").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_set_type = set_type_impl::get_instance(utf8_type, true);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
auto values = my_set_type->decompose(make_set_value(my_set_type, {value}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_map_elements_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql1 = fmt::format("INSERT INTO tbl (a, b) VALUES(1, {{'10-10-2010' : '{}'}})", value);
|
|
auto cql2 = fmt::format("INSERT INTO tbl (a, b) VALUES(1, {{'{}' : '10-10-2010'}})", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql1).get(), exceptions::invalid_request_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql2).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql1).get());
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql2).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (a int, b map<date, date>, PRIMARY KEY (a))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b map<text, text>, PRIMARY KEY (a))").get();
|
|
auto id = e.prepare("INSERT INTO tbl2 (a, b) VALUES(?, ?)").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_map_type = map_type_impl::get_instance(utf8_type, utf8_type, true);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
auto values =
|
|
my_map_type->decompose(make_map_value(my_map_type,
|
|
{std::make_pair(value, sstring("foo"))}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
raw_values.pop_back();
|
|
values =
|
|
my_map_type->decompose(make_map_value(my_map_type,
|
|
{std::make_pair(sstring("foo"), value)}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
e.execute_prepared(id, raw_values).get();
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_in_clause_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql = fmt::format("SELECT r1 FROM tbl WHERE (c1,r1) IN ((1, '{}')) ALLOW FILTERING", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (p1 int, c1 int, r1 date, PRIMARY KEY (p1, c1,r1))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (p1 int, c1 int, r1 text, PRIMARY KEY (p1, c1,r1))").get();
|
|
auto id = e.prepare("SELECT r1 FROM tbl2 WHERE (c1,r1) IN ? ALLOW FILTERING").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_tuple_type = tuple_type_impl::get_instance({int32_type, utf8_type});
|
|
auto my_list_type = list_type_impl::get_instance(my_tuple_type, true);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
auto values = make_tuple_value(my_tuple_type, tuple_type_impl::native_type({int32_t{2}, value}));
|
|
auto in_values_list = my_list_type->decompose(make_list_value(my_list_type, {values}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(in_values_list));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_in_clause_cartesian_product_limits) {
|
|
do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tab1 (pk1 int, pk2 int, PRIMARY KEY ((pk1, pk2)))").get();
|
|
|
|
// 100 partitions, should pass
|
|
e.execute_cql("SELECT * FROM tab1 WHERE pk1 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)"
|
|
" AND pk2 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)").get();
|
|
// 110 partitions, should fail
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql("SELECT * FROM tab1 WHERE pk1 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10)"
|
|
" AND pk2 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)").get(),
|
|
std::runtime_error);
|
|
|
|
e.execute_cql("CREATE TABLE tab2 (pk1 int, ck1 int, ck2 int, PRIMARY KEY (pk1, ck1, ck2))").get();
|
|
|
|
// 100 clustering rows, should pass
|
|
e.execute_cql("SELECT * FROM tab2 WHERE pk1 = 1"
|
|
" AND ck1 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)"
|
|
" AND ck2 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)").get();
|
|
// 110 clustering rows, should fail
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql("SELECT * FROM tab2 WHERE pk1 = 1"
|
|
" AND ck1 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10)"
|
|
" AND ck2 IN (0, 1, 2, 3, 4, 5, 6, 7, 8, 9)").get(),
|
|
std::runtime_error);
|
|
auto make_tuple = [] (unsigned count) -> sstring {
|
|
std::ostringstream os;
|
|
os << "(0";
|
|
for (unsigned i = 1; i < count; ++i) {
|
|
os << "," << i;
|
|
}
|
|
os << ")";
|
|
return os.str();
|
|
};
|
|
e.execute_cql("CREATE TABLE tab3 (pk1 int, ck1 int, PRIMARY KEY (pk1, ck1))").get();
|
|
// tuple with 100 keys, should pass
|
|
e.execute_cql(fmt::format("SELECT * FROM tab3 WHERE pk1 IN {}", make_tuple(100))).get();
|
|
e.execute_cql(fmt::format("SELECT * FROM tab3 WHERE pk1 = 1 AND ck1 IN {}", make_tuple(100))).get();
|
|
// tuple with 101 keys, should fail
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql(fmt::format("SELECT * FROM tab3 WHERE pk1 IN {}", make_tuple(101))).get(),
|
|
std::runtime_error
|
|
);
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql(fmt::format("SELECT * FROM tab3 WHERE pk1 = 3 AND ck1 IN {}", make_tuple(101))).get(),
|
|
std::runtime_error
|
|
);
|
|
}).get();
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_tuple_elements_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql = fmt::format("INSERT INTO tbl (a, b) VALUES(1, (1, '{}'))", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (a int, b tuple<int, date>, PRIMARY KEY (a))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b tuple<int, text>, PRIMARY KEY (a))").get();
|
|
auto id = e.prepare("INSERT INTO tbl2 (a, b) VALUES(?, ?)").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_tuple_type = tuple_type_impl::get_instance({int32_type, utf8_type});
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
auto values = my_tuple_type->decompose(make_tuple_value(my_tuple_type, tuple_type_impl::native_type({int32_t{2}, value})));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_vector_elements_validation) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
auto test_inline = [&] (sstring value, bool should_throw) {
|
|
auto cql = fmt::format("INSERT INTO tbl (a, b) VALUES(1, ['{}'])", value);
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql).get(), exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_cql(cql).get());
|
|
}
|
|
};
|
|
e.execute_cql("CREATE TABLE tbl (a int, b vector<date, 1>, PRIMARY KEY (a))").get();
|
|
test_inline("definitely not a date value", true);
|
|
test_inline("2015-05-03", false);
|
|
e.execute_cql("CREATE TABLE tbl2 (a int, b vector<text, 1>, PRIMARY KEY (a))").get();
|
|
auto id = e.prepare("INSERT INTO tbl2 (a, b) VALUES(?, ?)").get();
|
|
auto test_bind = [&] (sstring value, bool should_throw) {
|
|
auto my_vector_type = vector_type_impl::get_instance(utf8_type, 1);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
auto values = my_vector_type->decompose(make_vector_value(my_vector_type, {value}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(values));
|
|
if (should_throw) {
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(id, raw_values).get(),
|
|
exceptions::invalid_request_exception);
|
|
} else {
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(id, raw_values).get());
|
|
}
|
|
};
|
|
test_bind(sstring(1, '\255'), true);
|
|
test_bind("proper utf8 string", false);
|
|
});
|
|
}
|
|
|
|
/// Reproduces #4209
|
|
SEASTAR_TEST_CASE(test_list_of_tuples_with_bound_var) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create table cf (pk int PRIMARY KEY, c1 list<frozen<tuple<int,int>>>);").get();
|
|
e.prepare("update cf SET c1 = c1 + [(?,9999)] where pk = 999;").get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bound_var_in_collection_literal) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create table list_t (pk int PRIMARY KEY, c1 list<int>);").get();
|
|
e.execute_cql("create table set_t (pk int PRIMARY KEY, c1 set<int>);").get();
|
|
e.execute_cql("create table map_t (pk int PRIMARY KEY, c1 map<int, int>);").get();
|
|
|
|
auto insert_list = e.prepare("insert into list_t (pk, c1) values (112, [997, ?])").get();
|
|
auto insert_set = e.prepare("insert into set_t (pk, c1) values (112, {997, ?})").get();
|
|
auto insert_map_key = e.prepare("insert into map_t (pk, c1) values (112, {997: 112, ?: 112})").get();
|
|
auto insert_map_value = e.prepare("insert into map_t (pk, c1) values (112, {997: 112, 112: ?})").get();
|
|
BOOST_REQUIRE_THROW(
|
|
e.prepare("insert into map_t (pk, c1) values (112, {997: 112, ?})").get(),
|
|
exceptions::syntax_exception
|
|
);
|
|
|
|
std::vector<cql3::prepared_cache_key_type> stmts = {insert_list, insert_set, insert_map_key, insert_map_value};
|
|
|
|
for (const auto& stmt : stmts) {
|
|
// Null value is not allowed as a collections element
|
|
const auto null_value = cql3::raw_value::make_null();
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(stmt, {null_value}).get(),
|
|
exceptions::invalid_request_exception
|
|
);
|
|
|
|
// Check if types mismatch is detected
|
|
const auto short_value = cql3::raw_value::make_value(short_type->decompose(data_value((int16_t)1)));
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(stmt, {short_value}).get(),
|
|
exceptions::invalid_request_exception
|
|
);
|
|
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_prepared(stmt, cql3::raw_value_vector_with_unset({null_value}, {true})).get(),
|
|
exceptions::invalid_request_exception
|
|
);
|
|
|
|
// Inserting a valid value has to be successful
|
|
const auto int_value = cql3::raw_value::make_value(int32_type->decompose(data_value(2)));
|
|
BOOST_REQUIRE_NO_THROW(e.execute_prepared(stmt, {int_value}).get());
|
|
}
|
|
});
|
|
}
|
|
|
|
/// The number of distinct values in a list is limited. Test the
|
|
// limit.
|
|
SEASTAR_TEST_CASE(test_list_append_limit) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t (pk int PRIMARY KEY, l list<int>);").get();
|
|
std::string value_list = "0";
|
|
for (int i = 0; i < utils::UUID_gen::SUBMICRO_LIMIT; i++) {
|
|
value_list.append(",0");
|
|
}
|
|
// Use a local copy of query_options to avoid impact on
|
|
// adjacent tests: that's where Scylla stores the list
|
|
// append sequence, which will be exceeded it in this
|
|
// test.
|
|
auto qo = std::make_unique<cql3::query_options>(db::consistency_level::LOCAL_ONE,
|
|
std::vector<cql3::raw_value>{},
|
|
cql3::query_options::specific_options{1, nullptr, {}, api::new_timestamp()});
|
|
auto cql = fmt::format("UPDATE t SET l = l + [{}] WHERE pk = 0;", value_list);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(cql, std::move(qo)).get(), exceptions::invalid_request_exception);
|
|
e.execute_cql("DROP TABLE t;").get();
|
|
});
|
|
}
|
|
|
|
|
|
SEASTAR_TEST_CASE(test_insert_statement) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (p1 varchar, c1 int, r1 int, PRIMARY KEY (p1, c1));").discard_result().then([&e] {
|
|
return e.execute_cql("insert into cf (p1, c1, r1) values ('key1', 1, 100);").discard_result();
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {1}, "r1", 100);
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set r1 = 66 where p1 = 'key1' and c1 = 1;").discard_result();
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {1}, "r1", 66);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_select_statement) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.create_table([](std::string_view ks_name) {
|
|
// CQL: create table cf (p1 varchar, c1 int, c2 int, r1 int, PRIMARY KEY (p1, c1, c2));
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("p1", utf8_type, column_kind::partition_key)
|
|
.with_column("c1", int32_type, column_kind::clustering_key)
|
|
.with_column("c2", int32_type, column_kind::clustering_key)
|
|
.with_column("r1", int32_type)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, c1, c2, r1) values ('key1', 1, 2, 3);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, c1, c2, r1) values ('key2', 1, 2, 13);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, c1, c2, r1) values ('key3', 1, 2, 23);").discard_result();
|
|
}).then([&e] {
|
|
// Test wildcard
|
|
return e.execute_cql("select * from cf where p1 = 'key1' and c2 = 2 and c1 = 1;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key1"))},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// Test with only regular column
|
|
return e.execute_cql("select r1 from cf where p1 = 'key1' and c2 = 2 and c1 = 1;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{int32_type->decompose(3)}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// Test full partition range, singular clustering range
|
|
return e.execute_cql("select * from cf where c1 = 1 and c2 = 2 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(3)
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key1"))},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)}})
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key2"))},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(13)}})
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key3"))},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(23)}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_cassandra_stress_like_write_and_read) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto execute_update_for_key = [&e](sstring key) {
|
|
return e.execute_cql(fmt::format("UPDATE cf SET "
|
|
"\"C0\" = 0x8f75da6b3dcec90c8a404fb9a5f6b0621e62d39c69ba5758e5f41b78311fbb26cc7a,"
|
|
"\"C1\" = 0xa8761a2127160003033a8f4f3d1069b7833ebe24ef56b3beee728c2b686ca516fa51,"
|
|
"\"C2\" = 0x583449ce81bfebc2e1a695eb59aad5fcc74d6d7311fc6197b10693e1a161ca2e1c64,"
|
|
"\"C3\" = 0x62bcb1dbc0ff953abc703bcb63ea954f437064c0c45366799658bd6b91d0f92908d7,"
|
|
"\"C4\" = 0x222fcbe31ffa1e689540e1499b87fa3f9c781065fccd10e4772b4c7039c2efd0fb27 "
|
|
"WHERE \"KEY\"={};", key)).discard_result();
|
|
};
|
|
|
|
auto verify_row_for_key = [&e](sstring key) {
|
|
return e.execute_cql(
|
|
format("select \"C0\", \"C1\", \"C2\", \"C3\", \"C4\" from cf where \"KEY\" = {}", key)).then(
|
|
[](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{from_hex(
|
|
"8f75da6b3dcec90c8a404fb9a5f6b0621e62d39c69ba5758e5f41b78311fbb26cc7a")},
|
|
{from_hex(
|
|
"a8761a2127160003033a8f4f3d1069b7833ebe24ef56b3beee728c2b686ca516fa51")},
|
|
{from_hex(
|
|
"583449ce81bfebc2e1a695eb59aad5fcc74d6d7311fc6197b10693e1a161ca2e1c64")},
|
|
{from_hex(
|
|
"62bcb1dbc0ff953abc703bcb63ea954f437064c0c45366799658bd6b91d0f92908d7")},
|
|
{from_hex("222fcbe31ffa1e689540e1499b87fa3f9c781065fccd10e4772b4c7039c2efd0fb27")}
|
|
});
|
|
});
|
|
};
|
|
|
|
return e.create_table([](std::string_view ks_name) {
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("KEY", bytes_type, column_kind::partition_key)
|
|
.with_column("C0", bytes_type)
|
|
.with_column("C1", bytes_type)
|
|
.with_column("C2", bytes_type)
|
|
.with_column("C3", bytes_type)
|
|
.with_column("C4", bytes_type)
|
|
.build();
|
|
}).then([execute_update_for_key, verify_row_for_key] {
|
|
static auto make_key = [](int suffix) { return format("0xdeadbeefcafebabe{:02d}", suffix); };
|
|
auto suffixes = std::views::iota(0, 10);
|
|
return parallel_for_each(suffixes.begin(), suffixes.end(), [execute_update_for_key](int suffix) {
|
|
return execute_update_for_key(make_key(suffix));
|
|
}).then([suffixes, verify_row_for_key] {
|
|
return parallel_for_each(suffixes.begin(), suffixes.end(), [verify_row_for_key](int suffix) {
|
|
return verify_row_for_key(make_key(suffix));
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_range_queries) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.create_table([](std::string_view ks_name) {
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("k", bytes_type, column_kind::partition_key)
|
|
.with_column("c0", bytes_type, column_kind::clustering_key)
|
|
.with_column("c1", bytes_type, column_kind::clustering_key)
|
|
.with_column("v", bytes_type)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x01 where k = 0x00 and c0 = 0x01 and c1 = 0x01;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x02 where k = 0x00 and c0 = 0x01 and c1 = 0x02;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x03 where k = 0x00 and c0 = 0x01 and c1 = 0x03;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x04 where k = 0x00 and c0 = 0x02 and c1 = 0x02;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x05 where k = 0x00 and c0 = 0x02 and c1 = 0x03;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x06 where k = 0x00 and c0 = 0x02 and c1 = 0x04;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x07 where k = 0x00 and c0 = 0x03 and c1 = 0x04;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x08 where k = 0x00 and c0 = 0x03 and c1 = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_rows({
|
|
{from_hex("01")},
|
|
{from_hex("02")},
|
|
{from_hex("03")},
|
|
{from_hex("04")},
|
|
{from_hex("05")},
|
|
{from_hex("06")},
|
|
{from_hex("07")},
|
|
{from_hex("08")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 = 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")}, {from_hex("05")}, {from_hex("06")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 > 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("07")}, {from_hex("08")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 >= 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")}, {from_hex("05")}, {from_hex("06")}, {from_hex("07")}, {from_hex("08")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 >= 0x02 and c0 < 0x03 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")}, {from_hex("05")}, {from_hex("06")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 > 0x02 and c0 <= 0x03 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("07")}, {from_hex("08")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 >= 0x02 and c0 <= 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")}, {from_hex("05")}, {from_hex("06")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 < 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")}, {from_hex("02")}, {from_hex("03")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 = 0x02 and c1 > 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("05")}, {from_hex("06")}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 and c0 = 0x02 and c1 >= 0x02 and c1 <= 0x02 allow filtering;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_ordering_of_composites_with_variable_length_components) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.create_table([](std::string_view ks) {
|
|
return *schema_builder(ks, "cf")
|
|
.with_column("k", bytes_type, column_kind::partition_key)
|
|
// We need more than one clustering column so that the single-element tuple format optimisation doesn't kick in
|
|
.with_column("c0", bytes_type, column_kind::clustering_key)
|
|
.with_column("c1", bytes_type, column_kind::clustering_key)
|
|
.with_column("v", bytes_type)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x01 where k = 0x00 and c0 = 0x0001 and c1 = 0x00;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x02 where k = 0x00 and c0 = 0x03 and c1 = 0x00;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x03 where k = 0x00 and c0 = 0x035555 and c1 = 0x00;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x04 where k = 0x00 and c0 = 0x05 and c1 = 0x00;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where k = 0x00 allow filtering;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")}, {from_hex("02")}, {from_hex("03")}, {from_hex("04")}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_query_with_static_columns) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (k blob, c blob, v blob, s1 blob static, s2 blob static, primary key (k, c));").discard_result().then([&e] {
|
|
return e.execute_cql("update cf set s1 = 0x01 where k = 0x00;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x02 where k = 0x00 and c = 0x01;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = 0x03 where k = 0x00 and c = 0x02;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1, v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01"), from_hex("02")},
|
|
{from_hex("01"), from_hex("03")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1 from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")},
|
|
{from_hex("01")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1 from cf limit 1;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1, v from cf limit 1;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01"), from_hex("02")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set v = null where k = 0x00 and c = 0x02;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1 from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (k, c) values (0x00, 0x02);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select s1 from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")},
|
|
{from_hex("01")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// Try 'in' restriction out
|
|
return e.execute_cql("select s1, v from cf where k = 0x00 and c in (0x01, 0x02);").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01"), from_hex("02")},
|
|
{from_hex("01"), {}}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// Verify that limit is respected for multiple clustering ranges and that static columns
|
|
// are populated when limit kicks in.
|
|
return e.execute_cql("select s1, v from cf where k = 0x00 and c in (0x01, 0x02) limit 1;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01"), from_hex("02")}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_insert_without_clustering_key) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (k blob, v blob, primary key (k));").discard_result().then([&e] {
|
|
return e.execute_cql("insert into cf (k) values (0x01);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01"), {}}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select k from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_limit_is_respected_across_partitions) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (k blob, c blob, v blob, s1 blob static, primary key (k, c));").discard_result().then([&e] {
|
|
return e.execute_cql("update cf set s1 = 0x01 where k = 0x01;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set s1 = 0x02 where k = 0x02;").discard_result();
|
|
}).then([&e] {
|
|
// Determine partition order
|
|
return e.execute_cql("select k from cf;");
|
|
}).then([&e](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(msg);
|
|
BOOST_REQUIRE(rows);
|
|
std::vector<bytes> keys;
|
|
const auto& rs = rows->rs().result_set();
|
|
for (auto&& row : rs.rows()) {
|
|
BOOST_REQUIRE(row[0]);
|
|
keys.push_back(to_bytes(*row[0]));
|
|
}
|
|
BOOST_REQUIRE(keys.size() == 2);
|
|
bytes k1 = keys[0];
|
|
bytes k2 = keys[1];
|
|
|
|
return now().then([k1, k2, &e] {
|
|
return e.execute_cql("select s1 from cf limit 1;").then([k1, k2](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{k1},
|
|
});
|
|
});
|
|
}).then([&e, k1, k2] {
|
|
return e.execute_cql("select s1 from cf limit 2;").then([k1, k2](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{k1}, {k2}
|
|
});
|
|
});
|
|
}).then([&e, k1, k2] {
|
|
return e.execute_cql(format("update cf set s1 = null where k = 0x{};", to_hex(k1))).discard_result();
|
|
}).then([&e, k1, k2] {
|
|
return e.execute_cql("select s1 from cf limit 1;").then([k1, k2](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{k2}
|
|
});
|
|
});
|
|
}).then([&e, k1, k2] {
|
|
return e.execute_cql(format("update cf set s1 = null where k = 0x{};", to_hex(k2))).discard_result();
|
|
}).then([&e, k1, k2] {
|
|
return e.execute_cql("select s1 from cf limit 1;").then([k1, k2](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_partitions_have_consistent_ordering_in_range_query) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (k blob, v int, primary key (k));").discard_result().then([&e] {
|
|
return e.execute_cql(
|
|
"begin unlogged batch \n"
|
|
" insert into cf (k, v) values (0x01, 0); \n"
|
|
" insert into cf (k, v) values (0x02, 0); \n"
|
|
" insert into cf (k, v) values (0x03, 0); \n"
|
|
" insert into cf (k, v) values (0x04, 0); \n"
|
|
" insert into cf (k, v) values (0x05, 0); \n"
|
|
" insert into cf (k, v) values (0x06, 0); \n"
|
|
"apply batch;"
|
|
).discard_result();
|
|
}).then([&e] {
|
|
// Determine partition order
|
|
return e.execute_cql("select k from cf;");
|
|
}).then([&e](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(msg);
|
|
BOOST_REQUIRE(rows);
|
|
std::vector<bytes> keys;
|
|
const auto& rs = rows->rs().result_set();
|
|
for (auto&& row : rs.rows()) {
|
|
BOOST_REQUIRE(row[0]);
|
|
keys.push_back(to_bytes(*row[0]));
|
|
}
|
|
BOOST_REQUIRE(keys.size() == 6);
|
|
|
|
return now().then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 1;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]}
|
|
});
|
|
});
|
|
}).then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 2;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]}
|
|
});
|
|
});
|
|
}).then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 3;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]}
|
|
});
|
|
});
|
|
}).then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 4;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]},
|
|
{keys[3]}
|
|
});
|
|
});
|
|
}).then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 5;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]},
|
|
{keys[3]},
|
|
{keys[4]}
|
|
});
|
|
});
|
|
}).then([keys, &e] {
|
|
return e.execute_cql("select k from cf limit 6;").then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]},
|
|
{keys[3]},
|
|
{keys[4]},
|
|
{keys[5]}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_partition_range_queries_with_bounds) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (k blob, v int, primary key (k));").discard_result().then([&e] {
|
|
return e.execute_cql(
|
|
"begin unlogged batch \n"
|
|
" insert into cf (k, v) values (0x01, 0); \n"
|
|
" insert into cf (k, v) values (0x02, 0); \n"
|
|
" insert into cf (k, v) values (0x03, 0); \n"
|
|
" insert into cf (k, v) values (0x04, 0); \n"
|
|
" insert into cf (k, v) values (0x05, 0); \n"
|
|
"apply batch;"
|
|
).discard_result();
|
|
}).then([&e] {
|
|
// Determine partition order
|
|
return e.execute_cql("select k, token(k) from cf;");
|
|
}).then([&e](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(msg);
|
|
BOOST_REQUIRE(rows);
|
|
std::vector<bytes> keys;
|
|
std::vector<int64_t> tokens;
|
|
const auto& rs = rows->rs().result_set();
|
|
for (auto&& row : rs.rows()) {
|
|
BOOST_REQUIRE(row[0]);
|
|
BOOST_REQUIRE(row[1]);
|
|
keys.push_back(to_bytes(*row[0]));
|
|
tokens.push_back(value_cast<int64_t>(long_type->deserialize(*row[1])));
|
|
}
|
|
BOOST_REQUIRE(keys.size() == 5);
|
|
|
|
return now().then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) > {:d};", tokens[1])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[2]},
|
|
{keys[3]},
|
|
{keys[4]}
|
|
});
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) >= {:d};", tokens[1])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[1]},
|
|
{keys[2]},
|
|
{keys[3]},
|
|
{keys[4]}
|
|
});
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) > {:d} and token(k) < {:d};",
|
|
tokens[1], tokens[4])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[2]},
|
|
{keys[3]},
|
|
});
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) < {:d};", tokens[3])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]}
|
|
});
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) = {:d};", tokens[3])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[3]}
|
|
});
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) < {:d} and token(k) > {:d};", tokens[3], tokens[3])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
return e.execute_cql(format("select k from cf where token(k) >= {:d} and token(k) <= {:d};", tokens[4], tokens[2])).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
}).then([keys, tokens, &e] {
|
|
auto min_token = std::numeric_limits<int64_t>::min();
|
|
return e.execute_cql(format("select k from cf where token(k) > {:d} and token (k) < {:d};", min_token, min_token)).then([keys](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{keys[0]},
|
|
{keys[1]},
|
|
{keys[2]},
|
|
{keys[3]},
|
|
{keys[4]}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_deletion_scenarios) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.create_table([](std::string_view ks) {
|
|
// CQL: create table cf (k bytes, c bytes, v bytes, primary key (k, c));
|
|
return *schema_builder(ks, "cf")
|
|
.with_column("k", bytes_type, column_kind::partition_key)
|
|
.with_column("c", bytes_type, column_kind::clustering_key)
|
|
.with_column("v", bytes_type)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (k, c, v) values (0x00, 0x05, 0x01) using timestamp 1;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("01")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf using timestamp 2 set v = null where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{}},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// same tampstamp, dead cell wins
|
|
return e.execute_cql("update cf using timestamp 2 set v = 0x02 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{}},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf using timestamp 3 set v = 0x02 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("02")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// same timestamp, greater value wins
|
|
return e.execute_cql("update cf using timestamp 3 set v = 0x03 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("03")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// same tampstamp, delete whole row, delete should win
|
|
return e.execute_cql("delete from cf using timestamp 3 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
}).then([&e] {
|
|
// same timestamp, update should be shadowed by range tombstone
|
|
return e.execute_cql("update cf using timestamp 3 set v = 0x04 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf using timestamp 4 set v = 0x04 where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{from_hex("04")},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// deleting an orphan cell (row is considered as deleted) yields no row
|
|
return e.execute_cql("update cf using timestamp 5 set v = null where k = 0x00 and c = 0x05;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().is_empty();
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_range_deletion_scenarios) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table cf (p int, c int, v text, primary key (p, c));").get();
|
|
for (auto i = 0; i < 10; ++i) {
|
|
e.execute_cql(format("insert into cf (p, c, v) values (1, {:d}, 'abc');", i)).get();
|
|
}
|
|
|
|
e.execute_cql("delete from cf where p = 1 and c <= 3").get();
|
|
e.execute_cql("delete from cf where p = 1 and c >= 8").get();
|
|
|
|
e.execute_cql("delete from cf where p = 1 and c >= 0 and c <= 5").get();
|
|
auto msg = e.execute_cql("select * from cf").get();
|
|
assert_that(msg).is_rows().with_size(2);
|
|
e.execute_cql("delete from cf where p = 1 and c > 3 and c < 10").get();
|
|
msg = e.execute_cql("select * from cf").get();
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
e.execute_cql("insert into cf (p, c, v) values (1, 1, '1');").get();
|
|
e.execute_cql("insert into cf (p, c, v) values (1, 3, '3');").get();
|
|
e.execute_cql("delete from cf where p = 1 and c >= 2 and c <= 3").get();
|
|
e.execute_cql("insert into cf (p, c, v) values (1, 2, '2');").get();
|
|
msg = e.execute_cql("select * from cf").get();
|
|
assert_that(msg).is_rows().with_size(2);
|
|
e.execute_cql("delete from cf where p = 1 and c >= 2 and c <= 3").get();
|
|
msg = e.execute_cql("select * from cf").get();
|
|
assert_that(msg).is_rows().with_rows({{ {int32_type->decompose(1)}, {int32_type->decompose(1)}, {utf8_type->decompose("1")} }});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_range_deletion_scenarios_with_compact_storage) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("update system.config SET value='true' where name='enable_create_table_with_compact_storage';").get();
|
|
e.execute_cql("create table cf (p int, c int, v text, primary key (p, c)) with compact storage;").get();
|
|
for (auto i = 0; i < 10; ++i) {
|
|
e.execute_cql(format("insert into cf (p, c, v) values (1, {:d}, 'abc');", i)).get();
|
|
}
|
|
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c <= 3").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c >= 0").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c > 0 and c <= 3").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c >= 0 and c < 3").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c > 0 and c < 3").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
try {
|
|
e.execute_cql("delete from cf where p = 1 and c >= 0 and c <= 3").get();
|
|
BOOST_FAIL("should've thrown");
|
|
} catch (...) { }
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_map_insert_update) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto make_my_map_type = [] { return map_type_impl::get_instance(int32_type, int32_type, true); };
|
|
auto my_map_type = make_my_map_type();
|
|
return e.create_table([make_my_map_type] (std::string_view ks_name) {
|
|
// CQL: create table cf (p1 varchar primary key, map1 map<int, int>);
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("p1", utf8_type, column_kind::partition_key)
|
|
.with_column("map1", make_my_map_type())
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, map1) values ('key1', { 1001: 2001 });").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type, map_type_impl::native_type({{1001, 2001}})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set map1[1002] = 2002 where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{1001, 2001},
|
|
{1002, 2002}})));
|
|
}).then([&e] {
|
|
// overwrite an element
|
|
return e.execute_cql("update cf set map1[1001] = 3001 where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{1001, 3001},
|
|
{1002, 2002}})));
|
|
}).then([&e] {
|
|
// overwrite whole map
|
|
return e.execute_cql("update cf set map1 = {1003: 4003} where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{1003, 4003}})));
|
|
}).then([&e] {
|
|
// overwrite whole map, but bad syntax
|
|
return e.execute_cql("update cf set map1 = {1003, 4003} where p1 = 'key1';");
|
|
}).then_wrapped([&e](future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
BOOST_REQUIRE(f.failed());
|
|
f.ignore_ready_future();
|
|
// overwrite whole map
|
|
return e.execute_cql(
|
|
"update cf set map1 = {1001: 5001, 1002: 5002, 1003: 5003} where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{1001, 5001},
|
|
{1002, 5002},
|
|
{1003, 5003}})));
|
|
}).then([&e] {
|
|
// discard some keys
|
|
return e.execute_cql("update cf set map1 = map1 - {1001, 1003, 1005} where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{{1002, 5002}}})));
|
|
}).then([&e, my_map_type] {
|
|
return e.execute_cql("select * from cf where p1 = 'key1';").then([my_map_type](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key1"))},
|
|
{my_map_type->decompose(make_map_value(my_map_type, map_type_impl::native_type{{{1002, 5002}}}))},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
// overwrite an element
|
|
return e.execute_cql("update cf set map1[1009] = 5009 where p1 = 'key1';").discard_result();
|
|
}).then([&e] {
|
|
// delete a key
|
|
return e.execute_cql("delete map1[1002] from cf where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type,
|
|
map_type_impl::native_type({{{1009, 5009}}})));
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, map1) values ('key1', null);").discard_result();
|
|
}).then([&e, my_map_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"map1", make_map_value(my_map_type, map_type_impl::native_type({})));
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_set_insert_update) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto make_my_set_type = [] { return set_type_impl::get_instance(int32_type, true); };
|
|
auto my_set_type = make_my_set_type();
|
|
return e.create_table([make_my_set_type](std::string_view ks_name) {
|
|
// CQL: create table cf (p1 varchar primary key, set1 set<int>);
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("p1", utf8_type, column_kind::partition_key)
|
|
.with_column("set1", make_my_set_type())
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, set1) values ('key1', { 1001 });").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({{1001}})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set set1 = set1 + { 1002 } where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({1001, 1002})));
|
|
}).then([&e] {
|
|
// overwrite an element
|
|
return e.execute_cql("update cf set set1 = set1 + { 1001 } where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({1001, 1002})));
|
|
}).then([&e] {
|
|
// overwrite entire set
|
|
return e.execute_cql("update cf set set1 = { 1007, 1019 } where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({1007, 1019})));
|
|
}).then([&e] {
|
|
// discard keys
|
|
return e.execute_cql("update cf set set1 = set1 - { 1007, 1008 } where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({{1019}})));
|
|
}).then([&e, my_set_type] {
|
|
return e.execute_cql("select * from cf where p1 = 'key1';").then([my_set_type](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key1"))},
|
|
{my_set_type->decompose(make_set_value(my_set_type, set_type_impl::native_type{{1019}}))},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set set1 = set1 + { 1009 } where p1 = 'key1';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("delete set1[1019] from cf where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({{1009}})));
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, set1) values ('key1', null);").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({})));
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, set1) values ('key1', {});").discard_result();
|
|
}).then([&e, my_set_type] {
|
|
// Empty non-frozen set is indistinguishable from NULL
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"set1", make_set_value(my_set_type, set_type_impl::native_type({})));
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_list_insert_update) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto my_list_type = list_type_impl::get_instance(int32_type, true);
|
|
return e.execute_cql("create table cf (p1 varchar primary key, list1 list<int>);").discard_result().then([&e] {
|
|
return e.execute_cql("insert into cf (p1, list1) values ('key1', [ 1001 ]);").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({{1001}})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1 = [ 1002, 1003 ] where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({1002, 1003})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1[1] = 2003 where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({1002, 2003})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1 = list1 - [1002, 2004] where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({{2003}})));
|
|
}).then([&e, my_list_type] {
|
|
return e.execute_cql("select * from cf where p1 = 'key1';").then([my_list_type] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{utf8_type->decompose(sstring("key1"))},
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type({{2003}})))},
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1 = [2008, 2009, 2010] where p1 = 'key1';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("delete list1[1] from cf where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({2008, 2010})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1 = list1 + [2012, 2019] where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({2008, 2010, 2012, 2019})));
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf set list1 = [2001, 2002] + list1 where p1 = 'key1';").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({2001, 2002, 2008, 2010, 2012, 2019})));
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, list1) values ('key1', null);").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({})));
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, list1) values ('key1', []);").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
// Empty non-frozen list is indistinguishable from NULL
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {},
|
|
"list1", make_list_value(my_list_type, list_type_impl::native_type({})));
|
|
});
|
|
});
|
|
}
|
|
|
|
static const api::timestamp_type the_timestamp = 123456789;
|
|
SEASTAR_TEST_CASE(test_writetime_and_ttl) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (p1 varchar primary key, i int, fc frozen<set<int>>, c set<int>);").discard_result().then([&e] {
|
|
auto q = format("insert into cf (p1, i) values ('key1', 1) using timestamp {:d};", the_timestamp);
|
|
return e.execute_cql(q).discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select writetime(i) from cf where p1 in ('key1');");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_rows({{
|
|
{long_type->decompose(int64_t(the_timestamp))},
|
|
}});
|
|
}).then([&e] {
|
|
return async([&e] {
|
|
auto ts1 = the_timestamp + 1;
|
|
e.execute_cql(format("UPDATE cf USING TIMESTAMP {:d} SET fc = {{1}}, c = {{2}} WHERE p1 = 'key1'", ts1)).get();
|
|
auto msg1 = e.execute_cql("SELECT writetime(fc) FROM cf").get();
|
|
assert_that(msg1).is_rows()
|
|
.with_rows({{
|
|
{long_type->decompose(int64_t(ts1))},
|
|
}});
|
|
auto msg2f = futurize_invoke([&] { return e.execute_cql("SELECT writetime(c) FROM cf"); });
|
|
msg2f.wait();
|
|
assert_that_failed(msg2f);
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_time_overflow_with_default_ttl) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto verify = [&e] (int value, bool bypass_cache) -> future<> {
|
|
auto sq = format("select i from cf where p1 = 'key1' {};", bypass_cache ? "bypass cache" : "");
|
|
return e.execute_cql(sq).then([value] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{int32_type->decompose(value)},
|
|
});
|
|
});
|
|
};
|
|
|
|
auto cr = format("create table cf (p1 varchar primary key, i int) with default_time_to_live = {:d};", max_ttl.count());
|
|
e.execute_cql(cr).get();
|
|
auto q = format("insert into cf (p1, i) values ('key1', 1);");
|
|
e.execute_cql(q).get();
|
|
require_column_has_value(e, "cf", {sstring("key1")}, {}, "i", 1).get();
|
|
verify(1, false).get();
|
|
verify(1, true).get();
|
|
e.execute_cql("update cf set i = 2 where p1 = 'key1';").get();
|
|
verify(2, true).get();
|
|
verify(2, false).get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_time_overflow_using_ttl) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto verify = [&e] (int value, bool bypass_cache) -> future<> {
|
|
auto sq = format("select i from cf where p1 = 'key1' {};", bypass_cache ? "bypass cache" : "");
|
|
return e.execute_cql(sq).then([value] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_size(1)
|
|
.with_row({
|
|
{int32_type->decompose(value)},
|
|
});
|
|
});
|
|
};
|
|
|
|
auto cr = "create table cf (p1 varchar primary key, i int);";
|
|
e.execute_cql(cr).get();
|
|
auto q = format("insert into cf (p1, i) values ('key1', 1) using ttl {:d};", max_ttl.count());
|
|
e.execute_cql(q).get();
|
|
require_column_has_value(e, "cf", {sstring("key1")}, {}, "i", 1).get();
|
|
verify(1, true).get();
|
|
verify(1, false).get();
|
|
q = format("insert into cf (p1, i) values ('key2', 0);");
|
|
e.execute_cql(q).get();
|
|
q = format("update cf using ttl {:d} set i = 2 where p1 = 'key2';", max_ttl.count());
|
|
e.execute_cql(q).get();
|
|
require_column_has_value(e, "cf", {sstring("key2")}, {}, "i", 2).get();
|
|
verify(1, false).get();
|
|
verify(1, true).get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_batch) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (p1 varchar, c1 int, r1 int, PRIMARY KEY (p1, c1));").discard_result().then([&e] {
|
|
return e.execute_cql(
|
|
"begin unlogged batch \n"
|
|
" insert into cf (p1, c1, r1) values ('key1', 1, 100); \n"
|
|
" insert into cf (p1, c1, r1) values ('key1', 2, 200); \n"
|
|
"apply batch;"
|
|
).discard_result();
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {1}, "r1", 100);
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {2}, "r1", 200);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_tuples) {
|
|
auto make_tt = [] { return tuple_type_impl::get_instance({int32_type, long_type, utf8_type}); };
|
|
auto tt = make_tt();
|
|
return do_with_cql_env([tt, make_tt] (cql_test_env& e) {
|
|
return e.create_table([make_tt] (std::string_view ks_name) {
|
|
// this runs on all cores, so create a local tt for each core:
|
|
auto tt = make_tt();
|
|
// CQL: "create table cf (id int primary key, t tuple<int, bigint, text>);
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("id", int32_type, column_kind::partition_key)
|
|
.with_column("t", tt)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (id, t) values (1, (1001, 2001, 'abc1'));").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select t from cf where id = 1;");
|
|
}).then([&e, tt] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_rows({{
|
|
{tt->decompose(make_tuple_value(tt, tuple_type_impl::native_type({int32_t(1001), int64_t(2001), sstring("abc1")})))},
|
|
}});
|
|
return e.execute_cql("create table cf2 (p1 int PRIMARY KEY, r1 tuple<int, bigint, text>)").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r1) values (1, (1, 2, 'abc'));").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from cf2 where p1 = 1;");
|
|
}).then([tt] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(int32_t(1)), tt->decompose(make_tuple_value(tt, tuple_type_impl::native_type({int32_t(1), int64_t(2), sstring("abc")}))) }
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_vectors) {
|
|
auto make_vt = [] { return vector_type_impl::get_instance(int32_type, 3); };
|
|
auto vt = make_vt();
|
|
return do_with_cql_env([vt, make_vt] (cql_test_env& e) {
|
|
return e.create_table([make_vt] (std::string_view ks_name) {
|
|
// this runs on all cores, so create a local vt for each core:
|
|
auto vt = make_vt();
|
|
// CQL: "create table cf (id int primary key, v vector<int, 3>);
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("id", int32_type, column_kind::partition_key)
|
|
.with_column("v", vt)
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (id, v) values (1, [1001, 2001, 3001]);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select v from cf where id = 1;");
|
|
}).then([&e, vt] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows()
|
|
.with_rows({{
|
|
{vt->decompose(make_vector_value(vt, vector_type_impl::native_type({int32_t(1001), int32_t(2001), int32_t(3001)})))},
|
|
}});
|
|
return e.execute_cql("create table cf2 (p1 int PRIMARY KEY, r1 vector<int, 3>)").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r1) values (1, [1, 2, 3]);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from cf2 where p1 = 1;");
|
|
}).then([vt] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(int32_t(1)), vt->decompose(make_vector_value(vt, vector_type_impl::native_type({int32_t(1), int32_t(2), int32_t(3)}))) }
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_vectors_variable_length_elements) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t1 (id int PRIMARY KEY, v vector<text, 2>);").get();
|
|
e.execute_cql("INSERT INTO t1 (id, v) VALUES (1, ['abc', '']);").get();
|
|
|
|
auto msg = e.execute_cql("SELECT * FROM t1;").get();
|
|
auto vt = vector_type_impl::get_instance(utf8_type, 2);
|
|
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), vt->decompose(
|
|
make_vector_value(
|
|
vt,
|
|
vector_type_impl::native_type({
|
|
sstring("abc"),
|
|
sstring("")
|
|
})
|
|
)
|
|
)}
|
|
});
|
|
|
|
e.execute_cql("CREATE TABLE t2 (id int PRIMARY KEY, v vector<list<int>, 2>);").get();
|
|
e.execute_cql("INSERT INTO t2 (id, v) VALUES (1, [[1, 2], [3, 4, 5]]);").get();
|
|
|
|
msg = e.execute_cql("SELECT * FROM t2;").get();
|
|
vt = vector_type_impl::get_instance(list_type_impl::get_instance(int32_type, true), 2);
|
|
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), vt->decompose(
|
|
make_vector_value(
|
|
vt,
|
|
vector_type_impl::native_type({
|
|
make_list_value(list_type_impl::get_instance(int32_type, true), list_type_impl::native_type({1, 2})),
|
|
make_list_value(list_type_impl::get_instance(int32_type, true), list_type_impl::native_type({3, 4, 5}))
|
|
})
|
|
)
|
|
)}
|
|
});
|
|
|
|
e.execute_cql("CREATE TABLE t3 (id int PRIMARY KEY, v vector<tuple<int, text>, 2>);").get();
|
|
e.execute_cql("INSERT INTO t3 (id, v) VALUES (1, [(123, 'abc'), (456, '')]);").get();
|
|
|
|
msg = e.execute_cql("SELECT * FROM t3;").get();
|
|
vt = vector_type_impl::get_instance(tuple_type_impl::get_instance({int32_type, utf8_type}), 2);
|
|
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), vt->decompose(
|
|
make_vector_value(
|
|
vt,
|
|
vector_type_impl::native_type({
|
|
make_tuple_value(tuple_type_impl::get_instance({int32_type, utf8_type}), tuple_type_impl::native_type({123, sstring("abc")})),
|
|
make_tuple_value(tuple_type_impl::get_instance({int32_type, utf8_type}), tuple_type_impl::native_type({456, sstring("")}))
|
|
})
|
|
)
|
|
)}
|
|
});
|
|
|
|
});
|
|
}
|
|
|
|
namespace {
|
|
|
|
using std::source_location;
|
|
|
|
auto validate_request_failure(
|
|
cql_test_env& env,
|
|
const sstring& request,
|
|
const sstring& expected_message,
|
|
const source_location& loc = source_location::current()) {
|
|
return futurize_invoke([&] { return env.execute_cql(request); })
|
|
.then_wrapped([expected_message, loc] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
BOOST_REQUIRE_EXCEPTION(f.get(),
|
|
exceptions::invalid_request_exception,
|
|
exception_predicate::message_equals(expected_message, loc));
|
|
});
|
|
};
|
|
|
|
} // anonymous namespace
|
|
|
|
//
|
|
// Since durations don't have a well-defined ordering on their semantic value, a number of restrictions exist on their
|
|
// use.
|
|
//
|
|
SEASTAR_TEST_CASE(test_duration_restrictions) {
|
|
return do_with_cql_env([&] (cql_test_env& env) {
|
|
return make_ready_future<>().then([&] {
|
|
// Disallow "direct" use of durations in ordered collection types to avoid user confusion when their
|
|
// ordering doesn't match expectations.
|
|
return make_ready_future<>().then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create type my_type (a set<duration>);",
|
|
"Durations are not allowed inside sets: set<duration>");
|
|
}).then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create type my_type (a map<duration, int>);",
|
|
"Durations are not allowed as map keys: map<duration, int>");
|
|
});
|
|
}).then([&] {
|
|
// Disallow any type referring to a duration from being used in a primary key of a table or a materialized
|
|
// view.
|
|
return make_ready_future<>().then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create table my_table (direct_key duration PRIMARY KEY);",
|
|
"duration type is not supported for PRIMARY KEY part direct_key");
|
|
}).then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create table my_table (collection_key frozen<list<duration>> PRIMARY KEY);",
|
|
"duration type is not supported for PRIMARY KEY part collection_key");
|
|
}).then([&] {
|
|
return env.execute_cql("create type my_type0 (span duration);").discard_result().then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create table my_table (udt_key frozen<my_type0> PRIMARY KEY);",
|
|
"duration type is not supported for PRIMARY KEY part udt_key");
|
|
});
|
|
}).then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create table my_table (tuple_key tuple<int, duration, int> PRIMARY KEY);",
|
|
"duration type is not supported for PRIMARY KEY part tuple_key");
|
|
}).then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create table my_table (a int, b duration, PRIMARY KEY ((a), b)) WITH CLUSTERING ORDER BY (b DESC);",
|
|
"duration type is not supported for PRIMARY KEY part b");
|
|
}).then([&] {
|
|
return env.execute_cql("create table my_table0 (key int PRIMARY KEY, name text, span duration);")
|
|
.discard_result().then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"create materialized view my_mv as"
|
|
" select * from my_table0 "
|
|
" primary key (key, span);",
|
|
"Cannot use Duration column 'span' in PRIMARY KEY of materialized view");
|
|
});
|
|
});
|
|
}).then([&] {
|
|
// Disallow creating secondary indexes on durations.
|
|
return validate_request_failure(
|
|
env,
|
|
"create index my_index on my_table0 (span);",
|
|
"Secondary indexes are not supported on duration columns");
|
|
}).then([&] {
|
|
// Disallow slice-based restrictions and conditions on durations.
|
|
//
|
|
// Note that multi-column restrictions are only supported on clustering columns (which cannot be `duration`)
|
|
// and that multi-column conditions are not supported in the grammar.
|
|
return make_ready_future<>().then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"select * from my_table0 where key = 0 and span < 3d;",
|
|
"Duration type is unordered for span");
|
|
}).then([&] {
|
|
return validate_request_failure(
|
|
env,
|
|
"update my_table0 set name = 'joe' where key = 0 if span >= 5m",
|
|
"Duration type is unordered for span");
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_select_multiple_ranges) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (p1 varchar, r1 int, PRIMARY KEY (p1));").discard_result().then([&e] {
|
|
return e.execute_cql(
|
|
"begin unlogged batch \n"
|
|
" insert into cf (p1, r1) values ('key1', 100); \n"
|
|
" insert into cf (p1, r1) values ('key2', 200); \n"
|
|
"apply batch;"
|
|
).discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select r1 from cf where p1 in ('key1', 'key2');");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_size(2).with_row({
|
|
{int32_type->decompose(100)}
|
|
}).with_row({
|
|
{int32_type->decompose(200)}
|
|
});
|
|
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_validate_keyspace) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return make_ready_future<>().then([&e] {
|
|
sstring keyspace_name(schema::NAME_LENGTH + 1, 'k');
|
|
return e.execute_cql(format("create keyspace {} with replication = {{ 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 }};", keyspace_name));
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create keyspace ks3-1 with replication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create keyspace ks3 with replication = { 'replication_factor' : 1 };");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
BOOST_ASSERT(!f.failed());
|
|
return e.execute_cql("create keyspace ks3 with rreplication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create keyspace SyStEm with replication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };");
|
|
}).then_wrapped([] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_validate_table) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return make_ready_future<>().then([&e] {
|
|
sstring table_name(schema::NAME_LENGTH + 1, 't');
|
|
return e.execute_cql(format("create table {} (foo text PRIMARY KEY, bar text);", table_name));
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text PRIMARY KEY, foo text);");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb-1 (foo text PRIMARY KEY, bar text);");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text, bar text);");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text PRIMARY KEY, bar text PRIMARY KEY);");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text PRIMARY KEY, bar text) with commment = 'aaaa';");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text PRIMARY KEY, bar text) with min_index_interval = -1;");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("create table tb (foo text PRIMARY KEY, bar text) with min_index_interval = 1024 and max_index_interval = 128;");
|
|
}).then_wrapped([] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_table_compression) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tb1 (foo text PRIMARY KEY, bar text) with compression = { };").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb1"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb1")->get_compressor_params().get_algorithm() == compression_parameters::algorithm::none);
|
|
|
|
e.execute_cql("create table tb5 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : '' };").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb5"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb5")->get_compressor_params().get_algorithm() == compression_parameters::algorithm::none);
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql(
|
|
"create table tb2 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'LossyCompressor' };").get(),
|
|
std::exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(
|
|
"create table tb2 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'LZ4Compressor', 'chunk_length_kb' : -1 };").get(),
|
|
std::exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(
|
|
"create table tb2 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'LZ4Compressor', 'chunk_length_kb' : 3 };").get(),
|
|
std::exception);
|
|
|
|
e.execute_cql("create table tb2 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'LZ4Compressor', 'chunk_length_kb' : 2 };").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb2"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb2")->get_compressor_params().get_algorithm() == compression_parameters::algorithm::lz4);
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb2")->get_compressor_params().chunk_length() == 2 * 1024);
|
|
|
|
e.execute_cql("create table tb3 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'DeflateCompressor' };").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb3"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb3")->get_compressor_params().get_algorithm() == compression_parameters::algorithm::deflate);
|
|
|
|
e.execute_cql("create table tb4 (foo text PRIMARY KEY, bar text) with compression = { 'sstable_compression' : 'org.apache.cassandra.io.compress.DeflateCompressor' };").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb4"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb4")->get_compressor_params().get_algorithm() == compression_parameters::algorithm::deflate);
|
|
|
|
e.execute_cql("create table tb6 (foo text PRIMARY KEY, bar text);").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tb6"));
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tb6")->get_compressor_params().get_algorithm() == e.local_db().get_config().sstable_compression_user_table_options().get_algorithm());
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_ttl) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto make_my_list_type = [] { return list_type_impl::get_instance(utf8_type, true); };
|
|
auto my_list_type = make_my_list_type();
|
|
return e.create_table([make_my_list_type] (std::string_view ks_name) {
|
|
return *schema_builder(ks_name, "cf")
|
|
.with_column("p1", utf8_type, column_kind::partition_key)
|
|
.with_column("r1", utf8_type)
|
|
.with_column("r2", utf8_type)
|
|
.with_column("r3", make_my_list_type())
|
|
.build();
|
|
}).then([&e] {
|
|
return e.execute_cql(
|
|
"update cf using ttl 100000 set r1 = 'value1_1', r3 = ['a', 'b', 'c'] where p1 = 'key1';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql(
|
|
"update cf using ttl 100 set r1 = 'value1_3', r3 = ['a', 'b', 'c'] where p1 = 'key3';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf using ttl 100 set r3[1] = 'b' where p1 = 'key1';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf using ttl 100 set r1 = 'value1_2' where p1 = 'key2';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf (p1, r2) values ('key2', 'value2_2');").discard_result();
|
|
}).then([&e, my_list_type] {
|
|
return e.execute_cql("select r1 from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({utf8_type->decompose(sstring("value1_1"))})
|
|
.with_row({utf8_type->decompose(sstring("value1_2"))})
|
|
.with_row({utf8_type->decompose(sstring("value1_3"))});
|
|
});
|
|
}).then([&e, my_list_type] {
|
|
return e.execute_cql("select r3 from cf where p1 = 'key1';").then([my_list_type] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
auto my_list_type = list_type_impl::get_instance(utf8_type, true);
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type{{sstring("a"), sstring("b"), sstring("c")}}))}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
forward_jump_clocks(200s);
|
|
return e.execute_cql("select r1, r2 from cf;").then([](shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_size(2)
|
|
.with_row({{}, utf8_type->decompose(sstring("value2_2"))})
|
|
.with_row({utf8_type->decompose(sstring("value1_1")), {}});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select r2 from cf;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_size(2)
|
|
.with_row({ utf8_type->decompose(sstring("value2_2")) })
|
|
.with_row({ {} });
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select r1 from cf;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_size(2)
|
|
.with_row({ {} })
|
|
.with_row({ utf8_type->decompose(sstring("value1_1")) });
|
|
});
|
|
}).then([&e, my_list_type] {
|
|
return e.execute_cql("select r3 from cf where p1 = 'key1';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
auto my_list_type = list_type_impl::get_instance(utf8_type, true);
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type{{sstring("a"), sstring("c")}}))}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("create table cf2 (p1 text PRIMARY KEY, r1 text, r2 text);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r1) values ('foo', 'bar') using ttl 500;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select p1, r1 from cf2 where p1 = 'foo';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose(sstring("foo")), utf8_type->decompose(sstring("bar"))}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
forward_jump_clocks(600s);
|
|
return e.execute_cql("select p1, r1 from cf2 where p1 = 'foo';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({ });
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select p1, r1 from cf2;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({ });
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("select count(*) from cf2;").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(0))}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r1) values ('foo', 'bar') using ttl 500;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("update cf2 set r1 = null where p1 = 'foo';").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select p1, r1 from cf2 where p1 = 'foo';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose(sstring("foo")), { }}
|
|
});
|
|
});
|
|
}).then([&e] {
|
|
forward_jump_clocks(600s);
|
|
return e.execute_cql("select p1, r1 from cf2 where p1 = 'foo';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({ });
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r1) values ('foo', 'bar') using ttl 500;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into cf2 (p1, r2) values ('foo', null);").discard_result();
|
|
}).then([&e] {
|
|
forward_jump_clocks(600s);
|
|
return e.execute_cql("select p1, r1 from cf2 where p1 = 'foo';").then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose(sstring("foo")), { }}
|
|
});
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_types) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql(
|
|
"CREATE TABLE all_types ("
|
|
" a ascii PRIMARY KEY,"
|
|
" b bigint,"
|
|
" c blob,"
|
|
" d boolean,"
|
|
" e double,"
|
|
" f float,"
|
|
" g inet,"
|
|
" h int,"
|
|
" i text,"
|
|
" j timestamp,"
|
|
" k timeuuid,"
|
|
" l uuid,"
|
|
" m varchar,"
|
|
" n varint,"
|
|
" o decimal,"
|
|
" p tinyint,"
|
|
" q smallint,"
|
|
" r date,"
|
|
" s time,"
|
|
" u duration,"
|
|
");").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "all_types"));
|
|
|
|
e.execute_cql(
|
|
"INSERT INTO all_types (a, b, c, d, e, f, g, h, i, j, k, l, m, n, o, p, q, r, s, u) VALUES ("
|
|
" 'ascii',"
|
|
" 123456789,"
|
|
" 0xdeadbeef,"
|
|
" true,"
|
|
" 3.14,"
|
|
" 3.14,"
|
|
" '127.0.0.1',"
|
|
" 3,"
|
|
" 'zażółć gęślą jaźń',"
|
|
" '2001-10-18 14:15:55.134+0000',"
|
|
" d2177dd0-eaa2-11de-a572-001b779c76e3,"
|
|
" d2177dd0-eaa2-11de-a572-001b779c76e3,"
|
|
" 'varchar',"
|
|
" 123,"
|
|
" 1.23,"
|
|
" 3,"
|
|
" 3,"
|
|
" '1970-01-02',"
|
|
" '00:00:00.000000001',"
|
|
" 1y2mo3w4d5h6m7s8ms9us10ns"
|
|
");").get();
|
|
|
|
{
|
|
auto msg = e.execute_cql("SELECT * FROM all_types WHERE a = 'ascii'").get();
|
|
struct tm t = { 0 };
|
|
t.tm_year = 2001 - 1900;
|
|
t.tm_mon = 10 - 1;
|
|
t.tm_mday = 18;
|
|
t.tm_hour = 14;
|
|
t.tm_min = 15;
|
|
t.tm_sec = 55;
|
|
auto tp = db_clock::from_time_t(timegm(&t)) + std::chrono::milliseconds(134);
|
|
assert_that(msg).is_rows().with_rows({
|
|
{
|
|
ascii_type->decompose(sstring("ascii")), long_type->decompose(123456789l),
|
|
from_hex("deadbeef"), boolean_type->decompose(true),
|
|
double_type->decompose(3.14), float_type->decompose(3.14f),
|
|
inet_addr_type->decompose(net::inet_address("127.0.0.1")),
|
|
int32_type->decompose(3), utf8_type->decompose(sstring("zażółć gęślą jaźń")),
|
|
timestamp_type->decompose(tp),
|
|
timeuuid_type->decompose(utils::UUID(sstring("d2177dd0-eaa2-11de-a572-001b779c76e3"))),
|
|
uuid_type->decompose(utils::UUID(sstring("d2177dd0-eaa2-11de-a572-001b779c76e3"))),
|
|
utf8_type->decompose(sstring("varchar")), varint_type->decompose(utils::multiprecision_int(123)),
|
|
decimal_type->decompose(big_decimal { 2, utils::multiprecision_int(123) }),
|
|
byte_type->decompose(int8_t(3)),
|
|
short_type->decompose(int16_t(3)),
|
|
serialized(simple_date_native_type{0x80000001}),
|
|
time_type->decompose(int64_t(0x0000000000000001)),
|
|
duration_type->decompose(cql_duration("1y2mo3w4d5h6m7s8ms9us10ns"))
|
|
}
|
|
});
|
|
}
|
|
|
|
e.execute_cql(
|
|
"INSERT INTO all_types (a, b, c, d, e, f, g, h, i, j, k, l, m, n, o, p, q, r, s, u) VALUES ("
|
|
" blobAsAscii(asciiAsBlob('ascii2')),"
|
|
" blobAsBigint(bigintAsBlob(123456789)),"
|
|
" bigintAsBlob(12),"
|
|
" blobAsBoolean(booleanAsBlob(true)),"
|
|
" blobAsDouble(doubleAsBlob(3.14)),"
|
|
" blobAsFloat(floatAsBlob(3.14)),"
|
|
" blobAsInet(inetAsBlob('127.0.0.1')),"
|
|
" blobAsInt(intAsBlob(3)),"
|
|
" blobAsText(textAsBlob('zażółć gęślą jaźń')),"
|
|
" blobAsTimestamp(timestampAsBlob('2001-10-18 14:15:55.134+0000')),"
|
|
" blobAsTimeuuid(timeuuidAsBlob(d2177dd0-eaa2-11de-a572-001b779c76e3)),"
|
|
" blobAsUuid(uuidAsBlob(d2177dd0-eaa2-11de-a572-001b779c76e3)),"
|
|
" blobAsVarchar(varcharAsBlob('varchar')), blobAsVarint(varintAsBlob(123)),"
|
|
" blobAsDecimal(decimalAsBlob(1.23)),"
|
|
" blobAsTinyint(tinyintAsBlob(3)),"
|
|
" blobAsSmallint(smallintAsBlob(3)),"
|
|
" blobAsDate(dateAsBlob('1970-01-02')),"
|
|
" blobAsTime(timeAsBlob('00:00:00.000000001')),"
|
|
" blobAsDuration(durationAsBlob(10y9mo8w7d6h5m4s3ms2us1ns))"
|
|
");").get();
|
|
|
|
{
|
|
auto msg = e.execute_cql("SELECT * FROM all_types WHERE a = 'ascii2'").get();
|
|
struct tm t = {0};
|
|
t.tm_year = 2001 - 1900;
|
|
t.tm_mon = 10 - 1;
|
|
t.tm_mday = 18;
|
|
t.tm_hour = 14;
|
|
t.tm_min = 15;
|
|
t.tm_sec = 55;
|
|
auto tp = db_clock::from_time_t(timegm(&t)) + std::chrono::milliseconds(134);
|
|
assert_that(msg).is_rows().with_rows({
|
|
{
|
|
ascii_type->decompose(sstring("ascii2")), long_type->decompose(123456789l),
|
|
from_hex("000000000000000c"), boolean_type->decompose(true),
|
|
double_type->decompose(3.14), float_type->decompose(3.14f),
|
|
inet_addr_type->decompose(net::inet_address("127.0.0.1")),
|
|
int32_type->decompose(3), utf8_type->decompose(sstring("zażółć gęślą jaźń")),
|
|
timestamp_type->decompose(tp),
|
|
timeuuid_type->decompose(utils::UUID(sstring("d2177dd0-eaa2-11de-a572-001b779c76e3"))),
|
|
uuid_type->decompose(utils::UUID(sstring("d2177dd0-eaa2-11de-a572-001b779c76e3"))),
|
|
utf8_type->decompose(sstring("varchar")), varint_type->decompose(utils::multiprecision_int(123)),
|
|
decimal_type->decompose(big_decimal { 2, boost::multiprecision::cpp_int(123) }),
|
|
byte_type->decompose(int8_t(3)),
|
|
short_type->decompose(int16_t(3)),
|
|
serialized(simple_date_native_type{0x80000001}),
|
|
time_type->decompose(int64_t(0x0000000000000001)),
|
|
duration_type->decompose(cql_duration("10y9mo8w7d6h5m4s3ms2us1ns"))
|
|
}
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_order_by) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table torder (p1 int, c1 int, c2 int, r1 int, r2 int, PRIMARY KEY(p1, c1, c2));").discard_result().get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "torder"));
|
|
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (0, 1, 2, 3);").get();
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (0, 2, 1, 0);").get();
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 order by c1 asc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 order by c1 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
});
|
|
}
|
|
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (0, 1, 1, 4);").get();
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (0, 2, 2, 5);").get();
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 order by c1 desc, c2 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
{int32_type->decompose(1), int32_type->decompose(1), int32_type->decompose(4)},
|
|
});
|
|
}
|
|
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (1, 1, 0, 6);").get();
|
|
e.execute_cql("insert into torder (p1, c1, c2, r1) values (1, 2, 3, 7);").get();
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) order by c1 desc, c2 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(7)},
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
{int32_type->decompose(1), int32_type->decompose(1), int32_type->decompose(4)},
|
|
{int32_type->decompose(1), int32_type->decompose(0), int32_type->decompose(6)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) order by c1 asc, c2 asc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1), int32_type->decompose(0), int32_type->decompose(6)},
|
|
{int32_type->decompose(1), int32_type->decompose(1), int32_type->decompose(4)},
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
{int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(7)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) and c1 < 2 order by c1 desc, c2 desc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) and c1 >= 2 order by c1 asc, c2 asc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) order by c1 desc, c2 desc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(7)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 in (0, 1) order by c1 asc, c2 asc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1), int32_type->decompose(0), int32_type->decompose(6)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 > 1 order by c1 desc, c2 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 >= 2 order by c1 desc, c2 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 >= 2 order by c1 desc, c2 desc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 order by c1 desc, c2 desc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
});
|
|
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 > 1 order by c1 asc, c2 asc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 >= 2 order by c1 asc, c2 asc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
{int32_type->decompose(2), int32_type->decompose(2), int32_type->decompose(5)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 and c1 >= 2 order by c1 asc, c2 asc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2), int32_type->decompose(1), int32_type->decompose(0)},
|
|
});
|
|
}
|
|
|
|
{
|
|
auto msg = e.execute_cql("select c1, c2, r1 from torder where p1 = 0 order by c1 asc, c2 asc limit 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1), int32_type->decompose(1), int32_type->decompose(4)},
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_order_by_validate) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table torderv (p1 int, c1 int, c2 int, r1 int, r2 int, PRIMARY KEY(p1, c1, c2));").discard_result().then([&e] {
|
|
return e.execute_cql("select c2, r1 from torderv where p1 = 0 order by c desc;");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("select c2, r1 from torderv where p1 = 0 order by c2 desc;");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("select c2, r1 from torderv where p1 = 0 order by c1 desc, c2 asc;");
|
|
}).then_wrapped([&e] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("select c2, r1 from torderv order by c1 asc;");
|
|
}).then_wrapped([] (future<shared_ptr<cql_transport::messages::result_message>> f) {
|
|
assert_that_failed(f);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_multi_column_restrictions) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tmcr (p1 int, c1 int, c2 int, c3 int, r1 int, PRIMARY KEY (p1, c1, c2, c3));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tmcr"));
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 0, 0, 0, 0);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 0, 0, 1, 1);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 0, 1, 0, 2);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 0, 1, 1, 3);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 1, 0, 0, 4);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 1, 0, 1, 5);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 1, 1, 0, 6);").get();
|
|
e.execute_cql("insert into tmcr (p1, c1, c2, c3, r1) values (0, 1, 1, 1, 7);").get();
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2, c3) = (0, 1, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2) = (0, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2, c3) in ((0, 1, 0), (1, 0, 1), (0, 1, 0));").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(5)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2) in ((0, 1), (1, 0), (0, 1));").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
{int32_type->decompose(4)},
|
|
{int32_type->decompose(5)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2, c3) >= (1, 0, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(5)},
|
|
{int32_type->decompose(6)},
|
|
{int32_type->decompose(7)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2, c3) >= (0, 1, 1) and (c1, c2, c3) < (1, 1, 0);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(3)},
|
|
{int32_type->decompose(4)},
|
|
{int32_type->decompose(5)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2) >= (0, 1) and (c1, c2, c3) < (1, 0, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
{int32_type->decompose(4)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tmcr where p1 = 0 and (c1, c2, c3) > (0, 1, 0) and (c1, c2) <= (0, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_select_distinct) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tsd (p1 int, c1 int, r1 int, PRIMARY KEY (p1, c1));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tsd"));
|
|
e.execute_cql("insert into tsd (p1, c1, r1) values (0, 0, 0);").get();
|
|
e.execute_cql("insert into tsd (p1, c1, r1) values (1, 1, 1);").get();
|
|
e.execute_cql("insert into tsd (p1, c1, r1) values (1, 1, 2);").get();
|
|
e.execute_cql("insert into tsd (p1, c1, r1) values (2, 2, 2);").get();
|
|
e.execute_cql("insert into tsd (p1, c1, r1) values (2, 3, 3);").get();
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1 from tsd;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(2)});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1 from tsd limit 3;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(2)});
|
|
}
|
|
|
|
e.execute_cql("create table tsd2 (p1 int, p2 int, c1 int, r1 int, PRIMARY KEY ((p1, p2), c1));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tsd2"));
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (0, 0, 0, 0);").get();
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (0, 0, 1, 1);").get();
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (1, 1, 0, 0);").get();
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (1, 1, 1, 1);").get();
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (2, 2, 0, 0);").get();
|
|
e.execute_cql("insert into tsd2 (p1, p2, c1, r1) values (2, 2, 1, 1);").get();
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1, p2 from tsd2;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0), int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(1), int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(2), int32_type->decompose(2)});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1, p2 from tsd2 limit 3;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0), int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(1), int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(2), int32_type->decompose(2)});
|
|
}
|
|
|
|
e.execute_cql("create table tsd3 (p1 int, r1 int, PRIMARY KEY (p1));").get();
|
|
e.execute_cql("insert into tsd3 (p1, r1) values (0, 0);").get();
|
|
e.execute_cql("insert into tsd3 (p1, r1) values (1, 1);").get();
|
|
e.execute_cql("insert into tsd3 (p1, r1) values (1, 2);").get();
|
|
e.execute_cql("insert into tsd3 (p1, r1) values (2, 2);").get();
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1 from tsd3;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(2)});
|
|
}
|
|
|
|
e.execute_cql("create table tsd4 (p1 int, c1 int, s1 int static, r1 int, PRIMARY KEY (p1, c1));").get();
|
|
e.execute_cql("insert into tsd4 (p1, c1, s1, r1) values (0, 0, 0, 0);").get();
|
|
e.execute_cql("insert into tsd4 (p1, c1, r1) values (0, 1, 1);").get();
|
|
e.execute_cql("insert into tsd4 (p1, s1) values (2, 1);").get();
|
|
e.execute_cql("insert into tsd4 (p1, s1) values (3, 2);").get();
|
|
{
|
|
auto msg = e.execute_cql("select distinct p1, s1 from tsd4;").get();
|
|
assert_that(msg).is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(0), int32_type->decompose(0)})
|
|
.with_row({int32_type->decompose(2), int32_type->decompose(1)})
|
|
.with_row({int32_type->decompose(3), int32_type->decompose(2)});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_select_distinct_with_where_clause) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE cf (k int, a int, b int, PRIMARY KEY (k, a))").get();
|
|
for (int i = 0; i < 10; i++) {
|
|
e.execute_cql(format("INSERT INTO cf (k, a, b) VALUES ({:d}, {:d}, {:d})", i, i, i)).get();
|
|
e.execute_cql(format("INSERT INTO cf (k, a, b) VALUES ({:d}, {:d}, {:d})", i, i * 10, i * 10)).get();
|
|
}
|
|
BOOST_REQUIRE_THROW(e.execute_cql("SELECT DISTINCT k FROM cf WHERE a >= 80 ALLOW FILTERING").get(), std::exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("SELECT DISTINCT k FROM cf WHERE k IN (1, 2, 3) AND a = 10").get(), std::exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("SELECT DISTINCT k FROM cf WHERE b = 5").get(), std::exception);
|
|
|
|
assert_that(e.execute_cql("SELECT DISTINCT k FROM cf WHERE k = 1").get())
|
|
.is_rows().with_size(1)
|
|
.with_row({int32_type->decompose(1)});
|
|
|
|
assert_that(e.execute_cql("SELECT DISTINCT k FROM cf WHERE k IN (5, 6, 7)").get())
|
|
.is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(5)})
|
|
.with_row({int32_type->decompose(6)})
|
|
.with_row({int32_type->decompose(7)});
|
|
|
|
// static columns
|
|
e.execute_cql("CREATE TABLE cf2 (k int, a int, s int static, b int, PRIMARY KEY (k, a))").get();
|
|
for (int i = 0; i < 10; i++) {
|
|
e.execute_cql(format("INSERT INTO cf2 (k, a, b, s) VALUES ({:d}, {:d}, {:d}, {:d})", i, i, i, i)).get();
|
|
e.execute_cql(format("INSERT INTO cf2 (k, a, b, s) VALUES ({:d}, {:d}, {:d}, {:d})", i, i * 10, i * 10, i * 10)).get();
|
|
}
|
|
assert_that(e.execute_cql("SELECT DISTINCT s FROM cf2 WHERE k = 5").get())
|
|
.is_rows().with_size(1)
|
|
.with_row({int32_type->decompose(50)});
|
|
|
|
assert_that(e.execute_cql("SELECT DISTINCT s FROM cf2 WHERE k IN (5, 6, 7)").get())
|
|
.is_rows().with_size(3)
|
|
.with_row({int32_type->decompose(50)})
|
|
.with_row({int32_type->decompose(60)})
|
|
.with_row({int32_type->decompose(70)});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_batch_insert_statement) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table cf (p1 varchar, c1 int, r1 int, PRIMARY KEY (p1, c1));").discard_result().then([&e] {
|
|
return e.execute_cql(R"(BEGIN BATCH
|
|
insert into cf (p1, c1, r1) values ('key1', 1, 100);
|
|
insert into cf (p1, c1, r1) values ('key2', 2, 200);
|
|
APPLY BATCH;)"
|
|
).discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql(R"(BEGIN BATCH
|
|
update cf set r1 = 66 where p1 = 'key1' and c1 = 1;
|
|
update cf set r1 = 33 where p1 = 'key2' and c1 = 2;
|
|
APPLY BATCH;)"
|
|
).discard_result();
|
|
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key1")}, {1}, "r1", 66);
|
|
}).then([&e] {
|
|
return require_column_has_value(e, "cf", {sstring("key2")}, {2}, "r1", 33);
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_in_restriction) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tir (p1 int, c1 int, r1 int, PRIMARY KEY (p1, c1));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tir"));
|
|
e.execute_cql("insert into tir (p1, c1, r1) values (0, 0, 0);").get();
|
|
e.execute_cql("insert into tir (p1, c1, r1) values (1, 0, 1);").get();
|
|
e.execute_cql("insert into tir (p1, c1, r1) values (1, 1, 2);").get();
|
|
e.execute_cql("insert into tir (p1, c1, r1) values (1, 2, 3);").get();
|
|
e.execute_cql("insert into tir (p1, c1, r1) values (2, 3, 4);").get();
|
|
{
|
|
auto msg = e.execute_cql("select * from tir where p1 in ();").get();
|
|
assert_that(msg).is_rows().with_size(0);
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tir where p1 in (2, 0, 2, 1);").get();
|
|
assert_that(msg).is_rows().with_rows_ignore_order({
|
|
{int32_type->decompose(4)},
|
|
{int32_type->decompose(0)},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tir where p1 = 1 and c1 in ();").get();
|
|
assert_that(msg).is_rows().with_size(0);
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tir where p1 = 1 and c1 in (2, 0, 2, 1);").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tir where p1 = 1 and c1 in (2, 0, 2, 1) order by c1 desc;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(3)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(1)},
|
|
});
|
|
}
|
|
{
|
|
auto prepared_id = e.prepare("select r1 from tir where p1 in ?;").get();
|
|
auto my_list_type = list_type_impl::get_instance(int32_type, true);
|
|
std::vector<cql3::raw_value> raw_values;
|
|
auto in_values_list = my_list_type->decompose(make_list_value(my_list_type,
|
|
list_type_impl::native_type{{int(2), int(0), int(2), int(1)}}));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(in_values_list));
|
|
auto msg = e.execute_prepared(prepared_id,raw_values).get();
|
|
assert_that(msg).is_rows().with_rows_ignore_order({
|
|
{int32_type->decompose(4)},
|
|
{int32_type->decompose(0)},
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
{int32_type->decompose(3)},
|
|
});
|
|
}
|
|
|
|
e.execute_cql("create table tir2 (p1 int, c1 int, r1 int, PRIMARY KEY (p1, c1,r1));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tir2"));
|
|
e.execute_cql("insert into tir2 (p1, c1, r1) values (0, 0, 0);").get();
|
|
e.execute_cql("insert into tir2 (p1, c1, r1) values (1, 0, 1);").get();
|
|
e.execute_cql("insert into tir2 (p1, c1, r1) values (1, 1, 2);").get();
|
|
e.execute_cql("insert into tir2 (p1, c1, r1) values (1, 2, 3);").get();
|
|
e.execute_cql("insert into tir2 (p1, c1, r1) values (2, 3, 4);").get();
|
|
{
|
|
auto msg = e.execute_cql("select r1 from tir2 where (c1,r1) in ((0, 1),(1,2),(0,1),(1,2),(3,3)) allow filtering;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
});
|
|
}
|
|
{
|
|
auto prepared_id = e.prepare("select r1 from tir2 where (c1,r1) in ? allow filtering;").get();
|
|
auto my_tuple_type = tuple_type_impl::get_instance({int32_type,int32_type});
|
|
auto my_list_type = list_type_impl::get_instance(my_tuple_type, true);
|
|
std::vector<tuple_type_impl::native_type> native_tuples = {
|
|
{int(0), int(1)},
|
|
{int(1), int(2)},
|
|
{int(0), int(1)},
|
|
{int(1), int(2)},
|
|
{int(3), int(3)},
|
|
};
|
|
std::vector<data_value> tuples;
|
|
for (auto&& native_tuple : native_tuples ) {
|
|
tuples.emplace_back(make_tuple_value(my_tuple_type,native_tuple));
|
|
}
|
|
|
|
std::vector<cql3::raw_value> raw_values;
|
|
auto in_values_list = my_list_type->decompose(make_list_value(my_list_type,tuples));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(in_values_list));
|
|
auto msg = e.execute_prepared(prepared_id,raw_values).get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(1)},
|
|
{int32_type->decompose(2)},
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_compact_storage) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("update system.config SET value='true' where name='enable_create_table_with_compact_storage';").get();
|
|
e.execute_cql("create table tcs (p1 int, c1 int, r1 int, PRIMARY KEY (p1, c1)) with compact storage;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tcs"));
|
|
e.execute_cql("insert into tcs (p1, c1, r1) values (1, 2, 3);").get();
|
|
require_column_has_value(e, "tcs", {1}, {2}, "r1", 3).get();
|
|
e.execute_cql("update tcs set r1 = 4 where p1 = 1 and c1 = 2;").get();
|
|
require_column_has_value(e, "tcs", {1}, {2}, "r1", 4).get();
|
|
auto msg = e.execute_cql("select * from tcs where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(4) },
|
|
});
|
|
e.execute_cql("create table tcs2 (p1 int, c1 int, PRIMARY KEY (p1, c1)) with compact storage;").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "tcs2"));
|
|
e.execute_cql("insert into tcs2 (p1, c1) values (1, 2);").get();
|
|
msg = e.execute_cql("select * from tcs2 where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2) },
|
|
});
|
|
e.execute_cql("create table tcs3 (p1 int, c1 int, c2 int, r1 int, PRIMARY KEY (p1, c1, c2)) with compact storage;").get();
|
|
e.execute_cql("insert into tcs3 (p1, c1, c2, r1) values (1, 2, 3, 4);").get();
|
|
e.execute_cql("insert into tcs3 (p1, c1, r1) values (1, 2, 5);").get();
|
|
e.execute_cql("insert into tcs3 (p1, c1, r1) values (1, 3, 6);").get();
|
|
e.execute_cql("insert into tcs3 (p1, c1, c2, r1) values (1, 3, 5, 7);").get();
|
|
e.execute_cql("insert into tcs3 (p1, c1, c2, r1) values (1, 3, blobasint(0x), 8);").get();
|
|
msg = e.execute_cql("select * from tcs3 where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), {}, int32_type->decompose(5) },
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(4) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), {}, int32_type->decompose(6) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), bytes(), int32_type->decompose(8) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), int32_type->decompose(5), int32_type->decompose(7) },
|
|
});
|
|
e.execute_cql("delete from tcs3 where p1 = 1 and c1 = 2;").get();
|
|
msg = e.execute_cql("select * from tcs3 where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(3), {}, int32_type->decompose(6) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), bytes(), int32_type->decompose(8) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), int32_type->decompose(5), int32_type->decompose(7) },
|
|
});
|
|
e.execute_cql("delete from tcs3 where p1 = 1 and c1 = 3 and c2 = 5;").get();
|
|
msg = e.execute_cql("select * from tcs3 where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(3), {}, int32_type->decompose(6) },
|
|
{ int32_type->decompose(1), int32_type->decompose(3), bytes(), int32_type->decompose(8) },
|
|
});
|
|
e.execute_cql("delete from tcs3 where p1 = 1 and c1 = 3 and c2 = blobasint(0x);").get();
|
|
msg = e.execute_cql("select * from tcs3 where p1 = 1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(3), {}, int32_type->decompose(6) },
|
|
});
|
|
e.execute_cql("create table tcs4 (p1 int PRIMARY KEY, c1 int, c2 int) with compact storage;").get();
|
|
e.execute_cql("insert into tcs4 (p1) values (1);").discard_result().get();
|
|
msg = e.execute_cql("select * from tcs4;").get();
|
|
assert_that(msg).is_rows().with_rows({ });
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_collections_of_collections) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto set_of_ints = set_type_impl::get_instance(int32_type, true);
|
|
auto set_of_sets = set_type_impl::get_instance(set_of_ints, true);
|
|
auto map_of_sets = map_type_impl::get_instance(set_of_ints, int32_type, true);
|
|
return e.execute_cql("create table cf_sos (p1 int PRIMARY KEY, v set<frozen<set<int>>>);").discard_result().then([&e] {
|
|
return e.execute_cql("insert into cf_sos (p1, v) values (1, {{1, 2}, {3, 4}, {5, 6}});").discard_result();
|
|
}).then([&e, set_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_sos", {1}, {},
|
|
"v", make_set_value(set_of_sets, set_type_impl::native_type({
|
|
make_set_value(set_of_ints, set_type_impl::native_type({1, 2})),
|
|
make_set_value(set_of_ints, set_type_impl::native_type({3, 4})),
|
|
make_set_value(set_of_ints, set_type_impl::native_type({5, 6})),
|
|
})));
|
|
}).then([&e, set_of_sets, set_of_ints] {
|
|
return e.execute_cql("delete v[{3, 4}] from cf_sos where p1 = 1;").discard_result();
|
|
}).then([&e, set_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_sos", {1}, {},
|
|
"v", make_set_value(set_of_sets, set_type_impl::native_type({
|
|
make_set_value(set_of_ints, set_type_impl::native_type({1, 2})),
|
|
make_set_value(set_of_ints, set_type_impl::native_type({5, 6})),
|
|
})));
|
|
}).then([&e, set_of_sets, set_of_ints] {
|
|
return e.execute_cql("update cf_sos set v = v - {{1, 2}, {5}} where p1 = 1;").discard_result();
|
|
}).then([&e, set_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_sos", {1}, {},
|
|
"v", make_set_value(set_of_sets, set_type_impl::native_type({
|
|
make_set_value(set_of_ints, set_type_impl::native_type({5, 6})),
|
|
})));
|
|
}).then([&e] {
|
|
return e.execute_cql("create table cf_mos (p1 int PRIMARY KEY, v map<frozen<set<int>>, int>);").discard_result();
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return e.execute_cql("insert into cf_mos (p1, v) values (1, {{1, 2}: 7, {3, 4}: 8, {5, 6}: 9});").discard_result();
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_mos", {1}, {},
|
|
"v", make_map_value(map_of_sets, map_type_impl::native_type({
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({1, 2})), 7 },
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({3, 4})), 8 },
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({5, 6})), 9 },
|
|
})));
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return e.execute_cql("delete v[{3, 4}] from cf_mos where p1 = 1;").discard_result();
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_mos", {1}, {},
|
|
"v", make_map_value(map_of_sets, map_type_impl::native_type({
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({1, 2})), 7 },
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({5, 6})), 9 },
|
|
})));
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return e.execute_cql("update cf_mos set v = v - {{1, 2}, {5}} where p1 = 1;").discard_result();
|
|
}).then([&e, map_of_sets, set_of_ints] {
|
|
return require_column_has_value(e, "cf_mos", {1}, {},
|
|
"v", make_map_value(map_of_sets, map_type_impl::native_type({
|
|
{ make_set_value(set_of_ints, set_type_impl::native_type({5, 6})), 9 },
|
|
})));
|
|
});
|
|
});
|
|
}
|
|
|
|
|
|
SEASTAR_TEST_CASE(test_result_order) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("update system.config SET value='true' where name='enable_create_table_with_compact_storage';").get();
|
|
e.execute_cql("create table tro (p1 int, c1 text, r1 int, PRIMARY KEY (p1, c1)) with compact storage;").discard_result().then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'z', 1);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'bbbb', 2);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'a', 3);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'aaa', 4);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'bb', 5);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tro (p1, c1, r1) values (1, 'cccc', 6);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from tro where p1 = 1;");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("a")), int32_type->decompose(3) },
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("aaa")), int32_type->decompose(4) },
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("bb")), int32_type->decompose(5) },
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("bbbb")), int32_type->decompose(2) },
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("cccc")), int32_type->decompose(6) },
|
|
{ int32_type->decompose(1), utf8_type->decompose(sstring("z")), int32_type->decompose(1) },
|
|
});
|
|
}).get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_frozen_collections) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
auto set_of_ints = set_type_impl::get_instance(int32_type, false);
|
|
auto list_of_ints = list_type_impl::get_instance(int32_type, false);
|
|
auto frozen_map_of_set_and_list = map_type_impl::get_instance(set_of_ints, list_of_ints, false);
|
|
return e.execute_cql("CREATE TABLE tfc (a int, b int, c frozen<map<set<int>, list<int>>> static, d int, PRIMARY KEY (a, b));").discard_result().then([&e] {
|
|
return e.execute_cql("INSERT INTO tfc (a, b, c, d) VALUES (0, 0, {}, 0);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("SELECT * FROM tfc;");
|
|
}).then([frozen_map_of_set_and_list] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(0),
|
|
int32_type->decompose(0),
|
|
frozen_map_of_set_and_list->decompose(make_map_value(
|
|
frozen_map_of_set_and_list, map_type_impl::native_type({}))),
|
|
int32_type->decompose(0) },
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
|
|
SEASTAR_TEST_CASE(test_alter_table) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table tat (pk1 int, c1 int, ck2 int, r1 int, r2 int, PRIMARY KEY (pk1, c1, ck2));").discard_result().then([&e] {
|
|
return e.execute_cql("insert into tat (pk1, c1, ck2, r1, r2) values (1, 2, 3, 4, 5);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("alter table tat with comment = 'This is a comment.';").discard_result();
|
|
}).then([&e] {
|
|
BOOST_REQUIRE_EQUAL(e.local_db().find_schema("ks", "tat")->comment(), sstring("This is a comment."));
|
|
return e.execute_cql("alter table tat alter r2 type blob;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select pk1, c1, ck2, r1, r2 from tat;");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(4), int32_type->decompose(5) },
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tat (pk1, c1, ck2, r2) values (1, 2, 3, 0x1234567812345678);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select pk1, c1, ck2, r1, r2 from tat;");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(4), from_hex("1234567812345678") },
|
|
});
|
|
}).then([&e] {
|
|
return e.execute_cql("alter table tat rename pk1 to p1 and ck2 to c2;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select p1, c1, c2, r1, r2 from tat;");
|
|
}).then([&e] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(4), from_hex("1234567812345678") },
|
|
});
|
|
return e.execute_cql("alter table tat add r1_2 int;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into tat (p1, c1, c2, r1_2) values (1, 2, 3, 6);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from tat;");
|
|
}).then([&e] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(4), int32_type->decompose(6), from_hex("1234567812345678") },
|
|
});
|
|
return e.execute_cql("alter table tat drop r1;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from tat;");
|
|
}).then([&e] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), int32_type->decompose(6), from_hex("1234567812345678") },
|
|
});
|
|
return e.execute_cql("alter table tat add r1 int;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from tat;");
|
|
}).then([&e] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), {}, int32_type->decompose(6), from_hex("1234567812345678") },
|
|
});
|
|
return e.execute_cql("alter table tat drop r2;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("alter table tat add r2 int;").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("select * from tat;");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ int32_type->decompose(1), int32_type->decompose(2), int32_type->decompose(3), {}, int32_type->decompose(6), {} },
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_map_query) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE xx (k int PRIMARY KEY, m map<text, int>);").get();
|
|
e.execute_cql("insert into xx (k, m) values (0, {'v2': 1});").get();
|
|
auto m_type = map_type_impl::get_instance(utf8_type, int32_type, true);
|
|
assert_that(e.execute_cql("select m from xx where k = 0;").get())
|
|
.is_rows().with_rows({
|
|
{ make_map_value(m_type, map_type_impl::native_type({{sstring("v2"), 1}})).serialize() }
|
|
});
|
|
e.execute_cql("delete m['v2'] from xx where k = 0;").get();
|
|
assert_that(e.execute_cql("select m from xx where k = 0;").get())
|
|
.is_rows().with_rows({{{}}});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_drop_table) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tmp (pk int, v int, PRIMARY KEY (pk));").get();
|
|
e.execute_cql("drop columnfamily tmp;").get();
|
|
e.execute_cql("create table tmp (pk int, v int, PRIMARY KEY (pk));").get();
|
|
e.execute_cql("drop columnfamily tmp;").get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_reversed_slice_with_empty_range_before_all_rows) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE test (a int, b int, c int, s1 int static, s2 int static, PRIMARY KEY (a, b));").get();
|
|
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 0, 0, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 1, 1, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 2, 2, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 3, 3, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 4, 4, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 5, 5, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 6, 6, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 7, 7, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 8, 8, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 9, 9, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 10, 10, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 11, 11, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 12, 12, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 13, 13, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 14, 14, 17, 42);").get();
|
|
e.execute_cql("INSERT INTO test (a, b, c, s1, s2) VALUES (99, 15, 15, 17, 42);").get();
|
|
|
|
assert_that(e.execute_cql("select * from test WHERE a = 99 and b < 0 ORDER BY b DESC limit 2;").get())
|
|
.is_rows().is_empty();
|
|
|
|
assert_that(e.execute_cql("select * from test WHERE a = 99 order by b desc;").get())
|
|
.is_rows().with_size(16);
|
|
|
|
assert_that(e.execute_cql("select * from test;").get())
|
|
.is_rows().with_size(16);
|
|
});
|
|
}
|
|
|
|
// Test that the sstable layer correctly handles reversed slices, in particular
|
|
// slices that read many clustering ranges, such that there is a large enough gap
|
|
// between the ranges for the reader to attempt to use the promoted index for
|
|
// skipping between them.
|
|
// For this reason, the test writes a large partition (10MB), then issues a
|
|
// reverse query which reads 4 singular clustering ranges from it. The ranges are
|
|
// constructed such that there is many clustering rows between them: roughly 20%
|
|
// which is ~2MB.
|
|
// See #6171
|
|
SEASTAR_TEST_CASE(test_reversed_slice_with_many_clustering_ranges) {
|
|
cql_test_config cfg;
|
|
cfg.db_config->max_memory_for_unlimited_query_soft_limit(std::numeric_limits<uint64_t>::max());
|
|
cfg.db_config->max_memory_for_unlimited_query_hard_limit(std::numeric_limits<uint64_t>::max());
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE test (pk int, ck int, v text, PRIMARY KEY (pk, ck));").get();
|
|
auto id = e.prepare("INSERT INTO test (pk, ck, v) VALUES (?, ?, ?);").get();
|
|
|
|
const int pk = 0;
|
|
const auto raw_pk = int32_type->decompose(data_value(pk));
|
|
const auto cql3_pk = cql3::raw_value::make_value(raw_pk);
|
|
|
|
const auto value = sstring(1024, 'a');
|
|
const auto raw_value = utf8_type->decompose(data_value(value));
|
|
const auto cql3_value = cql3::raw_value::make_value(raw_value);
|
|
|
|
const int num_rows = 10 * 1024;
|
|
|
|
for (int i = 0; i != num_rows; ++i) {
|
|
const auto cql3_ck = cql3::raw_value::make_value(int32_type->decompose(data_value(i)));
|
|
e.execute_prepared(id, {cql3_pk, cql3_ck, cql3_value}).get();
|
|
}
|
|
|
|
e.db().invoke_on_all([] (replica::database& db) {
|
|
return db.flush_all_memtables();
|
|
}).get();
|
|
|
|
const std::vector<int> selected_cks{
|
|
2 * (num_rows / 10),
|
|
4 * (num_rows / 10),
|
|
6 * (num_rows / 10),
|
|
8 * (num_rows / 10)};
|
|
|
|
const auto make_expected_row = [&] (int ck) -> std::vector<bytes_opt> {
|
|
return {raw_pk, int32_type->decompose(ck), raw_value};
|
|
};
|
|
|
|
// Many singular ranges - to check that the right range is used for
|
|
// determining the disk read-range upper bound.
|
|
{
|
|
const auto select_query = fmt::format(
|
|
"SELECT * FROM test WHERE pk = {} and ck IN ({}) ORDER BY ck DESC BYPASS CACHE;",
|
|
pk,
|
|
fmt::join(selected_cks | std::views::transform([] (int ck) { return format("{}", ck); }), ", "));
|
|
assert_that(e.execute_cql(select_query).get())
|
|
.is_rows()
|
|
.with_rows(selected_cks
|
|
| std::views::reverse
|
|
| std::views::transform(make_expected_row)
|
|
| std::ranges::to<std::vector<std::vector<bytes_opt>>>());
|
|
}
|
|
|
|
// A single wide range - to check that the right range bound is used for
|
|
// determining the disk read-range upper bound.
|
|
{
|
|
const auto select_query = format(
|
|
"SELECT * FROM test WHERE pk = {} and ck >= {} and ck <= {} ORDER BY ck DESC BYPASS CACHE;",
|
|
pk,
|
|
selected_cks[0],
|
|
selected_cks[1]);
|
|
|
|
assert_that(e.execute_cql(select_query).get())
|
|
.is_rows()
|
|
.with_rows(std::views::iota(selected_cks[0], selected_cks[1] + 1)
|
|
| std::views::reverse
|
|
| std::views::transform(make_expected_row)
|
|
| std::ranges::to<std::vector<std::vector<bytes_opt>>>());
|
|
}
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_query_with_range_tombstones) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE test (pk int, ck int, v int, PRIMARY KEY (pk, ck));").get();
|
|
|
|
e.execute_cql("INSERT INTO test (pk, ck, v) VALUES (0, 0, 0);").get();
|
|
e.execute_cql("INSERT INTO test (pk, ck, v) VALUES (0, 2, 2);").get();
|
|
e.execute_cql("INSERT INTO test (pk, ck, v) VALUES (0, 4, 4);").get();
|
|
e.execute_cql("INSERT INTO test (pk, ck, v) VALUES (0, 5, 5);").get();
|
|
e.execute_cql("INSERT INTO test (pk, ck, v) VALUES (0, 6, 6);").get();
|
|
|
|
e.execute_cql("DELETE FROM test WHERE pk = 0 AND ck >= 1 AND ck <= 3;").get();
|
|
e.execute_cql("DELETE FROM test WHERE pk = 0 AND ck > 4 AND ck <= 8;").get();
|
|
e.execute_cql("DELETE FROM test WHERE pk = 0 AND ck > 0 AND ck <= 1;").get();
|
|
|
|
assert_that(e.execute_cql("SELECT v FROM test WHERE pk = 0 ORDER BY ck DESC;").get())
|
|
.is_rows()
|
|
.with_rows({
|
|
{ int32_type->decompose(4) },
|
|
{ int32_type->decompose(0) },
|
|
});
|
|
|
|
assert_that(e.execute_cql("SELECT v FROM test WHERE pk = 0;").get())
|
|
.is_rows()
|
|
.with_rows({
|
|
{ int32_type->decompose(0) },
|
|
{ int32_type->decompose(4) },
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_alter_table_validation) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table tatv (p1 int, c1 int, c2 int, r1 int, r2 set<int>, PRIMARY KEY (p1, c1, c2));").discard_result().then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv drop r2;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv add r2 list<int>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv add r2 set<text>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv add r2 set<int>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv rename r2 to r3;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv alter r1 type bigint;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv alter r2 type map<int, int>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv add r3 map<int, int>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv add r4 set<text>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv drop r3;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv drop r4;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv add r3 map<int, text>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv add r4 set<int>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("alter table tatv add r3 map<int, blob>;").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
return e.execute_cql("alter table tatv add r4 set<blob>;").discard_result();
|
|
}).then_wrapped([] (future<> f) {
|
|
SCYLLA_ASSERT(!f.failed());
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_pg_style_string_literal) {
|
|
return do_with_cql_env([] (cql_test_env& e) {
|
|
return e.execute_cql("create table test (p1 text, PRIMARY KEY (p1));").discard_result().then([&e] {
|
|
return e.execute_cql("insert into test (p1) values ($$Apostrophe's$ $ not$ $ '' escaped$$);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into test (p1) values ($$$''valid$_$key$$);").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into test (p1) values ('$normal$valid$$$$key$');").discard_result();
|
|
}).then([&e] {
|
|
return e.execute_cql("insert into test (p1) values ($ $invalid$$);").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("insert into test (p1) values ($$invalid$ $);").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("insert into test (p1) values ($ $invalid$$$);").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("insert into test (p1) values ($$ \n\n$invalid);").discard_result();
|
|
}).then_wrapped([&e] (future<> f) {
|
|
assert_that_failed(f);
|
|
return e.execute_cql("select * from test;");
|
|
}).then([] (shared_ptr<cql_transport::messages::result_message> msg) {
|
|
assert_that(msg).is_rows().with_rows({
|
|
{ utf8_type->decompose(sstring("Apostrophe's$ $ not$ $ '' escaped")) },
|
|
{ utf8_type->decompose(sstring("$''valid$_$key")) },
|
|
{ utf8_type->decompose(sstring("$normal$valid$$$$key$")) },
|
|
});
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_long_text_value) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto prepared = e.execute_cql("CREATE TABLE t (id int PRIMARY KEY, v text, v2 varchar)").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "t"));
|
|
sstring big_one(17324, 'x');
|
|
sstring bigger_one(29123, 'y');
|
|
e.execute_cql(format("INSERT INTO t (id, v, v2) values (1, '{}', '{}')", big_one, big_one)).get();
|
|
e.execute_cql(format("INSERT INTO t (id, v, v2) values (2, '{}', '{}')", bigger_one, bigger_one)).get();
|
|
auto msg = e.execute_cql("select v, v2 from t where id = 1").get();
|
|
assert_that(msg).is_rows().with_rows({{utf8_type->decompose(big_one), utf8_type->decompose(big_one)}});
|
|
msg = e.execute_cql("select v, v2 from t where id = 2").get();
|
|
assert_that(msg).is_rows().with_rows({{utf8_type->decompose(bigger_one), utf8_type->decompose(bigger_one)}});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_time_conversions) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto prepared = e.execute_cql(
|
|
"CREATE TABLE time_data (id timeuuid PRIMARY KEY, d date, ts timestamp);").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "time_data"));
|
|
|
|
e.execute_cql("INSERT INTO time_data (id, d, ts) VALUES (f4e30f80-6958-11e8-96d6-000000000000, '2017-06-11', '2018-06-05 00:00:00+0000');").get();
|
|
|
|
struct tm t = { 0 };
|
|
t.tm_year = 2018 - 1900;
|
|
t.tm_mon = 6 - 1;
|
|
t.tm_mday = 6;
|
|
t.tm_hour = 7;
|
|
t.tm_min = 12;
|
|
t.tm_sec = 22;
|
|
auto tp1 = db_clock::from_time_t(timegm(&t)) + std::chrono::milliseconds(136);
|
|
t.tm_year = 2017 - 1900;
|
|
t.tm_mday = 11;
|
|
t.tm_hour = 0;
|
|
t.tm_min = 0;
|
|
t.tm_sec = 0;
|
|
auto tp2 = db_clock::from_time_t(timegm(&t));
|
|
|
|
auto msg = e.execute_cql("select todate(id), todate(ts), totimestamp(id), totimestamp(d), tounixtimestamp(id),"
|
|
"tounixtimestamp(ts), tounixtimestamp(d), tounixtimestamp(totimestamp(todate(totimestamp(todate(id))))) from time_data;").get();
|
|
assert_that(msg).is_rows().with_rows({{
|
|
serialized(simple_date_native_type{0x80004518}),
|
|
serialized(simple_date_native_type{0x80004517}),
|
|
timestamp_type->decompose(tp1),
|
|
timestamp_type->decompose(tp2),
|
|
long_type->decompose(int64_t(1528269142136)),
|
|
long_type->decompose(int64_t(1528156800000)),
|
|
long_type->decompose(int64_t(1497139200000)),
|
|
long_type->decompose(int64_t(1528243200000))
|
|
}});
|
|
|
|
});
|
|
}
|
|
|
|
// Corner-case test that checks for the paging code's preparedness for an empty
|
|
// range list.
|
|
SEASTAR_TEST_CASE(test_empty_partition_range_scan) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create keyspace empty_partition_range_scan with replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1};").get();
|
|
e.execute_cql("create table empty_partition_range_scan.tb (a int, b int, c int, val int, PRIMARY KEY ((a,b),c) );").get();
|
|
|
|
|
|
auto qo = std::make_unique<cql3::query_options>(db::consistency_level::LOCAL_ONE, std::vector<cql3::raw_value>{},
|
|
cql3::query_options::specific_options{1, nullptr, {}, api::new_timestamp()});
|
|
auto res = e.execute_cql("select * from empty_partition_range_scan.tb where token (a,b) > 1 and token(a,b) <= 1;", std::move(qo)).get();
|
|
assert_that(res).is_rows().is_empty();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_allow_filtering_contains) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t (p frozen<map<text, text>>, c1 frozen<list<int>>, c2 frozen<set<int>>, v map<text, text>, PRIMARY KEY(p, c1, c2));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "t"));
|
|
|
|
e.execute_cql("INSERT INTO t (p, c1, c2, v) VALUES ({'a':'a'}, [1,2,3], {1, 5}, {'x':'xyz', 'y1':'abc'});").get();
|
|
e.execute_cql("INSERT INTO t (p, c1, c2, v) VALUES ({'b':'b'}, [2,3,4], {3, 4}, {'d':'def', 'y1':'abc'});").get();
|
|
e.execute_cql("INSERT INTO t (p, c1, c2, v) VALUES ({'c':'c'}, [3,4,5], {1, 2, 3}, {});").get();
|
|
|
|
auto my_list_type = list_type_impl::get_instance(int32_type, false);
|
|
auto my_set_type = set_type_impl::get_instance(int32_type, false);
|
|
auto my_map_type = map_type_impl::get_instance(utf8_type, utf8_type, false);
|
|
auto my_nonfrozen_map_type = map_type_impl::get_instance(utf8_type, utf8_type, true);
|
|
|
|
auto msg = e.execute_cql("SELECT p FROM t WHERE p CONTAINS KEY 'a' ALLOW FILTERING").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_map_type->decompose(make_map_value(my_map_type, map_type_impl::native_type({{sstring("a"), sstring("a")}})))}
|
|
});
|
|
|
|
msg = e.execute_cql("SELECT c1 FROM t WHERE c1 CONTAINS 3 ALLOW FILTERING").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type({{1, 2, 3}})))},
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type({{2, 3, 4}})))},
|
|
{my_list_type->decompose(make_list_value(my_list_type, list_type_impl::native_type({{3, 4, 5}})))}
|
|
});
|
|
|
|
msg = e.execute_cql("SELECT c2 FROM t WHERE c2 CONTAINS 1 ALLOW FILTERING").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_set_type->decompose(make_set_value(my_set_type, set_type_impl::native_type({{1, 5}})))},
|
|
{my_set_type->decompose(make_set_value(my_set_type, set_type_impl::native_type({{1, 2, 3}})))}
|
|
});
|
|
|
|
msg = e.execute_cql("SELECT v FROM t WHERE v CONTAINS KEY 'y1' ALLOW FILTERING").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{my_nonfrozen_map_type->decompose(make_map_value(my_nonfrozen_map_type, map_type_impl::native_type({{sstring("x"), sstring("xyz")}, {sstring("y1"), sstring("abc")}})))},
|
|
{my_nonfrozen_map_type->decompose(make_map_value(my_nonfrozen_map_type, map_type_impl::native_type({{sstring("d"), sstring("def")}, {sstring("y1"), sstring("abc")}})))}
|
|
});
|
|
|
|
msg = e.execute_cql("SELECT c2, v FROM t WHERE v CONTAINS KEY 'y1' AND c2 CONTAINS 5 ALLOW FILTERING").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{
|
|
my_set_type->decompose(make_set_value(my_set_type, set_type_impl::native_type({{1, 5}}))),
|
|
my_nonfrozen_map_type->decompose(make_map_value(my_nonfrozen_map_type, map_type_impl::native_type({{sstring("x"), sstring("xyz")}, {sstring("y1"), sstring("abc")}})))
|
|
}
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_in_restriction_on_not_last_partition_key) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t (a int,b int,c int,d int,PRIMARY KEY ((a, b), c));").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "t"));
|
|
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,1,1,100); ").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,1,2,200); ").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,1,3,300); ").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,2,1,300); ").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,3,1,1300);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (1,3,2,1400);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (2,3,2,1400);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (2,1,2,1400);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (2,1,3,1300);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (2,2,3,1300);").get();
|
|
e.execute_cql("INSERT INTO t (a,b,c,d) VALUES (3,1,3,1300);").get();
|
|
|
|
{
|
|
auto msg = e.execute_cql("SELECT * FROM t WHERE a IN (1,2) AND b IN (2,3) AND c>=2 AND c<=3;").get();
|
|
assert_that(msg).is_rows().with_rows_ignore_order({
|
|
{
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(1400),
|
|
},
|
|
{
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(1300),
|
|
},
|
|
{
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(1400),
|
|
}
|
|
});
|
|
}
|
|
{
|
|
auto msg = e.execute_cql("SELECT * FROM t WHERE a IN (1,3) AND b=1 AND c>=2 AND c<=3;").get();
|
|
assert_that(msg).is_rows().with_rows_ignore_order({
|
|
{
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(2),
|
|
int32_type->decompose(200),
|
|
},
|
|
{
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(300),
|
|
},
|
|
{
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(1),
|
|
int32_type->decompose(3),
|
|
int32_type->decompose(1300),
|
|
}
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_static_multi_cell_static_lists_with_ckey) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t (p int, c int, slist list<int> static, v int, PRIMARY KEY (p, c));").get();
|
|
e.execute_cql("INSERT INTO t (p, c, slist, v) VALUES (1, 1, [1], 1); ").get();
|
|
|
|
{
|
|
e.execute_cql("UPDATE t SET slist[0] = 3, v = 3 WHERE p = 1 AND c = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({{3}}))) },
|
|
{ int32_type->decompose(3) }
|
|
});
|
|
}
|
|
{
|
|
e.execute_cql("UPDATE t SET slist = [4], v = 4 WHERE p = 1 AND c = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({{4}}))) },
|
|
{ int32_type->decompose(4) }
|
|
});
|
|
}
|
|
{
|
|
e.execute_cql("UPDATE t SET slist = [3] + slist , v = 5 WHERE p = 1 AND c = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({3, 4}))) },
|
|
{ int32_type->decompose(5) }
|
|
});
|
|
}
|
|
{
|
|
e.execute_cql("UPDATE t SET slist = slist + [5] , v = 6 WHERE p = 1 AND c = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({3, 4, 5}))) },
|
|
{ int32_type->decompose(6) }
|
|
});
|
|
}
|
|
{
|
|
e.execute_cql("DELETE slist[2] from t WHERE p = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({3, 4}))) },
|
|
{ int32_type->decompose(6) }
|
|
});
|
|
}
|
|
{
|
|
e.execute_cql("UPDATE t SET slist = slist - [4] , v = 7 WHERE p = 1 AND c = 1;").get();
|
|
auto msg = e.execute_cql("SELECT slist, v FROM t WHERE p = 1 AND c = 1;").get();
|
|
auto slist_type = list_type_impl::get_instance(int32_type, true);
|
|
assert_that(msg).is_rows().with_row({
|
|
{ slist_type->decompose(make_list_value(slist_type, list_type_impl::native_type({data_value(3)}))) },
|
|
{ int32_type->decompose(7) }
|
|
});
|
|
}
|
|
});
|
|
}
|
|
|
|
// std::pow() is not constexpr
|
|
static constexpr int constexpr_int_pow(int x, unsigned y) {
|
|
int ret = 1;
|
|
for (unsigned i = 0; i < y; ++i) {
|
|
ret *= x;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* A class to represent a single multy-column slice expression.
|
|
* The purpose of this class is to provide facilities for testing
|
|
* the multicolumn slice expression. The class uses an abstraction
|
|
* of integer tuples with predefined max values as a counting system
|
|
* with Digits num of digits and Base-1 as the maximum value.
|
|
*
|
|
* The table this class is designed to test is a table with exactly
|
|
* Digits clustering columns of type int and at least one none clustering
|
|
* of type int.
|
|
* The none clustering column should containing the conversion of the
|
|
* clustering key into an int according to base conversion rules where the
|
|
* last clustering key is the smallest digit.
|
|
* This conversion produces a mapping from int to the tuple space spanned
|
|
* by the clustering keys of the table (including prefix tuples).
|
|
*
|
|
* The test case represent a specific slice expression and utility functions to
|
|
* run and validate it.
|
|
* For a usage example see: test_select_with_mixed_order_table
|
|
*/
|
|
template <int Base,int Digits>
|
|
class slice_testcase
|
|
{
|
|
private:
|
|
std::vector<int> _gt_range;
|
|
bool _gt_inclusive;
|
|
std::vector<int> _lt_range;
|
|
bool _lt_inclusive;
|
|
public:
|
|
enum class column_ordering_for_results {
|
|
ASC,
|
|
DESC,
|
|
};
|
|
using ordering_spec = column_ordering_for_results[Digits];
|
|
/**
|
|
* The mapping of tuples is to integers between 0 and this value.
|
|
*/
|
|
static const int total_num_of_values = constexpr_int_pow(Base, Digits);
|
|
|
|
/**
|
|
* Constructor for the testcase
|
|
* @param gt_range - the tuple for the greater than part of the expression
|
|
* as a vector of integers.An empty vector indicates no grater than
|
|
* part.
|
|
* @param gt_inclusive - is the greater than inclusive (>=).
|
|
* @param lt_range - the tuple for the less than part of the expression
|
|
* as a vector of integers.An empty vector indicates no less than
|
|
* part.
|
|
* @param lt_inclusive - is the less than inclusive (<=)
|
|
*/
|
|
slice_testcase(std::vector<int> gt_range, bool gt_inclusive, std::vector<int> lt_range, bool lt_inclusive) {
|
|
_gt_range = gt_range;
|
|
_lt_range = lt_range;
|
|
_gt_inclusive = gt_inclusive;
|
|
_lt_inclusive = lt_inclusive;
|
|
}
|
|
|
|
/**
|
|
* Builds a vector of the expected result assuming a select with
|
|
* this slice expression. The select is assumed to be only on the
|
|
* none clustering column.
|
|
* @return
|
|
*/
|
|
auto genrate_results_for_validation(ordering_spec& orderings) {
|
|
std::vector<std::vector<bytes_opt>> vals;
|
|
for (auto val : generate_results(orderings)) {
|
|
vals.emplace_back(std::vector<bytes_opt>{ int32_type->decompose(val) });
|
|
}
|
|
return vals;
|
|
}
|
|
|
|
/**
|
|
* Generates the actual slice expression that can be embedded
|
|
* into an SQL select query.
|
|
* @param column_names - the mames of the table clustering columns
|
|
* @return the SQL expression as a text.
|
|
*/
|
|
auto generate_cql_slice_expresion(std::vector<std::string> column_names) {
|
|
std::string expression;
|
|
if (!_gt_range.empty()) {
|
|
expression += "(";
|
|
expression += column_names[0];
|
|
for(std::size_t i=1; i < _gt_range.size(); i++) {
|
|
expression += ", "+column_names[i];
|
|
}
|
|
expression += ") ";
|
|
expression += _gt_inclusive ? ">= " : "> ";
|
|
expression += "(";
|
|
expression += std::to_string(_gt_range[0]);
|
|
for(std::size_t i=1; i<_gt_range.size(); i++) {
|
|
expression += ", ";
|
|
expression += std::to_string(_gt_range[i]);
|
|
}
|
|
expression += ")";
|
|
}
|
|
if (!_gt_range.empty() && !_lt_range.empty()) {
|
|
expression += " AND ";
|
|
}
|
|
if (!_lt_range.empty()) {
|
|
expression += "(";
|
|
expression += column_names[0];
|
|
for(std::size_t i=1; i < _lt_range.size(); i++) {
|
|
expression += ", ";
|
|
expression += column_names[i];
|
|
}
|
|
expression += ") ";
|
|
expression += _lt_inclusive ? "<= " : "< ";
|
|
|
|
expression += "(";
|
|
expression += std::to_string(_lt_range[0]);
|
|
for(std::size_t i=1; i < _lt_range.size(); i++) {
|
|
expression += ", ";
|
|
expression += std::to_string(_lt_range[i]);
|
|
}
|
|
expression += ")";
|
|
}
|
|
return expression;
|
|
}
|
|
|
|
/**
|
|
* Maps a tuple of integers to integer
|
|
* @param tuple - the tuple to convert as a vector of integers.
|
|
* @return the integer this tuple is mapped to.
|
|
*/
|
|
static int tuple_to_bound_val(std::vector<int> tuple) {
|
|
int ret = 0;
|
|
int factor = std::pow(Base, Digits-1);
|
|
for (int val : tuple) {
|
|
ret += val * factor;
|
|
factor /= Base;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Maps back from integer space to tuple space.
|
|
* There can be more than one tuple mapped to the same int.
|
|
* There will never be more than one tuple of a certain size
|
|
* that is mapped to the same int.
|
|
* For example: (1) and (1,0) will be mapped to the same integer,
|
|
* but no other tuple of size 1 or 2 will be mapped to this int.
|
|
* @param val - the value to map
|
|
* @param num_componnents - the size of the produced tuple.
|
|
* @return the tuple of the requested size.
|
|
*/
|
|
static auto bound_val_to_tuple(int val, std::size_t num_componnents = Digits) {
|
|
std::vector<int> tuple;
|
|
int factor = std::pow(Base, Digits-1);
|
|
while (tuple.size() < num_componnents) {
|
|
tuple.emplace_back(val/factor);
|
|
val %= factor;
|
|
factor /= Base;
|
|
}
|
|
return tuple;
|
|
}
|
|
private:
|
|
/**
|
|
* A helper function to generate the expected results of a select
|
|
* statement with this slice. The select statement is assumed to
|
|
* select only the none clustering column.
|
|
* @return a vector of integers representing the expected results.
|
|
*/
|
|
std::vector<int> generate_results(ordering_spec& orderings) {
|
|
std::vector<int> vals;
|
|
int start_val = 0;
|
|
int end_val = total_num_of_values -1;
|
|
if (!_gt_range.empty()) {
|
|
start_val = tuple_to_bound_val(_gt_range);
|
|
if (!_gt_inclusive) {
|
|
start_val += std::pow(Base,Digits - _gt_range.size());
|
|
}
|
|
}
|
|
if (!_lt_range.empty()) {
|
|
end_val = tuple_to_bound_val(_lt_range);
|
|
if (!_lt_inclusive) {
|
|
end_val--;
|
|
} else {
|
|
end_val += std::pow(Base, Digits - _lt_range.size()) - 1;
|
|
}
|
|
}
|
|
for (int i=start_val; i<=end_val; i++) {
|
|
vals.emplace_back(i);
|
|
}
|
|
auto comparator_lt = [&orderings] (int lhs, int rhs){
|
|
auto lhs_t = bound_val_to_tuple(lhs);
|
|
auto rhs_t = bound_val_to_tuple(rhs);
|
|
for (int i=0; i < Digits; i++) {
|
|
if (lhs_t[i] == rhs_t[i]) {
|
|
continue ;
|
|
}
|
|
bool lt = (lhs_t[i] < rhs_t[i]);
|
|
// the reason this is correct is because at that point we know that:
|
|
// lhs_t[i] is not less and not equal to rhs_t[i])
|
|
bool gt = !lt;
|
|
bool reverse = orderings[i] == column_ordering_for_results::DESC;
|
|
return (lt && !reverse) || (gt && reverse);
|
|
}
|
|
return false;
|
|
};
|
|
std::sort(vals.begin(), vals.end(), comparator_lt);
|
|
return vals;
|
|
}
|
|
};
|
|
|
|
|
|
SEASTAR_TEST_CASE(test_select_with_mixed_order_table) {
|
|
using slice_test_type = slice_testcase<5,4>;
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
std::string select_query_template = "SELECT f FROM foo WHERE a=0 AND {};";
|
|
std::vector<std::string> column_names = { "b", "c", "d", "e" };
|
|
e.execute_cql("CREATE TABLE foo (a int, b int, c int,d int,e int,f int, PRIMARY KEY (a, b, c, d, e)) WITH CLUSTERING ORDER BY (b DESC, c ASC, d DESC,e ASC);").get();
|
|
BOOST_REQUIRE(e.local_db().has_schema("ks", "foo"));
|
|
|
|
// We convert the range 0-> max mapped integers to the mapped tuple,
|
|
// this will create a table satisfying the slice_testcase assumption.
|
|
for(int i=0; i < slice_test_type::total_num_of_values; i++) {
|
|
auto tuple = slice_test_type::bound_val_to_tuple(i);
|
|
e.execute_cql(format("INSERT INTO foo (a, b, c, d, e, f) VALUES (0, {}, {}, {}, {}, {});",
|
|
tuple[0],tuple[1],tuple[2],tuple[3],i)).get();
|
|
}
|
|
|
|
// a vector to hold all test cases.
|
|
std::vector<slice_test_type> test_cases;
|
|
|
|
//generates all inclusiveness permutations for the specified bounds
|
|
auto generate_with_inclusiveness_permutations = [&test_cases] (std::vector<int> gt_range,std::vector<int> lt_range) {
|
|
if(gt_range.empty() || lt_range.empty()) {
|
|
test_cases.emplace_back(slice_test_type{gt_range, false, lt_range,false});
|
|
test_cases.emplace_back(slice_test_type{gt_range, true, lt_range,true});
|
|
} else {
|
|
for(int i=0; i<=3; i++) {
|
|
test_cases.emplace_back(slice_test_type{gt_range, bool(i&1), lt_range, bool(i&2)});
|
|
}
|
|
}
|
|
};
|
|
|
|
slice_test_type::ordering_spec ordering = {
|
|
slice_test_type::column_ordering_for_results::DESC,
|
|
slice_test_type::column_ordering_for_results::ASC,
|
|
slice_test_type::column_ordering_for_results::DESC,
|
|
slice_test_type::column_ordering_for_results::ASC,
|
|
};
|
|
// no overlap in components equal num of components - (b,c,d,e) >/>= (0,1,2,3) and (b,c,d,e) </<= (1,2,3,4)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{1,2,3,4});
|
|
// overlap in components equal num of components - (b,c,d,e) >/>= (0,1,2,3) and (b,c,d,e) </<= (0,2,2,2)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{0,2,2,2});
|
|
// overlap in components equal num of components - (b,c,d,e) >/>= (0,1,2,3) and (b,c,d,e) </<= (0,1,2,2)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{0,1,2,2});
|
|
// no overlap less compnnents in </<= expression - (b,c,d,e) >/>= (0,1,2,3) and (b,c) </<= (1,2)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{1,2});
|
|
// overlap in compnnents for less components in </<= expression - (b,c,d,e) >/>= (0,1,2,3) and (b,c) </<= (0,2)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{0,2});
|
|
// lt side is a prefix of gt side </<= expression - (b,c,d,e) >/>= (0,1,2,3) and (b,c) </<= (0,1)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{0,1});
|
|
// gt side is a prefix of lt side </<= expression - (b,c,d,e) >/>= (0,1) and (b,c) </<= (0,1,2,3)
|
|
generate_with_inclusiveness_permutations({0,1},{0,1,2,3});
|
|
// no overlap less compnnents in >/>= expression - (b,c) >/>= (0,1) and (b,c,d,e) </<= (1,2,3,4)
|
|
generate_with_inclusiveness_permutations({0,1},{1,2,3,4});
|
|
// overlap in compnnents for less components in >/>= expression - (b,c) >/>= (0,1) and (b,c,d,e) </<= (0,2,3,4)
|
|
generate_with_inclusiveness_permutations({0,1},{0,2,3,4});
|
|
// one sided >/>= 1 expression - (b) >/>= (1)
|
|
generate_with_inclusiveness_permutations({1},{});
|
|
// one sided >/>= partial expression - (b,c) >/>= (0,1)
|
|
generate_with_inclusiveness_permutations({0,1},{});
|
|
// one sided >/>= full expression - (b,c,d,e) >/>= (0,1,2,3)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{});
|
|
// one sided </<= 1 expression - - (b) </<= (3)
|
|
generate_with_inclusiveness_permutations({},{3});
|
|
// one sided </<= partial expression - (b,c) </<= (3,4)
|
|
generate_with_inclusiveness_permutations({},{3,4});
|
|
// one sided </<= full expression - (b,c,d,e) </<= (2,3,4,4)
|
|
generate_with_inclusiveness_permutations({},{2,3,4,4});
|
|
// equality and empty - (b,c,d,e) >/>= (0,1,2,3) and (b,c,d,e) </<= (0,1,2,3)
|
|
generate_with_inclusiveness_permutations({0,1,2,3},{0,1,2,3});
|
|
|
|
for (auto&& test_case : test_cases) {
|
|
auto msg = e.execute_cql(fmt::format(fmt::runtime(select_query_template) ,test_case.generate_cql_slice_expresion(column_names))).get();
|
|
assert_that(msg).is_rows().with_rows(test_case.genrate_results_for_validation(ordering));
|
|
}
|
|
});
|
|
}
|
|
|
|
uint64_t
|
|
run_and_examine_cache_read_stats_change(cql_test_env& e, std::string_view cf_name, std::function<void (cql_test_env& e)> func) {
|
|
auto read_stat = [&] {
|
|
return e.db().map_reduce0([&cf_name] (const replica::database& db) {
|
|
auto& t = db.find_column_family("ks", cf_name);
|
|
auto& stats = t.get_row_cache().stats();
|
|
return stats.reads_with_misses.count() + stats.reads_with_no_misses.count();
|
|
}, uint64_t(0), std::plus<uint64_t>()).get();
|
|
};
|
|
auto before = read_stat();
|
|
func(e);
|
|
auto after = read_stat();
|
|
return after - before;
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_cache_bypass) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE t (k int PRIMARY KEY)").get();
|
|
auto with_cache = run_and_examine_cache_read_stats_change(e, "t", [] (cql_test_env& e) {
|
|
e.execute_cql("SELECT * FROM t").get();
|
|
});
|
|
BOOST_REQUIRE(with_cache >= smp::count); // scan may make multiple passes per shard
|
|
auto without_cache = run_and_examine_cache_read_stats_change(e, "t", [] (cql_test_env& e) {
|
|
e.execute_cql("SELECT * FROM t BYPASS CACHE").get();
|
|
});
|
|
BOOST_REQUIRE_EQUAL(without_cache, 0);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_describe_varchar) {
|
|
// Test that, like cassandra, a varchar column is represented as a text column.
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create table tbl (id int PRIMARY KEY, t text, v varchar);").get();
|
|
auto ks = utf8_type->decompose("ks");
|
|
auto tbl = utf8_type->decompose("tbl");
|
|
auto id = utf8_type->decompose("id");
|
|
auto t = utf8_type->decompose("t");
|
|
auto v = utf8_type->decompose("v");
|
|
auto none = utf8_type->decompose("NONE");
|
|
auto partition_key = utf8_type->decompose("partition_key");
|
|
auto regular = utf8_type->decompose("regular");
|
|
auto pos_0 = int32_type->decompose(0);
|
|
auto pos_m1 = int32_type->decompose(-1);
|
|
auto int_t = utf8_type->decompose("int");
|
|
auto text_t = utf8_type->decompose("text");
|
|
assert_that(e.execute_cql("select * from system_schema.columns where keyspace_name = 'ks';").get())
|
|
.is_rows()
|
|
.with_rows({
|
|
{ks, tbl, id, none, id, partition_key, pos_0, int_t},
|
|
{ks, tbl, t, none, t, regular, pos_m1, text_t},
|
|
{ks, tbl, v, none, v, regular, pos_m1, text_t}
|
|
});
|
|
});
|
|
}
|
|
|
|
namespace {
|
|
|
|
auto B(bool x) { return boolean_type->decompose(x); }
|
|
|
|
auto I(int32_t x) { return int32_type->decompose(x); }
|
|
|
|
auto L(int64_t x) { return long_type->decompose(x); }
|
|
|
|
auto T(const char* t) { return utf8_type->decompose(t); }
|
|
|
|
} // anonymous namespace
|
|
|
|
SEASTAR_TEST_CASE(test_aggregate_and_simple_selection_together) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table t (p int, c int, v int, primary key(p, c))");
|
|
cquery_nofail(e, "insert into t (p, c, v) values (1, 1, 11)");
|
|
cquery_nofail(e, "insert into t (p, c, v) values (1, 2, 12)");
|
|
cquery_nofail(e, "insert into t (p, c, v) values (1, 3, 13)");
|
|
cquery_nofail(e, "insert into t (p, c, v) values (2, 2, 22)");
|
|
require_rows(e, "select c, avg(c) from t", {{I(1), I(2)}});
|
|
require_rows(e, "select p, sum(v) from t", {{I(1), I(58)}});
|
|
require_rows(e, "select p, count(c) from t group by p", {{I(1), L(3)}, {I(2), L(1)}});
|
|
return make_ready_future<>();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_alter_type_on_compact_storage_with_no_regular_columns_does_not_crash) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("update system.config SET value='true' where name='enable_create_table_with_compact_storage';").get();
|
|
cquery_nofail(e, "CREATE TYPE my_udf (first text);");
|
|
cquery_nofail(e, "create table z (pk int, ck frozen<my_udf>, primary key(pk, ck)) with compact storage;");
|
|
cquery_nofail(e, "alter type my_udf add test_int int;");
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_rf_expand) {
|
|
constexpr static auto simple = "org.apache.cassandra.locator.SimpleStrategy";
|
|
constexpr static auto network_topology = "org.apache.cassandra.locator.NetworkTopologyStrategy";
|
|
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto get_replication = [&] (const sstring& ks) {
|
|
auto msg = e.execute_cql(
|
|
format("SELECT JSON replication FROM system_schema.keyspaces WHERE keyspace_name = '{}'", ks)).get();
|
|
auto res = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(msg);
|
|
auto rows = res->rs().result_set().rows();
|
|
BOOST_REQUIRE_EQUAL(rows.size(), 1);
|
|
auto row0 = rows[0];
|
|
BOOST_REQUIRE_EQUAL(row0.size(), 1);
|
|
|
|
auto parsed = rjson::parse(to_string_view(to_bytes(*row0[0])));
|
|
return std::move(rjson::get(parsed, "replication"));
|
|
};
|
|
|
|
auto assert_replication_contains = [&] (const sstring& ks, const std::map<sstring, sstring>& kvs) {
|
|
auto repl = get_replication(ks);
|
|
for (const auto& [k, v] : kvs) {
|
|
BOOST_REQUIRE_EQUAL(v, rjson::to_string_view(rjson::get(repl, k)));
|
|
}
|
|
};
|
|
|
|
// 'replication_factor' option should be translated to datacenter name for NetworkTopologyStrategy
|
|
e.execute_cql(format("CREATE KEYSPACE rf_expand_0 WITH replication = {{'class': '{}', 'replication_factor': 3}}", network_topology)).get();
|
|
assert_replication_contains("rf_expand_0", {
|
|
{"class", network_topology},
|
|
{"datacenter1", "3"}
|
|
});
|
|
|
|
e.execute_cql("CREATE KEYSPACE rf_expand_1 WITH replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 3}").get();
|
|
assert_replication_contains("rf_expand_1", {
|
|
{"class", network_topology},
|
|
{"datacenter1", "3"}
|
|
});
|
|
|
|
// The auto-expansion should not change existing replication factors.
|
|
e.execute_cql("ALTER KEYSPACE rf_expand_1 WITH replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 2}").get();
|
|
assert_replication_contains("rf_expand_1", {
|
|
{"class", network_topology},
|
|
{"datacenter1", "3"}
|
|
});
|
|
|
|
e.execute_cql("CREATE KEYSPACE rf_expand_2 WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 3}").get();
|
|
assert_replication_contains("rf_expand_2", {
|
|
{"class", simple},
|
|
{"replication_factor", "3"}
|
|
});
|
|
|
|
// Should auto-expand when switching from SimpleStrategy to NetworkTopologyStrategy without additional options.
|
|
e.execute_cql("ALTER KEYSPACE rf_expand_2 WITH replication = {'class': 'NetworkTopologyStrategy'}").get();
|
|
assert_replication_contains("rf_expand_2", {
|
|
{"class", network_topology},
|
|
{"datacenter1", "3"}
|
|
});
|
|
|
|
// Respect factors specified manually.
|
|
e.execute_cql("CREATE KEYSPACE rf_expand_3 WITH replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 3, 'datacenter1': 2}").get();
|
|
assert_replication_contains("rf_expand_3", {
|
|
{"class", network_topology},
|
|
{"datacenter1", "2"}
|
|
});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_int_sum_overflow) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, ck text, val int, primary key(pk, ck));");
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p1', 'c1', 2147483647);");
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p1', 'c2', 1);");
|
|
auto sum_query = "select sum(val) from cf;";
|
|
BOOST_REQUIRE_THROW(e.execute_cql(sum_query).get(), exceptions::overflow_error_exception);
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p2', 'c1', -1);");
|
|
auto result = e.execute_cql(sum_query).get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({int32_type->decompose(int32_t(2147483647))});
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p3', 'c1', 2147483647);");
|
|
BOOST_REQUIRE_THROW(e.execute_cql(sum_query).get(), exceptions::overflow_error_exception);
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p3', 'c2', -2147483648);");
|
|
result = e.execute_cql(sum_query).get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({int32_type->decompose(int32_t(2147483646))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bigint_sum_overflow) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, ck text, val bigint, primary key(pk, ck));");
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p1', 'c1', 9223372036854775807);");
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p1', 'c2', 1);");
|
|
auto sum_query = "select sum(val) from cf;";
|
|
BOOST_REQUIRE_THROW(e.execute_cql(sum_query).get(), exceptions::overflow_error_exception);
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p2', 'c1', -1);");
|
|
auto result = e.execute_cql(sum_query).get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(9223372036854775807))});
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p3', 'c1', 9223372036854775807);");
|
|
BOOST_REQUIRE_THROW(e.execute_cql(sum_query).get(), exceptions::overflow_error_exception);
|
|
|
|
cquery_nofail(e, "insert into cf (pk, ck, val) values ('p3', 'c2', -9223372036854775808);");
|
|
result = e.execute_cql(sum_query).get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(9223372036854775806))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bigint_sum) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val bigint, primary key(pk));");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('x', 2147483647);");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('y', 2147483647);");
|
|
auto sum_query = "select sum(val) from cf;";
|
|
assert_that(e.execute_cql(sum_query).get())
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(4294967294))});
|
|
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('z', -4294967295);");
|
|
assert_that(e.execute_cql(sum_query).get())
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(-1))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_int_sum_with_cast) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val int, primary key(pk));");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', 2147483647);");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('b', 2147483647);");
|
|
auto sum_as_bigint_query = "select sum(cast(val as bigint)) from cf;";
|
|
assert_that(e.execute_cql(sum_as_bigint_query).get())
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(4294967294))});
|
|
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', -2147483648);");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('b', -2147483647);");
|
|
assert_that(e.execute_cql(sum_as_bigint_query).get())
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(-4294967295))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_float_sum_overflow) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val float, primary key(pk));");
|
|
testlog.info("make sure we can sum close to the max value");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', 3.4028234e+38);");
|
|
auto result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(3.4028234e+38f)});
|
|
testlog.info("cause overflow");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('b', 1e+38);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(std::numeric_limits<float>::infinity())});
|
|
testlog.info("test maximum negative value");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', -3.4028234e+38);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(-2.4028234e+38f)});
|
|
testlog.info("cause negative overflow");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('c', -2e+38);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(-std::numeric_limits<float>::infinity())});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_double_sum_overflow) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val double, primary key(pk));");
|
|
testlog.info("make sure we can sum close to the max value");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', 1.79769313486231570814527423732e+308);");
|
|
auto result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(1.79769313486231570814527423732E308)});
|
|
testlog.info("cause overflow");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('b', 0.5e+308);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(std::numeric_limits<double>::infinity())});
|
|
testlog.info("test maximum negative value");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', -1.79769313486231570814527423732e+308);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(-1.29769313486231570814527423732e+308)});
|
|
testlog.info("cause negative overflow");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('c', -1e+308);");
|
|
result = e.execute_cql("select sum(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({serialized(-std::numeric_limits<double>::infinity())});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_int_avg) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val int, primary key(pk));");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('a', 2147483647);");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('b', 2147483647);");
|
|
auto result = e.execute_cql("select avg(val) from cf;").get();
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({int32_type->decompose(int32_t(2147483647))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bigint_avg) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "create table cf (pk text, val bigint, primary key(pk));");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('x', 9223372036854775807);");
|
|
cquery_nofail(e, "insert into cf (pk, val) values ('y', 9223372036854775807);");
|
|
assert_that(e.execute_cql("select avg(val) from cf;").get())
|
|
.is_rows()
|
|
.with_size(1)
|
|
.with_row({long_type->decompose(int64_t(9223372036854775807))});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_view_with_two_regular_base_columns_in_key) {
|
|
return do_with_cql_env_thread([] (auto& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (p int, c int, v1 int, v2 int, primary key(p,c))");
|
|
auto schema = e.local_db().find_schema("ks", "t");
|
|
|
|
// Create a CQL-illegal view with two regular base columns in the view key
|
|
schema_builder view_builder("ks", "tv");
|
|
view_builder.with_column(to_bytes("v1"), int32_type, column_kind::partition_key)
|
|
.with_column(to_bytes("v2"), int32_type, column_kind::clustering_key)
|
|
.with_column(to_bytes("p"), int32_type, column_kind::clustering_key)
|
|
.with_column(to_bytes("c"), int32_type, column_kind::clustering_key)
|
|
.with_view_info(schema, false, "v1 IS NOT NULL AND v2 IS NOT NULL AND p IS NOT NULL AND c IS NOT NULL");
|
|
|
|
schema_ptr view_schema = view_builder.build();
|
|
auto& mm = e.migration_manager().local();
|
|
auto group0_guard = mm.start_group0_operation().get();
|
|
auto ts = group0_guard.write_timestamp();
|
|
mm.announce(service::prepare_new_view_announcement(mm.get_storage_proxy(), view_ptr(view_schema), ts).get(), std::move(group0_guard), "").get();
|
|
|
|
// Verify that deleting and restoring columns behaves as expected - i.e. the row is deleted and regenerated
|
|
cquery_nofail(e, "INSERT INTO t (p, c, v1, v2) VALUES (1, 2, 3, 4)");
|
|
auto msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{int32_type->decompose(3), int32_type->decompose(4), int32_type->decompose(1), int32_type->decompose(2)}},
|
|
});
|
|
|
|
cquery_nofail(e, "UPDATE t SET v2 = NULL WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
cquery_nofail(e, "UPDATE t SET v2 = 7 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{int32_type->decompose(3), int32_type->decompose(7), int32_type->decompose(1), int32_type->decompose(2)}},
|
|
});
|
|
|
|
cquery_nofail(e, "UPDATE t SET v1 = NULL WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
|
|
cquery_nofail(e, "UPDATE t SET v1 = 9 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{int32_type->decompose(9), int32_type->decompose(7), int32_type->decompose(1), int32_type->decompose(2)}},
|
|
});
|
|
|
|
cquery_nofail(e, "UPDATE t SET v1 = NULL, v2 = NULL WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
cquery_nofail(e, "UPDATE t SET v1 = 11, v2 = 13 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{int32_type->decompose(11), int32_type->decompose(13), int32_type->decompose(1), int32_type->decompose(2)}},
|
|
});
|
|
|
|
// Reproduce issue #6008 - updates with not-previously-existing row,
|
|
// not setting both v1 and v2 - should not create a view row, and
|
|
// definitely not cause a crash as they did in #6008. Same for
|
|
// deletes when no previous row exists.
|
|
cquery_nofail(e, "DELETE FROM t WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
cquery_nofail(e, "UPDATE t SET v1 = 17 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
cquery_nofail(e, "DELETE FROM t WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
cquery_nofail(e, "UPDATE t SET v2 = 7 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
// Same tests as above, but with a row marker left behind, so there
|
|
// is an existing base row - it's just empty.
|
|
cquery_nofail(e, "INSERT INTO t (p, c, v1, v2) VALUES (1, 2, 3, 4)");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_rows({
|
|
{{int32_type->decompose(3), int32_type->decompose(4), int32_type->decompose(1), int32_type->decompose(2)}},
|
|
});
|
|
cquery_nofail(e, "UPDATE t SET v1 = NULL, v2 = NULL WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
cquery_nofail(e, "UPDATE t SET v1 = 17 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
|
|
cquery_nofail(e, "UPDATE t SET v1 = NULL, v2 = NULL WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
cquery_nofail(e, "UPDATE t SET v2 = 7 WHERE p = 1 AND c = 2");
|
|
msg = cquery_nofail(e, "SELECT * FROM tv");
|
|
assert_that(msg).is_rows().with_size(0);
|
|
});
|
|
}
|
|
|
|
/*!
|
|
* \brief this helper function changes all white space into a single space for
|
|
* a more robust testing.
|
|
*/
|
|
std::string normalize_white_space(const std::string& str) {
|
|
return boost::regex_replace(boost::regex_replace(" " + str + " ", boost::regex("\\s+"), " "), boost::regex(", "), ",");
|
|
}
|
|
|
|
cql_test_config describe_test_config() {
|
|
auto ext = std::make_shared<db::extensions>();
|
|
ext->add_schema_extension<db::tags_extension>(db::tags_extension::NAME);
|
|
ext->add_schema_extension<cdc::cdc_extension>(cdc::cdc_extension::NAME);
|
|
ext->add_schema_extension<db::paxos_grace_seconds_extension>(db::paxos_grace_seconds_extension::NAME);
|
|
ext->add_schema_extension<tombstone_gc_extension>(tombstone_gc_extension::NAME);
|
|
ext->add_schema_extension<db::per_partition_rate_limit_extension>(db::per_partition_rate_limit_extension::NAME);
|
|
|
|
auto cfg = seastar::make_shared<db::config>(ext);
|
|
return cql_test_config(cfg);
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_describe_simple_schema) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
std::unordered_map<std::string, std::string> cql_create_tables {
|
|
{"cf", "CREATE TABLE ks.cf (\n"
|
|
" pk blob,\n"
|
|
" \"COL2\" blob,\n"
|
|
" col1 blob,\n"
|
|
" PRIMARY KEY (pk)\n"
|
|
") WITH bloom_filter_fp_chance = 0.01\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 864000\n"
|
|
" AND max_index_interval = 2048\n"
|
|
" AND memtable_flush_period_in_ms = 0\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"
|
|
},
|
|
{"cf1", "CREATE TABLE ks.cf1 (\n"
|
|
" pk blob,\n"
|
|
" ck blob,\n"
|
|
" col1 blob,\n"
|
|
" col2 blob,\n"
|
|
" PRIMARY KEY (pk, ck)\n"
|
|
") WITH CLUSTERING ORDER BY (ck ASC)\n"
|
|
" AND bloom_filter_fp_chance = 0.01\n"
|
|
" AND caching = {'keys': 'ALL','rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 864000\n"
|
|
" AND max_index_interval = 2048\n"
|
|
" AND memtable_flush_period_in_ms = 0\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"
|
|
},
|
|
{"CF2", "CREATE TABLE ks.\"CF2\" (\n"
|
|
" pk blob,\n"
|
|
" \"CK\" blob,\n"
|
|
" col1 blob,\n"
|
|
" col2 blob,\n"
|
|
" PRIMARY KEY (pk, \"CK\")\n"
|
|
") WITH CLUSTERING ORDER BY (\"CK\" DESC)\n"
|
|
" AND bloom_filter_fp_chance = 0.02\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 954000\n"
|
|
" AND max_index_interval = 3048\n"
|
|
" AND memtable_flush_period_in_ms = 60000\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"
|
|
},
|
|
{"Cf3", "CREATE TABLE ks.\"Cf3\" (\n"
|
|
" pk blob,\n"
|
|
" pk2 blob,\n"
|
|
" \"CK\" blob,\n"
|
|
" col1 blob,\n"
|
|
" col2 blob,\n"
|
|
" PRIMARY KEY ((pk, pk2), \"CK\")\n"
|
|
") WITH CLUSTERING ORDER BY (\"CK\" DESC)\n"
|
|
" AND bloom_filter_fp_chance = 0.02\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 954000\n"
|
|
" AND max_index_interval = 3048\n"
|
|
" AND memtable_flush_period_in_ms = 60000\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"
|
|
},
|
|
{"cf4", "CREATE TABLE ks.cf4 (\n"
|
|
" pk blob,\n"
|
|
" col1 blob,\n"
|
|
" col2 blob,\n"
|
|
" pn phone,\n"
|
|
" PRIMARY KEY (pk)\n"
|
|
") WITH "
|
|
" bloom_filter_fp_chance = 0.02\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 954000\n"
|
|
" AND max_index_interval = 3048\n"
|
|
" AND memtable_flush_period_in_ms = 60000\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"
|
|
}
|
|
|
|
};
|
|
e.execute_cql("CREATE TYPE phone ("
|
|
"country_code int,"
|
|
"number text)"
|
|
).get();
|
|
for (auto &&ct : cql_create_tables) {
|
|
e.execute_cql(ct.second).get();
|
|
auto schema = e.local_db().find_schema("ks", ct.first);
|
|
auto schema_desc = schema->describe(replica::make_schema_describe_helper(schema, e.data_dictionary()), cql3::describe_option::STMTS);
|
|
|
|
BOOST_CHECK_EQUAL(normalize_white_space(schema_desc.create_statement.value().linearize()), normalize_white_space(ct.second));
|
|
}
|
|
}, describe_test_config());
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_describe_view_schema) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
std::string base_table = "CREATE TABLE \"KS\".\"cF\" (\n"
|
|
" pk blob,\n"
|
|
" pk1 blob,\n"
|
|
" ck blob,\n"
|
|
" ck1 blob,\n"
|
|
" \"COL1\" blob,\n"
|
|
" col2 blob,\n"
|
|
" PRIMARY KEY ((pk, pk1), ck, ck1)\n"
|
|
") WITH CLUSTERING ORDER BY (ck ASC, ck1 ASC)\n"
|
|
" AND bloom_filter_fp_chance = 0.01\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 864000\n"
|
|
" AND max_index_interval = 2048\n"
|
|
" AND memtable_flush_period_in_ms = 0\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n";
|
|
|
|
std::unordered_map<std::string, std::string> cql_create_tables {
|
|
{"cf_view", "CREATE MATERIALIZED VIEW \"KS\".cf_view AS\n"
|
|
" SELECT \"COL1\", pk, pk1, ck1, ck\n"
|
|
" FROM \"KS\".\"cF\"\n"
|
|
" WHERE pk IS NOT null AND \"COL1\" IS NOT null AND pk1 IS NOT null AND ck1 IS NOT null AND ck IS NOT null\n"
|
|
" PRIMARY KEY (\"COL1\", pk, pk1, ck1, ck)\n"
|
|
" WITH CLUSTERING ORDER BY (pk ASC, pk1 ASC, ck1 ASC, ck ASC)\n"
|
|
" AND bloom_filter_fp_chance = 0.01\n"
|
|
" AND caching = {'keys': 'ALL', 'rows_per_partition': 'ALL'}\n"
|
|
" AND comment = ''\n"
|
|
" AND compaction = {'class': 'SizeTieredCompactionStrategy'}\n"
|
|
" AND compression = {'sstable_compression': 'org.apache.cassandra.io.compress.LZ4Compressor'}\n"
|
|
" AND crc_check_chance = 1\n"
|
|
" AND default_time_to_live = 0\n"
|
|
" AND gc_grace_seconds = 864000\n"
|
|
" AND max_index_interval = 2048\n"
|
|
" AND memtable_flush_period_in_ms = 0\n"
|
|
" AND min_index_interval = 128\n"
|
|
" AND speculative_retry = '99.0PERCENTILE'\n"
|
|
" AND paxos_grace_seconds = 43200\n"
|
|
" AND tombstone_gc = {'mode': 'timeout', 'propagation_delay_in_seconds': '3600'};\n"},
|
|
{"cf_index_index", "CREATE INDEX cf_index ON \"KS\".\"cF\"(col2);"},
|
|
{"cf_index1_index", "CREATE INDEX cf_index1 ON \"KS\".\"cF\"(pk);"},
|
|
{"cf_index2_index", "CREATE INDEX cf_index2 ON \"KS\".\"cF\"(pk1);"},
|
|
{"cf_index3_index", "CREATE INDEX cf_index3 ON \"KS\".\"cF\"(ck1);"},
|
|
{"cf_index4_index", "CREATE INDEX cf_index4 ON \"KS\".\"cF\"((pk, pk1),col2);"}
|
|
};
|
|
|
|
e.execute_cql("CREATE KEYSPACE \"KS\" WITH replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 3}").get();
|
|
e.execute_cql(base_table).get();
|
|
for (auto &&ct : cql_create_tables) {
|
|
e.execute_cql(ct.second).get();
|
|
auto schema = e.local_db().find_schema("KS", ct.first);
|
|
auto schema_desc = schema->describe(replica::make_schema_describe_helper(schema, e.data_dictionary()), cql3::describe_option::STMTS);
|
|
|
|
BOOST_CHECK_EQUAL(normalize_white_space(schema_desc.create_statement.value().linearize()), normalize_white_space(ct.second));
|
|
|
|
auto base_schema = e.local_db().find_schema("KS", "cF");
|
|
auto base_schema_desc = base_schema->describe(replica::make_schema_describe_helper(base_schema, e.data_dictionary()), cql3::describe_option::STMTS);
|
|
|
|
BOOST_CHECK_EQUAL(normalize_white_space(base_schema_desc.create_statement.value().linearize()), normalize_white_space(base_table));
|
|
}
|
|
}, describe_test_config());
|
|
}
|
|
|
|
shared_ptr<cql_transport::messages::result_message> cql_func_require_nofail(
|
|
cql_test_env& env,
|
|
const seastar::sstring& fct,
|
|
const seastar::sstring& inp,
|
|
std::unique_ptr<cql3::query_options>&& qo = nullptr,
|
|
const std::source_location& loc = std::source_location::current()) {
|
|
auto res = shared_ptr<cql_transport::messages::result_message>(nullptr);
|
|
auto query = format("SELECT {}({}) FROM t;", fct, inp);
|
|
try {
|
|
if (qo) {
|
|
res = env.execute_cql(query, std::move(qo)).get();
|
|
} else {
|
|
res = env.execute_cql(query).get();
|
|
}
|
|
testlog.info("Query '{}' succeeded as expected", query);
|
|
} catch (...) {
|
|
BOOST_ERROR(format("query '{}' failed unexpectedly with error: {}\n{}:{}: originally from here",
|
|
query, std::current_exception(),
|
|
loc.file_name(), loc.line()));
|
|
}
|
|
return res;
|
|
}
|
|
|
|
// FIXME: should be in cql_assertions, but we don't want to call boost from cql_assertions.hh
|
|
template <typename Exception>
|
|
void cql_func_require_throw(
|
|
cql_test_env& env,
|
|
const seastar::sstring& fct,
|
|
const seastar::sstring& inp,
|
|
std::unique_ptr<cql3::query_options>&& qo = nullptr,
|
|
const std::source_location& loc = std::source_location::current()) {
|
|
auto query = format("SELECT {}({}) FROM t;", fct, inp);
|
|
try {
|
|
if (qo) {
|
|
env.execute_cql(query, std::move(qo)).get();
|
|
} else {
|
|
env.execute_cql(query).get();
|
|
}
|
|
BOOST_ERROR(format("query '{}' succeeded unexpectedly\n{}:{}: originally from here", query,
|
|
loc.file_name(), loc.line()));
|
|
} catch (Exception& e) {
|
|
testlog.info("Query '{}' failed as expected with error: {}", query, e);
|
|
} catch (...) {
|
|
BOOST_ERROR(format("query '{}' failed with unexpected error: {}\n{}:{}: originally from here",
|
|
query, std::current_exception(),
|
|
loc.file_name(), loc.line()));
|
|
}
|
|
}
|
|
|
|
static void create_time_uuid_fcts_schema(cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (id int primary key, t timestamp, l bigint, f float, u timeuuid, d date)");
|
|
cquery_nofail(e, "INSERT INTO t (id, t, l, f, u, d) VALUES "
|
|
"(1, 1579072460606, 1579072460606000, 1579072460606, a66525e0-3766-11ea-8080-808080808080, '2020-01-13')");
|
|
cquery_nofail(e, "SELECT * FROM t;");
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_basic_time_uuid_fcts) {
|
|
return do_with_cql_env_thread([] (auto& e) {
|
|
create_time_uuid_fcts_schema(e);
|
|
|
|
cql_func_require_nofail(e, "currenttime", "");
|
|
cql_func_require_nofail(e, "currentdate", "");
|
|
cql_func_require_nofail(e, "now", "");
|
|
cql_func_require_nofail(e, "currenttimeuuid", "");
|
|
cql_func_require_nofail(e, "currenttimestamp", "");
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_time_uuid_fcts_input_validation) {
|
|
return do_with_cql_env_thread([] (auto& e) {
|
|
create_time_uuid_fcts_schema(e);
|
|
|
|
// test timestamp arg
|
|
auto require_timestamp = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "t");
|
|
cql_func_require_throw<exceptions::server_exception>(e, fct, "l");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "f");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "u");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "d");
|
|
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttime()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currentdate()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "now()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttimeuuid()");
|
|
cql_func_require_nofail(e, fct, "currenttimestamp()");
|
|
};
|
|
|
|
require_timestamp("mintimeuuid");
|
|
require_timestamp("maxtimeuuid");
|
|
|
|
// test timeuuid arg
|
|
auto require_timeuuid = [&e] (const sstring& fct) {
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "t");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "l");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "f");
|
|
cql_func_require_nofail(e, fct, "u");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "d");
|
|
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttime()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currentdate()");
|
|
cql_func_require_nofail(e, fct, "now()");
|
|
cql_func_require_nofail(e, fct, "currenttimeuuid()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttimestamp()");
|
|
};
|
|
|
|
require_timeuuid("dateof");
|
|
require_timeuuid("unixtimestampof");
|
|
|
|
// test timeuuid or date arg
|
|
auto require_timeuuid_or_date = [&e] (const sstring& fct) {
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "t");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "l");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "f");
|
|
cql_func_require_nofail(e, fct, "u");
|
|
cql_func_require_nofail(e, fct, "d");
|
|
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttime()");
|
|
cql_func_require_nofail(e, fct, "currentdate()");
|
|
cql_func_require_nofail(e, fct, "now()");
|
|
cql_func_require_nofail(e, fct, "currenttimeuuid()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttimestamp()");
|
|
};
|
|
|
|
require_timeuuid_or_date("totimestamp");
|
|
|
|
// test timestamp or timeuuid arg
|
|
auto require_timestamp_or_timeuuid = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "t");
|
|
cql_func_require_throw<std::exception>(e, fct, "l");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "f");
|
|
cql_func_require_nofail(e, fct, "u");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "d");
|
|
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttime()");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currentdate()");
|
|
cql_func_require_nofail(e, fct, "now()");
|
|
cql_func_require_nofail(e, fct, "currenttimeuuid()");
|
|
cql_func_require_nofail(e, fct, "currenttimestamp()");
|
|
};
|
|
|
|
require_timestamp_or_timeuuid("todate");
|
|
|
|
// test timestamp, timeuuid, or date arg
|
|
auto require_timestamp_timeuuid_or_date = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "t");
|
|
cql_func_require_throw<exceptions::server_exception>(e, fct, "l");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "f");
|
|
cql_func_require_nofail(e, fct, "u");
|
|
cql_func_require_nofail(e, fct, "d");
|
|
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "currenttime()");
|
|
cql_func_require_nofail(e, fct, "currentdate()");
|
|
cql_func_require_nofail(e, fct, "now()");
|
|
cql_func_require_nofail(e, fct, "currenttimeuuid()");
|
|
cql_func_require_nofail(e, fct, "currenttimestamp()");
|
|
};
|
|
|
|
require_timestamp_timeuuid_or_date("tounixtimestamp");
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_time_uuid_fcts_result) {
|
|
return do_with_cql_env_thread([] (auto& e) {
|
|
create_time_uuid_fcts_schema(e);
|
|
|
|
// test timestamp arg
|
|
auto require_timestamp = [&e] (const sstring& fct) {
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "mintimeuuid(t)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "maxtimeuuid(t)");
|
|
cql_func_require_nofail(e, fct, "dateof(u)");
|
|
cql_func_require_nofail(e, fct, "unixtimestampof(u)");
|
|
cql_func_require_nofail(e, fct, "totimestamp(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "todate(u)");
|
|
cql_func_require_nofail(e, fct, "tounixtimestamp(u)");
|
|
};
|
|
|
|
require_timestamp("mintimeuuid");
|
|
require_timestamp("maxtimeuuid");
|
|
|
|
// test timeuuid arg
|
|
auto require_timeuuid = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "mintimeuuid(t)");
|
|
cql_func_require_nofail(e, fct, "maxtimeuuid(t)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "dateof(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "unixtimestampof(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "totimestamp(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "todate(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "tounixtimestamp(u)");
|
|
};
|
|
|
|
require_timeuuid("dateof");
|
|
require_timeuuid("unixtimestampof");
|
|
|
|
// test timeuuid or date arg
|
|
auto require_timeuuid_or_date = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "mintimeuuid(t)");
|
|
cql_func_require_nofail(e, fct, "maxtimeuuid(t)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "dateof(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "unixtimestampof(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "totimestamp(u)");
|
|
cql_func_require_nofail(e, fct, "todate(u)");
|
|
cql_func_require_throw<exceptions::invalid_request_exception>(e, fct, "tounixtimestamp(u)");
|
|
};
|
|
|
|
require_timeuuid_or_date("totimestamp");
|
|
|
|
// test timestamp or timeuuid arg
|
|
auto require_timestamp_or_timeuuid = [] (const sstring& fct) {
|
|
};
|
|
|
|
require_timestamp_or_timeuuid("todate");
|
|
|
|
// test timestamp, timeuuid, or date arg
|
|
auto require_timestamp_timeuuid_or_date = [&e] (const sstring& fct) {
|
|
cql_func_require_nofail(e, fct, "mintimeuuid(t)");
|
|
cql_func_require_nofail(e, fct, "maxtimeuuid(t)");
|
|
cql_func_require_nofail(e, fct, "dateof(u)");
|
|
cql_func_require_nofail(e, fct, "unixtimestampof(u)");
|
|
cql_func_require_nofail(e, fct, "totimestamp(u)");
|
|
cql_func_require_nofail(e, fct, "todate(u)");
|
|
cql_func_require_nofail(e, fct, "tounixtimestamp(u)");
|
|
};
|
|
|
|
require_timestamp_timeuuid_or_date("tounixtimestamp");
|
|
});
|
|
}
|
|
|
|
static std::unique_ptr<cql3::query_options> q_serial_opts(
|
|
std::vector<cql3::raw_value> values,
|
|
db::consistency_level cl) {
|
|
|
|
const auto& so = cql3::query_options::specific_options::DEFAULT;
|
|
auto qo = std::make_unique<cql3::query_options>(
|
|
cl,
|
|
values,
|
|
// Ensure (optional) serial consistency is always specified.
|
|
cql3::query_options::specific_options{
|
|
so.page_size,
|
|
so.state,
|
|
db::consistency_level::SERIAL,
|
|
so.timestamp,
|
|
}
|
|
);
|
|
return qo;
|
|
}
|
|
|
|
// Run parametrized query on the appropriate shard
|
|
static void prepared_on_shard(cql_test_env& e, const sstring& query,
|
|
std::vector<bytes> params,
|
|
std::vector<std::vector<bytes_opt>> expected_rows,
|
|
db::consistency_level cl = db::consistency_level::ONE) {
|
|
auto execute = [&] () mutable {
|
|
return seastar::async([&] () mutable {
|
|
auto id = e.prepare(query).get();
|
|
|
|
std::vector<cql3::raw_value> raw_values;
|
|
for (auto& param : params) {
|
|
raw_values.emplace_back(cql3::raw_value::make_value(param));
|
|
}
|
|
|
|
auto qo = q_serial_opts(std::move(raw_values), cl);
|
|
auto msg = e.execute_prepared_with_qo(id, std::move(qo)).get();
|
|
if (!msg->move_to_shard()) {
|
|
assert_that(msg).is_rows().with_rows_ignore_order(expected_rows);
|
|
}
|
|
return make_foreign(msg);
|
|
});
|
|
};
|
|
|
|
auto msg = execute().get();
|
|
if (msg->move_to_shard()) {
|
|
unsigned shard = *msg->move_to_shard();
|
|
smp::submit_to(shard, std::move(execute)).get();
|
|
}
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_null_value_tuple_floating_types_and_uuids) {
|
|
cql_test_config cfg;
|
|
cfg.need_remote_proxy = true;
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto test_for_single_type = [&e] (const shared_ptr<const abstract_type>& type, auto update_value) {
|
|
cquery_nofail(e, format("CREATE TABLE IF NOT EXISTS t (k int PRIMARY KEY, test {})", type->cql3_type_name()));
|
|
cquery_nofail(e, "INSERT INTO t (k, test) VALUES (0, null)");
|
|
|
|
auto list_type = list_type_impl::get_instance(type, true);
|
|
// decomposed (null) value
|
|
auto arg_value = list_type->decompose(
|
|
make_list_value(list_type, {data_value::make_null(type)}));
|
|
|
|
prepared_on_shard(e, format("UPDATE t SET test={} WHERE k=0 IF test IN ?", update_value),
|
|
{std::move(arg_value)},
|
|
{{boolean_type->decompose(true), std::nullopt}});
|
|
|
|
cquery_nofail(e, "DROP TABLE t");
|
|
};
|
|
|
|
test_for_single_type(double_type, 1.0);
|
|
test_for_single_type(float_type, 1.0f);
|
|
test_for_single_type(uuid_type, utils::make_random_uuid());
|
|
test_for_single_type(timeuuid_type, utils::UUID("00000000-0000-1000-0000-000000000000"));
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_like_parameter_marker) {
|
|
cql_test_config cfg;
|
|
cfg.need_remote_proxy = true;
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (pk int PRIMARY KEY, col text)");
|
|
cquery_nofail(e, "INSERT INTO t (pk, col) VALUES (1, 'aaa')");
|
|
cquery_nofail(e, "INSERT INTO t (pk, col) VALUES (2, 'bbb')");
|
|
cquery_nofail(e, "INSERT INTO t (pk, col) VALUES (3, 'ccc')");
|
|
|
|
const sstring query("UPDATE t SET col = ? WHERE pk = ? IF col LIKE ?");
|
|
prepared_on_shard(e, query, {T("err"), I(9), T("e%")}, {{B(false), {}}});
|
|
prepared_on_shard(e, query, {T("err"), I(9), T("e%")}, {{B(false), {}}});
|
|
prepared_on_shard(e, query, {T("chg"), I(1), T("a%")}, {{B(true), "aaa"}});
|
|
prepared_on_shard(e, query, {T("err"), I(1), T("a%")}, {{B(false), "chg"}});
|
|
prepared_on_shard(e, query, {T("chg"), I(2), T("b%")}, {{B(true), "bbb"}});
|
|
prepared_on_shard(e, query, {T("err"), I(1), T("a%")}, {{B(false), "chg"}});
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_list_parameter_marker) {
|
|
cql_test_config cfg;
|
|
cfg.need_remote_proxy = true;
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (k int PRIMARY KEY, v list<int>)");
|
|
cquery_nofail(e, "INSERT INTO t (k, v) VALUES (1, [10, 20, 30])");
|
|
|
|
const sstring query("UPDATE t SET v=:upd_v WHERE k=1 IF v[:i] in :v");
|
|
|
|
auto list_of = [](const std::vector<data_value>& data) -> bytes {
|
|
auto list_type_int = list_type_impl::get_instance(int32_type, false);
|
|
return make_list_value(list_type_int, data).serialize_nonnull();
|
|
};
|
|
|
|
prepared_on_shard(e,
|
|
query,
|
|
{list_of({100, 200, 300}), I(1), list_of({21, 22, 23})},
|
|
{{B(false), list_of({10, 20, 30})}});
|
|
prepared_on_shard(e, query,
|
|
{list_of({100, 200, 300}), I(1), list_of({20, 21, 22})},
|
|
{{B(true), list_of({10, 20, 30})}});
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_select_serial_consistency) {
|
|
cql_test_config cfg;
|
|
cfg.need_remote_proxy = true;
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (a int, b int, primary key (a,b))");
|
|
cquery_nofail(e, "INSERT INTO t (a, b) VALUES (1, 1)");
|
|
cquery_nofail(e, "INSERT INTO t (a, b) VALUES (1, 2)");
|
|
cquery_nofail(e, "INSERT INTO t (a, b) VALUES (2, 1)");
|
|
cquery_nofail(e, "INSERT INTO t (a, b) VALUES (2, 2)");
|
|
|
|
auto check_fails = [&e] (const sstring& query, const source_location& loc = source_location::current()) {
|
|
try {
|
|
prepared_on_shard(e, query, {}, {}, db::consistency_level::SERIAL);
|
|
SCYLLA_ASSERT(false);
|
|
} catch (const exceptions::invalid_request_exception& e) {
|
|
testlog.info("Query '{}' failed as expected with error: {}", query, e);
|
|
} catch (...) {
|
|
BOOST_ERROR(format("query '{}' failed with unexpected error: {}\n{}:{}: originally from here",
|
|
query, std::current_exception(),
|
|
loc.file_name(), loc.line()));
|
|
}
|
|
};
|
|
check_fails("select * from t allow filtering");
|
|
check_fails("select * from t where b > 0 allow filtering");
|
|
check_fails("select * from t where a in (1, 3)");
|
|
prepared_on_shard(e, "select * from t where a = 1", {}, {{I(1), I(1)}, {I(1), I(2)}}, db::consistency_level::SERIAL);
|
|
}, std::move(cfg));
|
|
}
|
|
|
|
|
|
SEASTAR_TEST_CASE(test_range_deletions_for_specific_column) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (pk int, ck int, col text, PRIMARY KEY(pk, ck))");
|
|
cquery_nofail(e, "INSERT INTO t (pk, ck, col) VALUES (1, 1, 'aaa')");
|
|
cquery_nofail(e, "INSERT INTO t (pk, ck, col) VALUES (1, 2, 'bbb')");
|
|
cquery_nofail(e, "INSERT INTO t (pk, ck, col) VALUES (1, 3, 'ccc')");
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql("DELETE col FROM t WHERE pk = 0 AND ck > 1 AND ck <= 3").get(),
|
|
exceptions::invalid_request_exception);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_alter_table_default_ttl_reset) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a)) WITH default_time_to_live=10").get();
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tbl")->default_time_to_live().count() == 10);
|
|
e.execute_cql("ALTER TABLE tbl WITH gc_grace_seconds=0").get();
|
|
BOOST_REQUIRE(e.local_db().find_schema("ks", "tbl")->default_time_to_live().count() == 10);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_internal_schema_changes_on_a_distributed_table) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
cquery_nofail(e, "create table t (p int primary key, v int)");
|
|
const auto local_err = exception_predicate::message_contains("internal query");
|
|
BOOST_REQUIRE_EXCEPTION(e.local_qp().execute_internal("alter table ks.t add col abcd", cql3::query_processor::cache_internal::yes).get(), std::logic_error, local_err);
|
|
BOOST_REQUIRE_EXCEPTION(e.local_qp().execute_internal("create table ks.t2 (id int primary key)", cql3::query_processor::cache_internal::yes).get(), std::logic_error, local_err);
|
|
BOOST_REQUIRE_EXCEPTION(e.local_qp().execute_internal("create index on ks.t(v)", cql3::query_processor::cache_internal::yes).get(), std::logic_error, local_err);
|
|
BOOST_REQUIRE_EXCEPTION(e.local_qp().execute_internal("drop table ks.t", cql3::query_processor::cache_internal::yes).get(), std::logic_error, local_err);
|
|
BOOST_REQUIRE_EXCEPTION(e.local_qp().execute_internal("drop keyspace ks", cql3::query_processor::cache_internal::yes).get(), std::logic_error, local_err);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_impossible_where) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t(p int PRIMARY KEY, r int)");
|
|
cquery_nofail(e, "INSERT INTO t(p,r) VALUES (0, 0)");
|
|
cquery_nofail(e, "INSERT INTO t(p,r) VALUES (1, 10)");
|
|
cquery_nofail(e, "INSERT INTO t(p,r) VALUES (2, 20)");
|
|
require_rows(e, "SELECT * FROM t WHERE r>10 AND r<10 ALLOW FILTERING", {});
|
|
require_rows(e, "SELECT * FROM t WHERE r>=10 AND r<=0 ALLOW FILTERING", {});
|
|
|
|
cquery_nofail(e, "CREATE TABLE t2(p int, c int, PRIMARY KEY(p, c)) WITH CLUSTERING ORDER BY (c DESC)");
|
|
cquery_nofail(e, "INSERT INTO t2(p,c) VALUES (0, 0)");
|
|
cquery_nofail(e, "INSERT INTO t2(p,c) VALUES (1, 10)");
|
|
cquery_nofail(e, "INSERT INTO t2(p,c) VALUES (2, 20)");
|
|
require_rows(e, "SELECT * FROM t2 WHERE c>10 AND c<10 ALLOW FILTERING", {});
|
|
require_rows(e, "SELECT * FROM t2 WHERE c>=10 AND c<=0 ALLOW FILTERING", {});
|
|
});
|
|
}
|
|
|
|
} // cql_query_test namespace
|
|
|
|
// FIXME: copy-pasta
|
|
static bool has_more_pages(::shared_ptr<cql_transport::messages::result_message> res) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(res);
|
|
BOOST_REQUIRE(rows);
|
|
return rows->rs().get_metadata().flags().contains(cql3::metadata::flag::HAS_MORE_PAGES);
|
|
};
|
|
|
|
static size_t count_rows_fetched(::shared_ptr<cql_transport::messages::result_message> res) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(res);
|
|
BOOST_REQUIRE(rows);
|
|
return rows->rs().result_set().size();
|
|
};
|
|
|
|
static lw_shared_ptr<service::pager::paging_state> extract_paging_state(::shared_ptr<cql_transport::messages::result_message> res) {
|
|
auto rows = dynamic_pointer_cast<cql_transport::messages::result_message::rows>(res);
|
|
BOOST_REQUIRE(rows);
|
|
auto paging_state = rows->rs().get_metadata().paging_state();
|
|
if (!paging_state) {
|
|
return nullptr;
|
|
}
|
|
return make_lw_shared<service::pager::paging_state>(*paging_state);
|
|
};
|
|
|
|
namespace cql_query_test {
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_query_limit) {
|
|
cql_test_config cfg;
|
|
|
|
auto db_config = cfg.db_config;
|
|
|
|
do_with_cql_env_thread([db_config] (cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE test (pk int, ck int, v text, PRIMARY KEY (pk, ck));").get();
|
|
auto id = e.prepare("INSERT INTO test (pk, ck, v) VALUES (?, ?, ?);").get();
|
|
|
|
const int pk = 0;
|
|
const auto raw_pk = int32_type->decompose(data_value(pk));
|
|
const auto cql3_pk = cql3::raw_value::make_value(raw_pk);
|
|
|
|
const auto value = sstring(1024, 'a');
|
|
const auto raw_value = utf8_type->decompose(data_value(value));
|
|
const auto cql3_value = cql3::raw_value::make_value(raw_value);
|
|
|
|
const int num_rows = 10;
|
|
|
|
for (int i = 0; i != num_rows; ++i) {
|
|
const auto cql3_ck = cql3::raw_value::make_value(int32_type->decompose(data_value(i)));
|
|
e.execute_prepared(id, {cql3_pk, cql3_ck, cql3_value}).get();
|
|
}
|
|
|
|
const auto make_expected_row = [&] (int ck) -> std::vector<bytes_opt> {
|
|
return {raw_pk, int32_type->decompose(ck), raw_value};
|
|
};
|
|
|
|
const auto normal_rows = std::views::iota(0, num_rows) | std::views::transform(make_expected_row) | std::ranges::to<std::vector<std::vector<bytes_opt>>>();
|
|
const auto reversed_rows = std::views::iota(0, num_rows) | std::views::reverse | std::views::transform(make_expected_row) | std::ranges::to<std::vector<std::vector<bytes_opt>>>();
|
|
|
|
db_config->max_memory_for_unlimited_query_soft_limit.set(256, utils::config_file::config_source::CommandLine);
|
|
db_config->max_memory_for_unlimited_query_hard_limit.set(1024, utils::config_file::config_source::CommandLine);
|
|
|
|
auto groups = get_scheduling_groups().get();
|
|
|
|
for (auto is_paged : {true, false}) {
|
|
for (auto is_reversed : {true, false}) {
|
|
for (auto scheduling_group : {groups.statement_scheduling_group, groups.streaming_scheduling_group, default_scheduling_group()}) {
|
|
const auto should_fail = !is_paged && scheduling_group == groups.statement_scheduling_group;
|
|
testlog.info("checking: is_paged={}, is_reversed={}, scheduling_group={}, should_fail={}", is_paged, is_reversed, scheduling_group.name(), should_fail);
|
|
const auto select_query = format("SELECT * FROM test WHERE pk = {} ORDER BY ck {};", pk, is_reversed ? "DESC" : "ASC");
|
|
|
|
int32_t page_size = is_paged ? 10000 : -1;
|
|
|
|
const auto& expected_rows = is_reversed ? reversed_rows : normal_rows;
|
|
|
|
try {
|
|
bool has_more_pages = true;
|
|
lw_shared_ptr<service::pager::paging_state> paging_state = nullptr;
|
|
size_t next_expected_row_idx = 0;
|
|
while (has_more_pages) {
|
|
// FIXME: even though we chose a large page size, for reversed queries we may still obtain multiple pages.
|
|
// This happens because the query result size limit for reversed queries is set to the 'unlimited query hard limit'
|
|
// (even though single-partition reversed queries are no longer 'unlimited') which we set to a low value,
|
|
// causing reversed queries to finish early.
|
|
auto qo = std::make_unique<cql3::query_options>(db::consistency_level::LOCAL_ONE, std::vector<cql3::raw_value>{},
|
|
cql3::query_options::specific_options{page_size, paging_state, {}, api::new_timestamp()});
|
|
auto result = with_scheduling_group(scheduling_group, [&e] (const sstring& q, std::unique_ptr<cql3::query_options> qo) {
|
|
return e.execute_cql(q, std::move(qo));
|
|
}, select_query, std::move(qo)).get();
|
|
|
|
auto rows_fetched = count_rows_fetched(result);
|
|
BOOST_REQUIRE(next_expected_row_idx + rows_fetched <= expected_rows.size());
|
|
assert_that(result)
|
|
.is_rows()
|
|
.with_rows(std::vector<std::vector<bytes_opt>>{
|
|
expected_rows.begin() + next_expected_row_idx,
|
|
expected_rows.begin() + next_expected_row_idx + rows_fetched});
|
|
|
|
has_more_pages = ::has_more_pages(result);
|
|
paging_state = extract_paging_state(result);
|
|
BOOST_REQUIRE(!has_more_pages || paging_state);
|
|
next_expected_row_idx += rows_fetched;
|
|
}
|
|
BOOST_REQUIRE(next_expected_row_idx == expected_rows.size());
|
|
|
|
if (should_fail) {
|
|
BOOST_FAIL("Expected exception, but none was thrown.");
|
|
} else {
|
|
testlog.trace("No exception thrown, as expected.");
|
|
}
|
|
} catch (exceptions::read_failure_exception& e) {
|
|
if (should_fail) {
|
|
testlog.trace("Exception thrown, as expected: {}", e);
|
|
} else {
|
|
BOOST_FAIL(fmt::format("Expected no exception, but caught: {}", e));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}, std::move(cfg)).get();
|
|
}
|
|
|
|
// reproduces https://github.com/scylladb/scylla/issues/3552
|
|
// when clustering-key filtering is enabled in filter_sstable_for_reader
|
|
static future<> test_clustering_filtering_with_compaction_strategy(std::string_view cs) {
|
|
auto db_config = make_shared<db::config>();
|
|
db_config->sstable_format("me");
|
|
|
|
return do_with_cql_env_thread([cs] (cql_test_env& e) {
|
|
cquery_nofail(e, seastar::format("CREATE TABLE cf(pk text, ck int, v text, PRIMARY KEY(pk, ck)) WITH COMPACTION = {{'class': '{}'}}", cs));
|
|
cquery_nofail(e, "INSERT INTO cf(pk, ck, v) VALUES ('a', 1, 'a1')");
|
|
e.db().invoke_on_all([] (replica::database& db) { return db.flush_all_memtables(); }).get();
|
|
e.db().invoke_on_all([] (replica::database& db) { db.row_cache_tracker().clear(); }).get();
|
|
require_rows(e, "SELECT v FROM cf WHERE pk='a' AND ck=0 ALLOW FILTERING", {});
|
|
require_rows(e, "SELECT v FROM cf", {{T("a1")}});
|
|
}, cql_test_config(db_config));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_clustering_filtering) {
|
|
static std::array<std::string_view, 2> test_compaction_strategies = {
|
|
"SizeTieredCompactionStrategy",
|
|
"TimeWindowCompactionStrategy",
|
|
};
|
|
|
|
return do_for_each(test_compaction_strategies,
|
|
test_clustering_filtering_with_compaction_strategy);
|
|
}
|
|
|
|
static future<> test_clustering_filtering_2_with_compaction_strategy(std::string_view cs) {
|
|
auto db_config = make_shared<db::config>();
|
|
db_config->sstable_format("me");
|
|
|
|
return do_with_cql_env_thread([cs] (cql_test_env& e) {
|
|
cquery_nofail(e, seastar::format("CREATE TABLE cf(pk text, ck int, v text, PRIMARY KEY(pk, ck)) WITH COMPACTION = {{'class': '{}'}}", cs));
|
|
cquery_nofail(e, "INSERT INTO cf(pk, ck, v) VALUES ('a', 1, 'a1')");
|
|
cquery_nofail(e, "INSERT INTO cf(pk, ck, v) VALUES ('b', 2, 'b2')");
|
|
e.db().invoke_on_all([] (replica::database& db) { return db.flush_all_memtables(); }).get();
|
|
e.db().invoke_on_all([] (replica::database& db) { db.row_cache_tracker().clear(); }).get();
|
|
require_rows(e, "SELECT v FROM cf WHERE pk='a' AND ck=0 ALLOW FILTERING", {});
|
|
require_rows(e, "SELECT v FROM cf", {{T("a1")}, {T("b2")}});
|
|
}, cql_test_config(db_config));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_clustering_filtering_2) {
|
|
static std::array<std::string_view, 2> test_compaction_strategies = {
|
|
"SizeTieredCompactionStrategy",
|
|
"TimeWindowCompactionStrategy",
|
|
};
|
|
|
|
return do_for_each(test_compaction_strategies,
|
|
test_clustering_filtering_2_with_compaction_strategy);
|
|
}
|
|
|
|
static future<> test_clustering_filtering_3_with_compaction_strategy(std::string_view cs) {
|
|
auto db_config = make_shared<db::config>();
|
|
db_config->sstable_format("me");
|
|
|
|
return do_with_cql_env_thread([cs] (cql_test_env& e) {
|
|
cquery_nofail(e, seastar::format("CREATE TABLE cf(pk text, ck int, v text, PRIMARY KEY(pk, ck)) WITH COMPACTION = {{'class': '{}'}}", cs));
|
|
e.db().invoke_on_all([] (replica::database& db) {
|
|
auto& table = db.find_column_family("ks", "cf");
|
|
return table.disable_auto_compaction();
|
|
}).get();
|
|
cquery_nofail(e, "INSERT INTO cf(pk, ck, v) VALUES ('a', 1, 'a1')");
|
|
e.db().invoke_on_all([] (replica::database& db) { return db.flush_all_memtables(); }).get();
|
|
cquery_nofail(e, "INSERT INTO cf(pk, ck, v) VALUES ('b', 0, 'b0')");
|
|
e.db().invoke_on_all([] (replica::database& db) { return db.flush_all_memtables(); }).get();
|
|
e.db().invoke_on_all([] (replica::database& db) { db.row_cache_tracker().clear(); }).get();
|
|
require_rows(e, "SELECT v FROM cf WHERE pk='a' AND ck=0 ALLOW FILTERING", {});
|
|
require_rows(e, "SELECT v FROM cf", {{T("a1")}, {T("b0")}});
|
|
}, cql_test_config(db_config));
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_clustering_filtering_3) {
|
|
static std::array<std::string_view, 2> test_compaction_strategies = {
|
|
"SizeTieredCompactionStrategy",
|
|
"TimeWindowCompactionStrategy",
|
|
};
|
|
|
|
return do_for_each(test_compaction_strategies,
|
|
test_clustering_filtering_3_with_compaction_strategy);
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_counter_column_added_into_non_counter_table) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
cquery_nofail(e, "CREATE TABLE t (pk int, ck int, PRIMARY KEY(pk, ck))");
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER TABLE t ADD \"c\" counter;").get(),
|
|
exceptions::configuration_exception);
|
|
});
|
|
}
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_invalid_using_timestamps) {
|
|
do_with_cql_env_thread([] (cql_test_env& e) {
|
|
auto now_nano = std::chrono::duration_cast<std::chrono::nanoseconds>(db_clock::now().time_since_epoch()).count();
|
|
e.execute_cql("CREATE TABLE tbl (a int, b int, PRIMARY KEY (a))").get();
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql(format("INSERT INTO tbl (a, b) VALUES (1, 1) USING TIMESTAMP {}", now_nano)).get(), exceptions::invalid_request_exception);
|
|
|
|
e.execute_cql("INSERT INTO tbl (a, b) VALUES (1, 1)").get();
|
|
BOOST_REQUIRE_THROW(e.execute_cql(format("UPDATE tbl USING TIMESTAMP {} SET b = 10 WHERE a = 1", now_nano)).get(), exceptions::invalid_request_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(format("DELETE b FROM tbl USING TIMESTAMP {} WHERE a = 1", now_nano)).get(), exceptions::invalid_request_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql(format("BEGIN BATCH USING TIMESTAMP {} INSERT INTO TBL (a, b) VALUES (2, 2); APPLY BATCH", now_nano)).get(), exceptions::invalid_request_exception);
|
|
}).get();
|
|
}
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_twcs_non_optimal_query_path) {
|
|
do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql(
|
|
"CREATE TABLE tbl (pk int, ck int, v int, s int static, PRIMARY KEY (pk, ck))"
|
|
" WITH compaction = {"
|
|
" 'compaction_window_size': '1',"
|
|
" 'compaction_window_unit': 'MINUTES',"
|
|
" 'class': 'org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy'"
|
|
"}").get();
|
|
|
|
e.execute_cql("INSERT INTO tbl (pk, ck, v) VALUES (0, 0, 0)").get();
|
|
e.execute_cql("INSERT INTO tbl (pk, ck, v) VALUES (1, 0, 0)").get();
|
|
e.execute_cql("INSERT INTO tbl (pk, ck, v) VALUES (0, 1, 0)").get();
|
|
e.execute_cql("INSERT INTO tbl (pk, ck, v) VALUES (1, 1, 0)").get();
|
|
|
|
e.db().invoke_on_all([] (replica::database& db) {
|
|
return db.flush_all_memtables();
|
|
}).get();
|
|
|
|
// Not really testing anything, just ensure that we execute the non-optimal
|
|
// sstable read path of TWCS tables too, allowing ASAN to shake out any memory
|
|
// related bugs.
|
|
// The non-optimal read path is triggered by having a static column, see
|
|
// `sstables::time_series_sstable_set::create_single_key_sstable_reader()`.
|
|
e.execute_cql("SELECT * FROM tbl WHERE pk = 0 BYPASS CACHE").get();
|
|
}).get();
|
|
}
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_twcs_optimal_query_path) {
|
|
do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql(
|
|
"CREATE TABLE tbl (pk int, ck int, v int, PRIMARY KEY (pk, ck))"
|
|
" WITH compaction = {"
|
|
" 'compaction_window_size': '1',"
|
|
" 'compaction_window_unit': 'MINUTES',"
|
|
" 'class': 'org.apache.cassandra.db.compaction.TimeWindowCompactionStrategy'"
|
|
"}").get();
|
|
|
|
e.execute_cql("INSERT INTO tbl (pk, ck, v) VALUES (0, 0, 0)").get();
|
|
|
|
e.db().invoke_on_all([] (replica::database& db) {
|
|
return db.flush_all_memtables();
|
|
}).get();
|
|
|
|
// Not really testing anything, just ensure that we execute the optimal
|
|
// sstable read path of TWCS tables too, allowing ASAN to shake out any memory
|
|
// related bugs.
|
|
assert_that(e.execute_cql("SELECT * FROM tbl WHERE pk = 0 BYPASS CACHE").get())
|
|
.is_rows().with_size(1);
|
|
}).get();
|
|
}
|
|
|
|
SEASTAR_THREAD_TEST_CASE(test_query_unselected_columns) {
|
|
cql_test_config cfg;
|
|
|
|
cfg.db_config->max_memory_for_unlimited_query_soft_limit(1024 * 1024, utils::config_file::config_source::CommandLine);
|
|
cfg.db_config->max_memory_for_unlimited_query_hard_limit(1024 * 1024, utils::config_file::config_source::CommandLine);
|
|
|
|
cfg.dbcfg.emplace();
|
|
cfg.dbcfg->available_memory = memory::stats().total_memory();
|
|
|
|
do_with_cql_env_thread([] (cql_test_env& e) {
|
|
// Sanity test, this test-case has to run in the statement group that is != default group.
|
|
auto groups = get_scheduling_groups().get();
|
|
BOOST_REQUIRE(groups.statement_scheduling_group == current_scheduling_group());
|
|
BOOST_REQUIRE(default_scheduling_group() != current_scheduling_group());
|
|
|
|
e.execute_cql("CREATE TABLE tbl (pk int, ck int, v text, PRIMARY KEY (pk, ck))").get();
|
|
|
|
const int num_rows = 20;
|
|
const sstring val(100 * 1024, 'a');
|
|
const auto id = e.prepare(format("INSERT INTO tbl (pk, ck, v) VALUES (0, ?, '{}')", val)).get();
|
|
for (int ck = 0; ck < num_rows; ++ck) {
|
|
e.execute_prepared(id, {cql3::raw_value::make_value(int32_type->decompose(ck))}).get();
|
|
}
|
|
|
|
{
|
|
testlog.info("Single partition scan");
|
|
auto qo = std::make_unique<cql3::query_options>(db::consistency_level::LOCAL_ONE, std::vector<cql3::raw_value>{},
|
|
cql3::query_options::specific_options{-1, nullptr, {}, api::new_timestamp()});
|
|
assert_that(e.execute_cql("SELECT pk, ck FROM tbl WHERE pk = 0", std::move(qo)).get()).is_rows().with_size(num_rows);
|
|
}
|
|
|
|
{
|
|
testlog.info("Full scan");
|
|
auto qo = std::make_unique<cql3::query_options>(db::consistency_level::LOCAL_ONE, std::vector<cql3::raw_value>{},
|
|
cql3::query_options::specific_options{-1, nullptr, {}, api::new_timestamp()});
|
|
assert_that(e.execute_cql("SELECT pk, ck FROM tbl", std::move(qo)).get()).is_rows().with_size(num_rows);
|
|
}
|
|
}, std::move(cfg)).get();
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_user_based_sla_queries) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
// test create service level with defaults
|
|
e.execute_cql("CREATE SERVICE_LEVEL sl_1;").get();
|
|
auto msg = e.execute_cql("LIST SERVICE_LEVEL sl_1;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("sl_1"), {}, {}, int32_type->decompose(1000)},
|
|
});
|
|
//create and alter service levels
|
|
e.execute_cql("CREATE SERVICE_LEVEL sl_2 WITH SHARES = 200;").get();
|
|
e.execute_cql("ALTER SERVICE_LEVEL sl_1 WITH SHARES = 111;").get();
|
|
msg = e.execute_cql("LIST ALL SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("driver"), {}, {utf8_type->decompose("batch")}, int32_type->decompose(200), utf8_type->decompose("39.14%")},
|
|
{utf8_type->decompose("sl_1"), {}, {}, int32_type->decompose(111), utf8_type->decompose("21.72%")},
|
|
{utf8_type->decompose("sl_2"), {}, {}, int32_type->decompose(200), utf8_type->decompose("39.14%")},
|
|
});
|
|
//drop service levels
|
|
e.execute_cql("DROP SERVICE_LEVEL sl_1;").get();
|
|
msg = e.execute_cql("LIST ALL SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("driver"), {}, {utf8_type->decompose("batch")}, int32_type->decompose(200), utf8_type->decompose("50.00%")},
|
|
{utf8_type->decompose("sl_2"), {}, {}, int32_type->decompose(200), utf8_type->decompose("50.00%")},
|
|
});
|
|
|
|
// validate exceptions (illegal requests)
|
|
BOOST_REQUIRE_THROW(e.execute_cql("DROP SERVICE_LEVEL sl_1;").get(), qos::nonexistant_service_level_exception);
|
|
e.execute_cql("DROP SERVICE_LEVEL IF EXISTS sl_1;").get();
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE SERVICE_LEVEL sl_2;").get(), exceptions::invalid_request_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE SERVICE_LEVEL sl_2 WITH SHARES = 999;").get(), exceptions::invalid_request_exception);
|
|
e.execute_cql("CREATE SERVICE_LEVEL IF NOT EXISTS sl_2;").get();
|
|
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE SERVICE_LEVEL sl_1 WITH SHARES = 0;").get(), exceptions::syntax_exception);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("CREATE SERVICE_LEVEL sl_1 WITH SHARES = 1001;").get(), exceptions::syntax_exception);
|
|
|
|
// test attach role
|
|
e.execute_cql("ATTACH SERVICE_LEVEL sl_2 TO tester").get();
|
|
msg = e.execute_cql("LIST ATTACHED SERVICE_LEVEL OF tester;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("tester"), utf8_type->decompose("sl_2")},
|
|
});
|
|
msg = e.execute_cql("LIST ALL ATTACHED SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("tester"), utf8_type->decompose("sl_2")},
|
|
});
|
|
|
|
// test attachment illegal request
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ATTACH SERVICE_LEVEL sl_2 TO tester2;").get(), auth::nonexistant_role);
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ATTACH SERVICE_LEVEL sl_1 TO tester;").get(), qos::nonexistant_service_level_exception);
|
|
BOOST_CHECK(true);
|
|
// tests detaching service levels
|
|
e.execute_cql("CREATE ROLE tester2;").get();
|
|
e.execute_cql("CREATE SERVICE_LEVEL sl_1 WITH SHARES = 998;").get();
|
|
e.execute_cql("ATTACH SERVICE_LEVEL sl_1 TO tester2;").get();
|
|
e.execute_cql("DETACH SERVICE_LEVEL FROM tester;").get();
|
|
msg = e.execute_cql("LIST ATTACHED SERVICE_LEVEL OF tester2;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("tester2"), utf8_type->decompose("sl_1")},
|
|
});
|
|
BOOST_CHECK(true);
|
|
msg = e.execute_cql("LIST ATTACHED SERVICE_LEVEL OF tester;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
});
|
|
BOOST_CHECK(true);
|
|
msg = e.execute_cql("LIST ALL ATTACHED SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("tester2"), utf8_type->decompose("sl_1")},
|
|
});
|
|
BOOST_CHECK(true);
|
|
//test implicit detach when removing role
|
|
e.execute_cql("DROP ROLE tester2;").get();
|
|
msg = e.execute_cql("LIST ALL ATTACHED SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
});
|
|
BOOST_CHECK(true);
|
|
e.execute_cql("ATTACH SERVICE_LEVEL sl_1 TO tester;").get();
|
|
msg = e.execute_cql("LIST ALL ATTACHED SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{utf8_type->decompose("tester"), utf8_type->decompose("sl_1")},
|
|
});
|
|
BOOST_CHECK(true);
|
|
//test implicit detach when removing service level
|
|
e.execute_cql("DROP SERVICE_LEVEL sl_1;").get();
|
|
msg = e.execute_cql("LIST ALL ATTACHED SERVICE_LEVELS;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
});
|
|
BOOST_REQUIRE_THROW(e.execute_cql("ALTER SERVICE_LEVEL i_do_not_exist WITH shares = 1;").get(), exceptions::invalid_request_exception);
|
|
});
|
|
}
|
|
|
|
// Check that `current*()` CQL functions are re-evaluated on each execute
|
|
// even for prepared statements.
|
|
// Refs: #8816 (https://github.com/scylladb/scylla/issues/8816)
|
|
SEASTAR_TEST_CASE(timeuuid_fcts_prepared_re_evaluation) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
// We don't test the `currentdate()` function since we don't
|
|
// have a way to supply a synthetic clock to query processor, hence
|
|
// can't move one day forward to demonstrate the desired behavior.
|
|
const std::vector<std::pair<std::string, std::string>> sub_tests = {
|
|
{"currenttimestamp", "timestamp"},
|
|
{"currenttime", "time"},
|
|
{"currenttimeuuid", "timeuuid"}
|
|
};
|
|
for (const auto& t : sub_tests) {
|
|
BOOST_TEST_CHECKPOINT(t.first);
|
|
e.execute_cql(seastar::format("CREATE TABLE test_{} (pk {} PRIMARY KEY)", t.first, t.second)).get();
|
|
auto drop_test_table = defer([&e, &t] { e.execute_cql(seastar::format("DROP TABLE test_{}", t.first)).get(); });
|
|
auto insert_stmt = e.prepare(seastar::format("INSERT INTO test_{0} (pk) VALUES ({0}())", t.first)).get();
|
|
e.execute_prepared(insert_stmt, {}).get();
|
|
sleep(1ms).get();
|
|
// Check that the second execution is evaluated again and yields a
|
|
// different value.
|
|
e.execute_prepared(insert_stmt, {}).get();
|
|
auto msg = e.execute_cql(seastar::format("SELECT * FROM test_{}", t.first)).get();
|
|
assert_that(msg).is_rows().with_size(2);
|
|
}
|
|
});
|
|
}
|
|
|
|
static future<> with_parallelized_aggregation_enabled_thread(std::function<void(cql_test_env&)>&& func) {
|
|
auto db_cfg_ptr = make_shared<db::config>();
|
|
auto& db_cfg = *db_cfg_ptr;
|
|
db_cfg.enable_parallelized_aggregation({true}, db::config::config_source::CommandLine);
|
|
return do_with_cql_env_thread(std::forward<std::function<void(cql_test_env&)>>(func), db_cfg_ptr);
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_count) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, v int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k, v) VALUES ({:d}, {:d});", i, i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT COUNT(*) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(value_count))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_min) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT MIN(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(int32_t(0))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_max) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT MAX(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(int32_t(value_count - 1))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_sum) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT SUM(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(int32_t((value_count - 1) * value_count / 2))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_non_parallelized_multiple_select) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT MIN(k), MAX(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(int32_t(0)), int32_type->decompose(int32_t(value_count - 1))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_sum_group_by) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, c int, v int, PRIMARY KEY (k, c));").get();
|
|
int value_count = 10;
|
|
for (int k = 0; k < 2; k++) {
|
|
for (int c = 0; c < value_count; c++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k, c, v) VALUES ({:d}, {:d}, {:d});", k, c, c)).get();
|
|
}
|
|
}
|
|
|
|
auto msg = e.execute_cql("SELECT k, SUM(v) FROM tbl GROUP BY k;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{int32_type->decompose(int32_t(1)), int32_type->decompose(int32_t((value_count - 1) * value_count / 2))},
|
|
{int32_type->decompose(int32_t(0)), int32_type->decompose(int32_t((value_count - 1) * value_count / 2))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_counter_type) {
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, c counter, PRIMARY KEY (k));").get();
|
|
e.execute_cql("UPDATE tbl SET c = c + 4 WHERE k = 0;").get();
|
|
e.execute_cql("UPDATE tbl SET c = c + 2 WHERE k = 1;").get();
|
|
|
|
auto msg_sum = e.execute_cql("SELECT SUM(c) FROM tbl;").get();
|
|
assert_that(msg_sum).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(6))}
|
|
});
|
|
auto msg_min = e.execute_cql("SELECT MIN(c) FROM tbl;").get();
|
|
assert_that(msg_min).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(2))}
|
|
});
|
|
auto msg_max = e.execute_cql("SELECT MAX(c) FROM tbl;").get();
|
|
assert_that(msg_max).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(4))}
|
|
});
|
|
auto msg_avg = e.execute_cql("SELECT AVG(c) FROM tbl;").get();
|
|
assert_that(msg_avg).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(3))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 4, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_single_partition_aggregation_is_not_parallelized) {
|
|
// It's pointless from performance pov to parallelize
|
|
// aggregation queries which reads only single partition.
|
|
|
|
return with_parallelized_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
const auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE TABLE tbl (pk int, ck int, col int, PRIMARY KEY (pk, ck));").get();
|
|
const int value_count = 10;
|
|
for (int pk = 0; pk < 2; pk++) {
|
|
for (int c = 0; c < value_count; c++) {
|
|
e.execute_cql(format("INSERT INTO tbl (pk, ck, col) VALUES ({:d}, {:d}, {:d});", pk, c, c)).get();
|
|
}
|
|
}
|
|
|
|
const auto result1 = e.execute_cql("SELECT COUNT(*) FROM tbl WHERE pk = 1;").get();
|
|
assert_that(result1).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(value_count))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
|
|
const auto result2 = e.execute_cql("SELECT COUNT(*) FROM tbl WHERE pk = 1 AND ck = 1;").get();
|
|
assert_that(result2).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(1))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
|
|
const auto result3 = e.execute_cql("SELECT COUNT(*) FROM tbl WHERE token(pk) = 1;").get();
|
|
// We don't check value of count(*) here but only if it wasn't parallelized
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
|
|
const auto result4 = e.execute_cql("SELECT COUNT(*) FROM tbl WHERE pk = 1 AND pk = 2;").get();
|
|
assert_that(result4).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(0))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
|
|
|
|
e.execute_cql("CREATE TABLE tbl2 (pk1 int, pk2 int, ck int, col int, PRIMARY KEY((pk1, pk2), ck));").get();
|
|
for (int pk1 = 0; pk1 < 2; pk1++) {
|
|
for (int pk2 = 0; pk2 < 2; pk2++) {
|
|
for (int c = 0; c < value_count; c++) {
|
|
e.execute_cql(format("INSERT INTO tbl2 (pk1, pk2, ck, col) VALUES ({:d}, {:d}, {:d}, {:d});", pk1, pk2, c, c)).get();
|
|
}
|
|
}
|
|
}
|
|
|
|
const auto result_pk12 = e.execute_cql("SELECT COUNT(*) FROM tbl2 WHERE pk1 = 1 AND pk2 = 0;").get();
|
|
assert_that(result_pk12).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(value_count))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
|
|
// Query with only partly restricted partition key requires `ALLOW FILTERING` clause
|
|
// and we doesn't parallelize queries which need filtering.
|
|
// See issue #19369.
|
|
const auto result_pk1 = e.execute_cql("SELECT COUNT(*) FROM tbl2 WHERE pk1 = 1 ALLOW FILTERING;").get();
|
|
// This query contains also column for pk1
|
|
assert_that(result_pk1).is_rows().with_rows({
|
|
{long_type->decompose(int64_t(value_count * 2)), int32_type->decompose(int32_t(1))}
|
|
});
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
static future<> with_udf_and_parallel_aggregation_enabled_thread(std::function<void(cql_test_env&)>&& func) {
|
|
auto db_cfg_ptr = make_shared<db::config>();
|
|
auto& db_cfg = *db_cfg_ptr;
|
|
db_cfg.enable_user_defined_functions({true}, db::config::config_source::CommandLine);
|
|
db_cfg.user_defined_function_time_limit_ms(1000);
|
|
db_cfg.experimental_features({db::experimental_features_t::feature::UDF}, db::config::config_source::CommandLine);
|
|
db_cfg.enable_parallelized_aggregation({true}, db::config::config_source::CommandLine);
|
|
return do_with_cql_env_thread(std::forward<std::function<void(cql_test_env&)>>(func), db_cfg_ptr);
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_parallelized_select_uda) {
|
|
return with_udf_and_parallel_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE FUNCTION row_fct(acc bigint, val int) "
|
|
"RETURNS NULL ON NULL INPUT "
|
|
"RETURNS bigint "
|
|
"LANGUAGE lua "
|
|
"AS $$ "
|
|
"return acc+val "
|
|
"$$;").get();
|
|
e.execute_cql("CREATE FUNCTION reduce_fct(acc1 bigint, acc2 bigint) "
|
|
"RETURNS NULL ON NULL INPUT "
|
|
"RETURNS bigint "
|
|
"LANGUAGE lua "
|
|
"AS $$ "
|
|
"return acc1+acc2 "
|
|
"$$;").get();
|
|
e.execute_cql("CREATE FUNCTION final_fct(acc bigint) "
|
|
"RETURNS NULL ON NULL INPUT "
|
|
"RETURNS bigint "
|
|
"LANGUAGE lua "
|
|
"AS $$ "
|
|
"return -acc "
|
|
"$$;").get();
|
|
e.execute_cql("CREATE AGGREGATE aggr(int) "
|
|
"SFUNC row_fct "
|
|
"STYPE bigint "
|
|
"REDUCEFUNC reduce_fct "
|
|
"FINALFUNC final_fct "
|
|
"INITCOND 0;").get();
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT aggr(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{long_type->decompose(-int64_t((value_count - 1) * value_count / 2))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized + 1, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_not_parallelized_select_uda) {
|
|
return with_udf_and_parallel_aggregation_enabled_thread([](cql_test_env& e) {
|
|
auto& qp = e.local_qp();
|
|
auto stat_parallelized = qp.get_cql_stats().select_parallelized;
|
|
|
|
e.execute_cql("CREATE FUNCTION row_fct(acc bigint, val int) "
|
|
"RETURNS NULL ON NULL INPUT "
|
|
"RETURNS bigint "
|
|
"LANGUAGE lua "
|
|
"AS $$ "
|
|
"return acc+val "
|
|
"$$;").get();
|
|
e.execute_cql("CREATE FUNCTION final_fct(acc bigint) "
|
|
"RETURNS NULL ON NULL INPUT "
|
|
"RETURNS bigint "
|
|
"LANGUAGE lua "
|
|
"AS $$ "
|
|
"return -acc "
|
|
"$$;").get();
|
|
e.execute_cql("CREATE AGGREGATE aggr(int) "
|
|
"SFUNC row_fct "
|
|
"STYPE bigint "
|
|
"FINALFUNC final_fct "
|
|
"INITCOND 0;").get();
|
|
|
|
e.execute_cql("CREATE TABLE tbl (k int, PRIMARY KEY (k));").get();
|
|
int value_count = 10;
|
|
for (int i = 0; i < value_count; i++) {
|
|
e.execute_cql(format("INSERT INTO tbl (k) VALUES ({:d});", i)).get();
|
|
}
|
|
auto msg = e.execute_cql("SELECT aggr(k) FROM tbl;").get();
|
|
assert_that(msg).is_rows().with_rows({
|
|
{long_type->decompose(-int64_t((value_count - 1) * value_count / 2))}
|
|
});
|
|
|
|
BOOST_CHECK_EQUAL(stat_parallelized, qp.get_cql_stats().select_parallelized);
|
|
});
|
|
}
|
|
|
|
cql3::raw_value make_collection_raw_value(size_t size_to_write, const std::vector<cql3::raw_value>& elements_to_write) {
|
|
size_t serialized_len = 0;
|
|
serialized_len += collection_size_len();
|
|
for (const cql3::raw_value& val : elements_to_write) {
|
|
serialized_len += collection_value_len();
|
|
if (val.is_value()) {
|
|
serialized_len += val.view().with_value([](const FragmentedView auto& view) {
|
|
return view.size_bytes();
|
|
});
|
|
}
|
|
}
|
|
|
|
bytes b(bytes::initialized_later(), serialized_len);
|
|
bytes::iterator out = b.begin();
|
|
|
|
write_collection_size(out, size_to_write);
|
|
for (const cql3::raw_value& val : elements_to_write) {
|
|
if (val.is_null()) {
|
|
write_int32(out, -1);
|
|
} else {
|
|
val.view().with_value([&](const FragmentedView auto& val_view) {
|
|
write_collection_value(out, linearized(val_view));
|
|
});
|
|
}
|
|
}
|
|
|
|
return cql3::raw_value::make_value(b);
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_null_and_unset_in_collections) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE null_in_col (p int primary key, l list<int>, s set<int>, m map<int, int>);").get();
|
|
|
|
// The predicate that checks the message has to be a lambda to preserve source_location
|
|
auto check_null_msg = [](std::source_location loc = std::source_location::current()) {
|
|
return exception_predicate::message_matches(".*(null|NULL).*", loc);
|
|
};
|
|
|
|
auto check_unset_msg = [](std::source_location loc = std::source_location::current()) {
|
|
return exception_predicate::message_contains("unset", loc);
|
|
};
|
|
|
|
// Test null when specified inside a collection literal
|
|
// It's impossible to specify unset value this way
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("INSERT INTO null_in_col (p, l) VALUES (0, [1, null, 3])").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("INSERT INTO null_in_col (p, s) VALUES (0, {1, null, 3})").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("INSERT INTO null_in_col (p, m) VALUES (0, {0:1, null:3, 4:5})").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("INSERT INTO null_in_col (p, m) VALUES (0, {0:1, 2:null, 4:5})").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
|
|
// Test null and unset when sent as bind marker for collection value
|
|
auto insert_list_with_marker = e.prepare("INSERT INTO null_in_col (p, l) VALUES (0, [1, ?, 3])").get();
|
|
auto insert_set_with_marker = e.prepare("INSERT INTO null_in_col (p, s) VALUES (0, {1, ?, 3})").get();
|
|
auto insert_map_with_key_marker = e.prepare("INSERT INTO null_in_col (p, m) VALUES (0, {0:1, ?:3, 4:5})").get();
|
|
auto insert_map_with_value_marker = e.prepare("INSERT INTO null_in_col (p, m) VALUES (0, {0:1, 2:?, 4:5})").get();
|
|
|
|
cql3::raw_value null_value = cql3::raw_value::make_null();
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_list_with_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_set_with_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map_with_key_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map_with_value_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
auto bind_variable_list_with_unset = cql3::raw_value_vector_with_unset({null_value}, {true});
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_list_with_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_set_with_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map_with_key_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map_with_value_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
|
|
|
|
// Test sending whole collections with null and unset inside as bound value
|
|
auto insert_list = e.prepare("INSERT INTO null_in_col (p, l) VALUES (0, ?)").get();
|
|
auto insert_set = e.prepare("INSERT INTO null_in_col (p, s) VALUES (0, ?)").get();
|
|
auto insert_map = e.prepare("INSERT INTO null_in_col (p, m) VALUES (0, ?)").get();
|
|
|
|
auto make_int = [](int val) -> cql3::raw_value {
|
|
return cql3::raw_value::make_value(int32_type->decompose(val));
|
|
};
|
|
|
|
cql3::raw_value list_with_null = make_collection_raw_value(3, {make_int(1), null_value, make_int(2)});
|
|
cql3::raw_value set_with_null = make_collection_raw_value(3, {make_int(1), null_value, make_int(2)});
|
|
|
|
cql3::raw_value map_with_null_key = make_collection_raw_value(3, {make_int(0), make_int(1),
|
|
null_value, make_int(3),
|
|
make_int(4), make_int(5)});
|
|
|
|
cql3::raw_value map_with_null_value = make_collection_raw_value(3, {make_int(0), make_int(1),
|
|
make_int(2), null_value,
|
|
make_int(4), make_int(5)});
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_list, {list_with_null}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_set, {set_with_null}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map, {map_with_null_key}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(insert_map, {map_with_null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
|
|
// Update setting to bad collection value
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET l = [1, null, 2] WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET s = {1, null, 2} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = {0:1, null:3, 4:5} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = {0:1, 2:null, 4:5} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
// Update adding a bad single-element collection value
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET l = l + [null] WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET s = s + {null} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = m + {null:3} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = m + {2:null} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
// Update adding a bad collection value
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET l = l + [1, null, 2] WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET s = s + {1, null, 2} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = m + {0:1, null:3, 4:5} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_cql("UPDATE null_in_col SET m = m + {0:1, 2:null, 4:5} WHERE p = 0").get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
// Update adding a collection value with bad bind marker
|
|
auto add_list_with_marker = e.prepare("UPDATE null_in_col SET l = l + [1, ?, 2] WHERE p = 0").get();
|
|
auto add_set_with_marker = e.prepare("UPDATE null_in_col SET s = s + {1, ?, 2} WHERE p = 0").get();
|
|
auto add_map_with_key_marker = e.prepare("UPDATE null_in_col SET m = m + {0:1, ?:3, 4:5} WHERE p = 0").get();
|
|
auto add_map_with_value_marker = e.prepare("UPDATE null_in_col SET m = m + {0:1, 2:?, 4:5} WHERE p = 0").get();
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_list_with_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_set_with_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map_with_key_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map_with_value_marker, {null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_list_with_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_set_with_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map_with_key_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map_with_value_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
|
|
// Update adding a collection value with bad bind marker
|
|
auto add_list = e.prepare("UPDATE null_in_col SET l = l + ? WHERE p = 0").get();
|
|
auto add_set = e.prepare("UPDATE null_in_col SET s = s + ? WHERE p = 0").get();
|
|
auto add_map = e.prepare("UPDATE null_in_col SET m = m + ? WHERE p = 0").get();
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_list, {list_with_null}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_set, {set_with_null}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map, {map_with_null_key}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(add_map, {map_with_null_value}).get(),
|
|
exceptions::invalid_request_exception, check_null_msg());
|
|
|
|
// List of IN values can contain NULL (which doesn't match anything)
|
|
auto msg1 = e.execute_cql("SELECT * FROM null_in_col WHERE p IN (1, null, 2)").get();
|
|
assert_that(msg1).is_rows().with_rows({});
|
|
|
|
auto where_in_list_with_marker = e.prepare("SELECT * FROM null_in_col WHERE p IN (1, ?, 2)").get();
|
|
|
|
auto msg2 = e.execute_prepared(where_in_list_with_marker, {null_value}).get();
|
|
assert_that(msg2).is_rows().with_rows({});
|
|
|
|
BOOST_REQUIRE_EXCEPTION(e.execute_prepared(where_in_list_with_marker, bind_variable_list_with_unset).get(),
|
|
exceptions::invalid_request_exception, check_unset_msg());
|
|
|
|
auto where_in_list_marker = e.prepare("SELECT * FROM null_in_col WHERE p IN ?").get();
|
|
|
|
auto msg = e.execute_prepared(where_in_list_marker, {list_with_null}).get();
|
|
assert_that(msg).is_rows().with_rows({});
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bind_variable_type_checking) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tab1 (p int primary key, a int, b text, c int)").get();
|
|
|
|
// The predicate that checks the message has to be a lambda to preserve source_location
|
|
auto check_type_conflict = [](std::source_location loc = std::source_location::current()) {
|
|
return exception_predicate::message_contains("variable :var has type", loc);
|
|
};
|
|
|
|
// Test :var needing to have two conflicting types
|
|
BOOST_REQUIRE_EXCEPTION(e.prepare("INSERT INTO tab1 (p, a, b) VALUES (0, :var, :var)").get(),
|
|
exceptions::invalid_request_exception, check_type_conflict());
|
|
BOOST_REQUIRE_EXCEPTION(e.prepare("SELECT * FROM tab1 WHERE a = :var AND b = :var ALLOW FILTERING").get(),
|
|
exceptions::invalid_request_exception, check_type_conflict());
|
|
|
|
// Test :var with a compatible type
|
|
e.prepare("INSERT INTO tab1 (p, a, c) VALUES (0, :var, :var)").get();
|
|
e.prepare("SELECT * FROM tab1 WHERE a = :var AND c = :var ALLOW FILTERING").get();
|
|
});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_bind_variable_type_checking_disabled) {
|
|
auto db_config = make_shared<db::config>();
|
|
db_config->cql_duplicate_bind_variable_names_refer_to_same_variable(false);
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("CREATE TABLE tab1 (p int primary key, a int, b text, c int)").get();
|
|
|
|
// Test :var needing to have two conflicting types; will fail without
|
|
// cql_duplicate_bind_variable_names_refer_to_same_variable = false
|
|
auto prepared = e.prepare("INSERT INTO tab1 (p, a, b) VALUES (0, :var, :var)").get();
|
|
|
|
// Verify that the parameters passed positionally work (non-positional won't make sense)
|
|
auto a = int32_type->decompose(1);
|
|
auto b = utf8_type->decompose("abc");
|
|
e.execute_prepared(prepared, {cql3::raw_value::make_value(a), cql3::raw_value::make_value(b)}).get();
|
|
|
|
auto msg = e.execute_cql("SELECT a, b FROM tab1 WHERE p = 0").get();
|
|
assert_that(msg).is_rows().with_rows({{a, b}});
|
|
}, cql_test_config{db_config});
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_setting_synchronous_updates_property) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create table base (k int, v int, primary key (k));").get();
|
|
|
|
// Check if setting synchronous_updates property works with CREATE
|
|
// MATERIALIZED VIEW and ALTER MATERIALIZED VIEW statements.
|
|
e.execute_cql("create materialized view mv as select * from base "
|
|
"where k is not null and v is not null primary key (v, k)"
|
|
"with synchronous_updates = true").get();
|
|
e.execute_cql("alter materialized view mv with synchronous_updates = true").get();
|
|
e.execute_cql("alter materialized view mv with synchronous_updates = false").get();
|
|
|
|
// Check if index can be altered
|
|
e.execute_cql("create index on base (v)").get();
|
|
e.execute_cql("alter materialized view base_v_idx_index with synchronous_updates = true").get();
|
|
|
|
// Setting synchronous_updates in CREATE TABLE or ALTER TABLE is
|
|
// invalid
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql(
|
|
"create table t (k int, v int, primary key (k)) with synchronous_updates = true"
|
|
).get(),
|
|
exceptions::invalid_request_exception
|
|
);
|
|
BOOST_REQUIRE_THROW(
|
|
e.execute_cql("alter table base with synchronous_updates = true").get(),
|
|
exceptions::invalid_request_exception
|
|
);
|
|
});
|
|
}
|
|
|
|
static
|
|
cql_test_config tablet_cql_test_config() {
|
|
cql_test_config c;
|
|
c.db_config->tablets_mode_for_new_keyspaces.set(db::tablets_mode_t::mode::enabled);
|
|
return c;
|
|
}
|
|
|
|
static
|
|
bool has_tablet_routing(::shared_ptr<cql_transport::messages::result_message> result) {
|
|
auto custom_payload = result->custom_payload();
|
|
if (!custom_payload.has_value() || custom_payload->find("tablets-routing-v1") == custom_payload->end()) {
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_sending_tablet_info_unprepared_insert) {
|
|
BOOST_ASSERT(smp::count == 2);
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks_tablet with replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1 } and tablets = {'initial': 8};").get();
|
|
e.execute_cql("create table ks_tablet.test_tablet (pk int, ck int, v int, PRIMARY KEY (pk, ck));").get();
|
|
|
|
smp::submit_to(0, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_cql("insert into ks_tablet.test_tablet (pk, ck, v) VALUES (1, 2, 3);").get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
|
|
smp::submit_to(1, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_cql("insert into ks_tablet.test_tablet (pk, ck, v) VALUES (1, 2, 3);").get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
}, tablet_cql_test_config());
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_sending_tablet_info_unprepared_select) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks_tablet with replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1 } and tablets = {'initial': 8};").get();
|
|
e.execute_cql("create table ks_tablet.test_tablet (pk int, ck int, v int, PRIMARY KEY (pk, ck));").get();
|
|
e.execute_cql("insert into ks_tablet.test_tablet (pk, ck, v) VALUES (1, 2, 3);").get();
|
|
|
|
smp::submit_to(0, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_cql("select pk, ck, v FROM ks_tablet.test_tablet WHERE pk = 1;").get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
|
|
smp::submit_to(1, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_cql("select pk, ck, v FROM ks_tablet.test_tablet WHERE pk = 1;").get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
}, tablet_cql_test_config());
|
|
}
|
|
|
|
// Reproduces #20768
|
|
SEASTAR_TEST_CASE(test_alter_keyspace_updates_in_memory_objects_with_data_from_system_schema_scylla_keyspaces) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks_tablet with replication = { 'class': 'NetworkTopologyStrategy', "
|
|
"'replication_factor': 1 } and tablets = { 'initial': 1 }").get();
|
|
e.execute_cql("alter keyspace ks_tablet with tablets = { 'initial': 2 }").get();
|
|
e.execute_cql("alter keyspace ks_tablet with tablets = { 'initial': 3 }").get();
|
|
auto& ks = e.local_db().find_keyspace("ks_tablet");
|
|
BOOST_REQUIRE(ks.metadata()->initial_tablets() == 3);
|
|
}, tablet_cql_test_config());
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_sending_tablet_info_insert) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks_tablet with replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1 } and tablets = {'initial': 8};").get();
|
|
e.execute_cql("create table ks_tablet.test_tablet (pk int, ck int, v int, PRIMARY KEY (pk, ck));").get();
|
|
auto insert = e.prepare("insert into ks_tablet.test_tablet (pk, ck, v) VALUES (?, ?, ?);").get();
|
|
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{2})));
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{3})));
|
|
|
|
const auto sptr = e.local_db().find_schema("ks_tablet", "test_tablet");
|
|
|
|
auto pk = partition_key::from_singular(*sptr, int32_t(1));
|
|
|
|
unsigned local_shard = sptr->table().shard_for_reads(dht::get_token(*sptr, pk.view()));
|
|
|
|
smp::submit_to(local_shard, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_prepared(insert, raw_values).get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
|
|
std::vector<cql3::raw_value> raw_values2;
|
|
raw_values2.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{2})));
|
|
raw_values2.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{3})));
|
|
raw_values2.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{4})));
|
|
|
|
auto pk2 = partition_key::from_singular(*sptr, int32_t(2));
|
|
|
|
unsigned local_shard2 = sptr->table().shard_for_reads(dht::get_token(*sptr, pk2.view()));
|
|
unsigned foreign_shard = (local_shard2 + 1) % smp::count;
|
|
|
|
smp::submit_to(foreign_shard, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_prepared(insert, raw_values2).get();
|
|
BOOST_ASSERT(has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
}, tablet_cql_test_config());
|
|
}
|
|
|
|
SEASTAR_TEST_CASE(test_sending_tablet_info_select) {
|
|
return do_with_cql_env_thread([](cql_test_env& e) {
|
|
e.execute_cql("create keyspace ks_tablet with replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1} and tablets = {'initial': 8};").get();
|
|
e.execute_cql("create table ks_tablet.test_tablet (pk int, ck int, v int, PRIMARY KEY (pk, ck));").get();
|
|
e.execute_cql("insert into ks_tablet.test_tablet (pk, ck, v) VALUES (1, 2, 3);").get();
|
|
|
|
auto select = e.prepare("select pk, ck, v FROM ks_tablet.test_tablet WHERE pk = ?;").get();
|
|
std::vector<cql3::raw_value> raw_values;
|
|
raw_values.emplace_back(cql3::raw_value::make_value(int32_type->decompose(int32_t{1})));
|
|
|
|
const auto sptr = e.local_db().find_schema("ks_tablet", "test_tablet");
|
|
|
|
auto pk = partition_key::from_singular(*sptr, int32_t(1));
|
|
|
|
unsigned local_shard = sptr->table().shard_for_reads(dht::get_token(*sptr, pk.view()));
|
|
unsigned foreign_shard = (local_shard + 1) % smp::count;
|
|
|
|
smp::submit_to(local_shard, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_prepared(select, raw_values).get();
|
|
BOOST_ASSERT(!has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
|
|
smp::submit_to(foreign_shard, [&] {
|
|
return seastar::async([&] {
|
|
auto result = e.execute_prepared(select, raw_values).get();
|
|
BOOST_ASSERT(has_tablet_routing(result));
|
|
});
|
|
}).get();
|
|
}, tablet_cql_test_config());
|
|
}
|
|
|
|
// check if create statements emit schema change event properly
|
|
// we emit it even if resource wasn't created due to github.com/scylladb/scylladb/issues/16909
|
|
SEASTAR_TEST_CASE(test_schema_change_events) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
using event_t = cql_transport::messages::result_message::schema_change;
|
|
// keyspace
|
|
auto res = e.execute_cql("create keyspace ks2 with replication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
res = e.execute_cql("create keyspace if not exists ks2 with replication = { 'class' : 'NetworkTopologyStrategy', 'replication_factor' : 1 };").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
|
|
// table
|
|
res = e.execute_cql("create table users (user_name varchar PRIMARY KEY);").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
res = e.execute_cql("create table if not exists users (user_name varchar PRIMARY KEY);").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
|
|
// view
|
|
res = e.execute_cql("create materialized view users_view as select user_name from users where user_name is not null primary key (user_name)").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
res = e.execute_cql("create materialized view if not exists users_view as select user_name from users where user_name is not null primary key (user_name)").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
|
|
// type
|
|
res = e.execute_cql("create type my_type (first text);").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
res = e.execute_cql("create type if not exists my_type (first text);").get();
|
|
BOOST_REQUIRE(dynamic_pointer_cast<event_t>(res));
|
|
});
|
|
}
|
|
|
|
// check that we can load sstable with mixed numerical and uuid generation types
|
|
SEASTAR_TEST_CASE(test_sstable_load_mixed_generation_type) {
|
|
return do_with_cql_env_thread([] (cql_test_env& e) {
|
|
// Create table
|
|
e.execute_cql("create table ks.test (k int PRIMARY KEY, v int);").get();
|
|
|
|
auto& tbl = e.local_db().find_column_family("ks", "test");
|
|
auto upload_dir = table_dir(tbl) / sstables::upload_dir;
|
|
|
|
// Load sstables with mixed generation types
|
|
copy_directory("test/resource/sstables/mixed_generation_type", upload_dir);
|
|
replica::distributed_loader::process_upload_dir(e.db(), e.view_builder(), e.view_building_worker(), "ks", "test", false, false).get();
|
|
|
|
// Verify the expected data is present
|
|
assert_that(e.execute_cql("SELECT * FROM ks.test").get()).is_rows()
|
|
.with_size(3)
|
|
.with_rows_ignore_order({
|
|
{int32_type->decompose(0), int32_type->decompose(0)},
|
|
{int32_type->decompose(1), int32_type->decompose(1)},
|
|
{int32_type->decompose(2), int32_type->decompose(2)}
|
|
});
|
|
|
|
// Run major compaction to ensure that the mixed generation types are handled correctly
|
|
auto& compaction_module = e.local_db().get_compaction_manager().get_task_manager_module();
|
|
std::vector<table_info> table_infos({{"test", tbl.schema()->id()}});
|
|
auto task = compaction_module.make_and_start_task<compaction::major_keyspace_compaction_task_impl>({}, "ks", tasks::task_id::create_null_id(), e.db(), table_infos, compaction::flush_mode::skip, false).get();
|
|
task->done().get();
|
|
|
|
// Verify the expected data again
|
|
assert_that(e.execute_cql("SELECT * FROM ks.test").get()).is_rows()
|
|
.with_size(3)
|
|
.with_rows_ignore_order({
|
|
{int32_type->decompose(0), int32_type->decompose(0)},
|
|
{int32_type->decompose(1), int32_type->decompose(1)},
|
|
{int32_type->decompose(2), int32_type->decompose(2)}
|
|
});
|
|
});
|
|
}
|
|
|
|
BOOST_AUTO_TEST_SUITE_END()
|