Files
scylladb/redis/query_utils.cc
Etienne Adam 63a1a4cbb9 redis: add hgetall and hdel commands
This patch adds support for 2 hash commands HDEL and HGETALL.

Internally it introduces the hashes_result_builder class to
read hashes and stored them in a std::map.

Other changes:
  - one exception return string was fixed
  - tests now use pytest.raises

Signed-off-by: Etienne Adam <etienne.adam@gmail.com>
Message-Id: <20200907202528.4985-1-etienne.adam@gmail.com>
2020-09-08 11:59:52 +03:00

178 lines
8.3 KiB
C++

/*
* Copyright (C) 2019 pengjian.uestc @ gmail.com
*/
/*
* This file is part of Scylla.
*
* Scylla is free software: you can redistribute it and/or modify
* it under the terms of the GNU Affero General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* Scylla is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with Scylla. If not, see <http://www.gnu.org/licenses/>.
*/
#include "redis/query_utils.hh"
#include "redis/options.hh"
#include "timeout_config.hh"
#include "service/client_state.hh"
#include "service/storage_proxy.hh"
#include "dht/i_partitioner.hh"
#include "partition_slice_builder.hh"
#include "query-result-reader.hh"
#include "gc_clock.hh"
#include "service_permit.hh"
#include "redis/keyspace_utils.hh"
namespace redis {
class strings_result_builder {
lw_shared_ptr<strings_result> _data;
const query::partition_slice& _partition_slice;
const schema_ptr _schema;
private:
void add_cell(const column_definition& col, const std::optional<query::result_atomic_cell_view>& cell)
{
if (cell) {
cell->value().with_linearized([this, &col, &cell] (bytes_view cell_view) {
auto&& dv = col.type->deserialize_value(cell_view);
auto&& d = dv.serialize_nonnull();
_data->_result = std::move(d);
if (cell->expiry().has_value()) {
_data->_ttl = cell->expiry().value() - gc_clock::now();
}
_data->_has_result = true;
});
}
}
public:
strings_result_builder(lw_shared_ptr<strings_result> data, const schema_ptr schema, const query::partition_slice& ps)
: _data(data)
, _partition_slice(ps)
, _schema(schema)
{
}
void accept_new_partition(const partition_key& key, uint32_t row_count) {}
void accept_new_partition(uint32_t row_count) {}
void accept_new_row(const clustering_key& key, const query::result_row_view& static_row, const query::result_row_view& row)
{
auto row_iterator = row.iterator();
for (auto&& id : _partition_slice.regular_columns) {
add_cell(_schema->regular_column_at(id), row_iterator.next_atomic_cell());
}
}
void accept_new_row(const query::result_row_view& static_row, const query::result_row_view& row) {}
void accept_partition_end(const query::result_row_view& static_row) {}
};
future<lw_shared_ptr<strings_result>> read_strings(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit) {
auto schema = get_schema(proxy, options.get_keyspace_name(), redis::STRINGs);
auto ps = partition_slice_builder(*schema).build();
return query_strings(proxy, options, key, permit, schema, ps);
}
future<lw_shared_ptr<strings_result>> query_strings(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit, schema_ptr schema, query::partition_slice ps) {
const auto max_result_size = proxy.get_max_result_size(ps);
query::read_command cmd(schema->id(), schema->version(), ps, 1, gc_clock::now(), std::nullopt, 1, utils::UUID(), query::is_first_page::no, max_result_size, 0);
auto pkey = partition_key::from_single_value(*schema, key);
auto partition_range = dht::partition_range::make_singular(dht::decorate_key(*schema, std::move(pkey)));
dht::partition_range_vector partition_ranges;
partition_ranges.emplace_back(std::move(partition_range));
auto read_consistency_level = options.get_read_consistency_level();
db::timeout_clock::time_point timeout = db::timeout_clock::now() + options.get_read_timeout();
return proxy.query(schema, make_lw_shared<query::read_command>(std::move(cmd)), std::move(partition_ranges), read_consistency_level, {timeout, permit, service::client_state::for_internal_calls()}).then([ps, schema] (auto qr) {
return query::result_view::do_with(*qr.query_result, [&] (query::result_view v) {
auto pd = make_lw_shared<strings_result>();
v.consume(ps, strings_result_builder(pd, schema, ps));
return pd;
});
});
}
class hashes_result_builder {
lw_shared_ptr<std::map<bytes, bytes>> _data;
const query::partition_slice& _partition_slice;
const schema_ptr _schema;
private:
void add_cell(const bytes& ckey, const column_definition& col, const std::optional<query::result_atomic_cell_view>& cell)
{
if (cell) {
cell->value().with_linearized([this, &col, &cell, &ckey] (bytes_view cell_view) {
auto&& dv = col.type->deserialize_value(cell_view);
auto&& d = dv.serialize_nonnull();
_data->emplace(std::move(ckey), std::move(d));
});
}
}
public:
hashes_result_builder(lw_shared_ptr<std::map<bytes, bytes>> data, const schema_ptr schema, const query::partition_slice& ps)
: _data(data)
, _partition_slice(ps)
, _schema(schema)
{
}
void accept_new_partition(const partition_key& key, uint32_t row_count) {}
void accept_new_partition(uint32_t row_count) {}
void accept_new_row(const clustering_key& key, const query::result_row_view& static_row, const query::result_row_view& row)
{
auto row_iterator = row.iterator();
for (auto&& v : key.explode()) {
for (auto&& id : _partition_slice.regular_columns) {
add_cell(std::move(v), _schema->regular_column_at(id), row_iterator.next_atomic_cell());
}
}
}
void accept_new_row(const query::result_row_view& static_row, const query::result_row_view& row) {}
void accept_partition_end(const query::result_row_view& static_row) {}
};
future<lw_shared_ptr<std::map<bytes, bytes>>> read_hashes(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit) {
auto schema = get_schema(proxy, options.get_keyspace_name(), redis::HASHes);
auto ps = partition_slice_builder(*schema)
.build();
return query_hashes(proxy, options, key, permit, schema, ps);
}
future<lw_shared_ptr<std::map<bytes, bytes>>> read_hashes(service::storage_proxy& proxy, const redis_options& options, const bytes& key, const bytes& field, service_permit permit) {
auto schema = get_schema(proxy, options.get_keyspace_name(), redis::HASHes);
auto ckey = clustering_key::from_single_value(*schema, field);
auto clustering_range = query::clustering_range::make_singular(ckey);
auto ps = partition_slice_builder(*schema)
.with_range(std::move(clustering_range))
.build();
return query_hashes(proxy, options, key, permit, schema, ps);
}
future<lw_shared_ptr<std::map<bytes, bytes>>> query_hashes(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit, schema_ptr schema, query::partition_slice ps) {
const auto max_result_size = proxy.get_max_result_size(ps);
query::read_command cmd(schema->id(), schema->version(), ps, std::numeric_limits<uint32_t>::max(), gc_clock::now(), std::nullopt, 1, utils::UUID(), query::is_first_page::no, max_result_size, 0);
auto pkey = partition_key::from_single_value(*schema, key);
auto partition_range = dht::partition_range::make_singular(dht::decorate_key(*schema, std::move(pkey)));
dht::partition_range_vector partition_ranges;
partition_ranges.emplace_back(std::move(partition_range));
auto read_consistency_level = options.get_read_consistency_level();
db::timeout_clock::time_point timeout = db::timeout_clock::now() + options.get_read_timeout();
return proxy.query(schema, make_lw_shared<query::read_command>(std::move(cmd)), std::move(partition_ranges), read_consistency_level, {timeout, permit, service::client_state::for_internal_calls()}).then([ps, schema] (auto qr) {
return query::result_view::do_with(*qr.query_result, [&] (query::result_view v) {
auto pd = make_lw_shared<std::map<bytes, bytes>>();
v.consume(ps, hashes_result_builder(pd, schema, ps));
return pd;
});
});
}
}