Files
scylladb/redis/query_utils.cc
Etienne Adam 19683d04c6 redis: add hget and hset commands
hget and hset commands using hashes internally, thus
they are not using the existing write_strings() function.

Limitations:
 - hset only supports 3 params, instead of multiple field/value
list that is available in official redis-server.
 - hset should return 0 when the key and field already exists,
but I am not sure it's possible to retrieve this information
without doing read-before-write, which would not be atomic.

I factorized a bit the query_* functions to reduce duplication, but
I am not 100% sure of the naming, it may still be a bit confusing
between the schema used (strings, hashes) and the returned format
(currently only string but array should come later with hgetall).

Signed-off-by: Etienne Adam <etienne.adam@gmail.com>
Message-Id: <20200830190128.18534-1-etienne.adam@gmail.com>
2020-08-30 22:05:41 +03:00

112 lines
5.0 KiB
C++

/*
* Copyright (C) 2019 pengjian.uestc @ gmail.com
*/
/*
* This file is part of Scylla.
*
* Scylla is free software: you can redistribute it and/or modify
* it under the terms of the GNU Affero General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* Scylla is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with Scylla. If not, see <http://www.gnu.org/licenses/>.
*/
#include "redis/query_utils.hh"
#include "redis/options.hh"
#include "timeout_config.hh"
#include "service/client_state.hh"
#include "service/storage_proxy.hh"
#include "dht/i_partitioner.hh"
#include "partition_slice_builder.hh"
#include "query-result-reader.hh"
#include "gc_clock.hh"
#include "service_permit.hh"
#include "redis/keyspace_utils.hh"
namespace redis {
class strings_result_builder {
lw_shared_ptr<strings_result> _data;
const query::partition_slice& _partition_slice;
const schema_ptr _schema;
private:
void add_cell(const column_definition& col, const std::optional<query::result_atomic_cell_view>& cell)
{
if (cell) {
cell->value().with_linearized([this, &col, &cell] (bytes_view cell_view) {
auto&& dv = col.type->deserialize_value(cell_view);
auto&& d = dv.serialize_nonnull();
_data->_result = std::move(d);
if (cell->expiry().has_value()) {
_data->_ttl = cell->expiry().value() - gc_clock::now();
}
_data->_has_result = true;
});
}
}
public:
strings_result_builder(lw_shared_ptr<strings_result> data, const schema_ptr schema, const query::partition_slice& ps)
: _data(data)
, _partition_slice(ps)
, _schema(schema)
{
}
void accept_new_partition(const partition_key& key, uint32_t row_count) {}
void accept_new_partition(uint32_t row_count) {}
void accept_new_row(const clustering_key& key, const query::result_row_view& static_row, const query::result_row_view& row)
{
auto row_iterator = row.iterator();
for (auto&& id : _partition_slice.regular_columns) {
add_cell(_schema->regular_column_at(id), row_iterator.next_atomic_cell());
}
}
void accept_new_row(const query::result_row_view& static_row, const query::result_row_view& row) {}
void accept_partition_end(const query::result_row_view& static_row) {}
};
future<lw_shared_ptr<strings_result>> read_strings(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit) {
auto schema = get_schema(proxy, options.get_keyspace_name(), redis::STRINGs);
auto ps = partition_slice_builder(*schema).build();
return query_strings(proxy, options, key, permit, schema, ps);
}
future<lw_shared_ptr<strings_result>> read_strings_from_hash(service::storage_proxy& proxy, const redis_options& options, const bytes& key, const bytes& field, service_permit permit) {
auto schema = get_schema(proxy, options.get_keyspace_name(), redis::HASHes);
auto ckey = clustering_key::from_single_value(*schema, field);
auto clustering_range = query::clustering_range::make_singular(ckey);
auto ps = partition_slice_builder(*schema)
.with_range(std::move(clustering_range))
.build();
return query_strings(proxy, options, key, permit, schema, ps);
}
future<lw_shared_ptr<strings_result>> query_strings(service::storage_proxy& proxy, const redis_options& options, const bytes& key, service_permit permit, schema_ptr schema, query::partition_slice ps) {
const auto max_result_size = proxy.get_max_result_size(ps);
query::read_command cmd(schema->id(), schema->version(), ps, 1, gc_clock::now(), std::nullopt, 1, utils::UUID(), query::is_first_page::no, max_result_size, 0);
auto pkey = partition_key::from_single_value(*schema, key);
auto partition_range = dht::partition_range::make_singular(dht::decorate_key(*schema, std::move(pkey)));
dht::partition_range_vector partition_ranges;
partition_ranges.emplace_back(std::move(partition_range));
auto read_consistency_level = options.get_read_consistency_level();
db::timeout_clock::time_point timeout = db::timeout_clock::now() + options.get_read_timeout();
return proxy.query(schema, make_lw_shared<query::read_command>(std::move(cmd)), std::move(partition_ranges), read_consistency_level, {timeout, permit, service::client_state::for_internal_calls()}).then([ps, schema] (auto qr) {
return query::result_view::do_with(*qr.query_result, [&] (query::result_view v) {
auto pd = make_lw_shared<strings_result>();
v.consume(ps, strings_result_builder(pd, schema, ps));
return pd;
});
});
}
}