Files
scylladb/types/tuple.hh
Avi Kivity f3eade2f62 treewide: relicense to ScyllaDB-Source-Available-1.0
Drop the AGPL license in favor of a source-available license.
See the blog post [1] for details.

[1] https://www.scylladb.com/2024/12/18/why-were-moving-to-a-source-available-license/
2024-12-18 17:45:13 +02:00

208 lines
6.6 KiB
C++

/*
* Copyright (C) 2014-present ScyllaDB
*/
/*
* SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
*/
#pragma once
#include <iterator>
#include <vector>
#include <string>
#include <ranges>
#include "types/types.hh"
struct tuple_deserializing_iterator {
public:
using iterator_category = std::input_iterator_tag;
using value_type = const managed_bytes_view_opt;
using difference_type = std::ptrdiff_t;
using pointer = const managed_bytes_view_opt*;
using reference = const managed_bytes_view_opt&;
private:
managed_bytes_view _v;
managed_bytes_view_opt _current;
public:
struct end_tag {};
tuple_deserializing_iterator() = default;
tuple_deserializing_iterator(managed_bytes_view v) : _v(v) {
parse();
}
tuple_deserializing_iterator(end_tag, managed_bytes_view v) : _v(v) {
_v.remove_prefix(_v.size());
}
static tuple_deserializing_iterator start(managed_bytes_view v) {
return tuple_deserializing_iterator(v);
}
static tuple_deserializing_iterator finish(managed_bytes_view v) {
return tuple_deserializing_iterator(end_tag(), v);
}
const managed_bytes_view_opt& operator*() const {
return _current;
}
const managed_bytes_view_opt* operator->() const {
return &_current;
}
tuple_deserializing_iterator& operator++() {
skip();
parse();
return *this;
}
void operator++(int) {
skip();
parse();
}
bool operator==(const tuple_deserializing_iterator& x) const {
return _v == x._v;
}
private:
void parse() {
_current = std::nullopt;
if (_v.empty()) {
return;
}
// we don't consume _v, otherwise operator==
// or the copy constructor immediately after
// parse() yields the wrong results.
auto tmp = _v;
auto s = read_simple<int32_t>(tmp);
if (s < 0) {
return;
}
_current = read_simple_bytes(tmp, s);
}
void skip() {
_v.remove_prefix(4 + (_current ? _current->size() : 0));
}
};
template <FragmentedView View>
std::optional<View> read_tuple_element(View& v) {
auto s = read_simple<int32_t>(v);
if (s < 0) {
return std::nullopt;
}
return read_simple_bytes(v, s);
}
template <FragmentedView View>
managed_bytes_opt get_nth_tuple_element(View v, size_t n) {
for (size_t i = 0; i < n; ++i) {
if (v.empty()) {
return std::nullopt;
}
read_tuple_element(v);
}
if (v.empty()) {
return std::nullopt;
}
auto el = read_tuple_element(v);
if (el) {
return managed_bytes(*el);
}
return std::nullopt;
}
class tuple_type_impl : public concrete_type<std::vector<data_value>> {
using intern = type_interning_helper<tuple_type_impl, std::vector<data_type>>;
protected:
std::vector<data_type> _types;
static std::ranges::subrange<tuple_deserializing_iterator> make_range(managed_bytes_view v) {
return { tuple_deserializing_iterator::start(v), tuple_deserializing_iterator::finish(v) };
}
tuple_type_impl(kind k, sstring name, std::vector<data_type> types, bool freeze_inner);
tuple_type_impl(std::vector<data_type> types, bool freze_inner);
public:
tuple_type_impl(std::vector<data_type> types);
static shared_ptr<const tuple_type_impl> get_instance(std::vector<data_type> types);
data_type type(size_t i) const {
return _types[i];
}
size_t size() const {
return _types.size();
}
const std::vector<data_type>& all_types() const {
return _types;
}
std::vector<bytes_opt> split(FragmentedView auto v) const {
std::vector<bytes_opt> elements;
while (!v.empty()) {
auto fragmented_element_optional = read_tuple_element(v);
if (fragmented_element_optional) {
elements.push_back(linearized(*fragmented_element_optional));
} else {
elements.push_back(std::nullopt);
}
}
return elements;
}
std::vector<managed_bytes_opt> split_fragmented(FragmentedView auto v) const {
std::vector<managed_bytes_opt> elements;
while (!v.empty()) {
auto fragmented_element_optional = read_tuple_element(v);
if (fragmented_element_optional) {
elements.push_back(managed_bytes(*fragmented_element_optional));
} else {
elements.push_back(std::nullopt);
}
}
return elements;
}
template <typename RangeOf_bytes_opt> // also accepts bytes_view_opt
static bytes build_value(RangeOf_bytes_opt&& range) {
auto item_size = [] (auto&& v) { return 4 + (v ? v->size() : 0); };
auto size = std::ranges::fold_left(range | std::views::transform(item_size), 0, std::plus());
auto ret = bytes(bytes::initialized_later(), size);
auto out = ret.begin();
auto put = [&out] (auto&& v) {
if (v) {
using val_type = std::remove_cvref_t<decltype(*v)>;
if constexpr (FragmentedView<val_type>) {
int32_t size = v->size_bytes();
write(out, size);
read_fragmented(*v, size, out);
out += size;
} else {
write(out, int32_t(v->size()));
out = std::copy(v->begin(), v->end(), out);
}
} else {
write(out, int32_t(-1));
}
};
std::ranges::for_each(range, put);
return ret;
}
template <typename Range> // range of managed_bytes_opt or managed_bytes_view_opt
requires requires (Range it) { {std::begin(it)->value()} -> std::convertible_to<managed_bytes_view>; }
static managed_bytes build_value_fragmented(Range&& range) {
size_t size = 0;
for (auto&& v : range) {
size += 4 + (v ? v->size() : 0);
}
auto ret = managed_bytes(managed_bytes::initialized_later(), size);
auto out = managed_bytes_mutable_view(ret);
for (auto&& v : range) {
if (v) {
write<int32_t>(out, v->size());
write_fragmented(out, managed_bytes_view(*v));
} else {
write<int32_t>(out, -1);
}
}
return ret;
}
private:
void set_contains_collections();
static sstring make_name(const std::vector<data_type>& types);
friend abstract_type;
};
data_value make_tuple_value(data_type tuple_type, tuple_type_impl::native_type value);