mirror of
https://github.com/scylladb/scylladb.git
synced 2026-04-27 03:45:11 +00:00
It turns out that `cql_table_large_data_handler::record_large_rows` and `cql_table_large_data_handler::record_large_cells` were broken for reporting static cells and static rows from the very beginning: In case a large static cell or a large static row is encountered, it tries to execute `db::try_record` with `nullptr` additional values, denoting that there is no clustering key to be recorded. These values are next passed to `qctx.execute_cql()`, which creates `data_value` instances for each statement parameter, hence invoking `data_value(nullptr)`. This uses `const char*` overload which delegates to `std::string_view` ctor overload. It is UB to pass `nullptr` pointer to `std::string_view` ctor. Hence leading to segmentation faults in the aforementioned large data reporting code. What we want here is to make a null `data_value` instead, so just add an overload specifically for `std::nullptr_t`, which will create a null `data_value` with `text` type. A regression test is provided for the issue (written in `cql-pytest` framework). Tests: test/cql-pytest/test_large_cells_rows.py Fixes: #6780 Signed-off-by: Pavel Solodovnikov <pa.solodovnikov@scylladb.com> Message-Id: <20201223204552.61081-1-pa.solodovnikov@scylladb.com>
43 lines
2.2 KiB
Python
43 lines
2.2 KiB
Python
# Copyright 2020 ScyllaDB
|
|
#
|
|
# This file is part of Scylla.
|
|
#
|
|
# Scylla is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU Affero General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# Scylla is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU Affero General Public License
|
|
# along with Scylla. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
from util import new_test_table
|
|
|
|
import requests
|
|
|
|
def test_create_large_static_cells_and_rows(cql, test_keyspace):
|
|
'''Test that `large_data_handler` successfully reports large static cells
|
|
and static rows and this doesn't cause a crash of Scylla server.
|
|
|
|
This is a regression test for https://github.com/scylladb/scylla/issues/6780'''
|
|
schema = "pk int, ck int, user_ids set<text> static, PRIMARY KEY (pk, ck)"
|
|
with new_test_table(cql, test_keyspace, schema) as table:
|
|
insert_stmt = cql.prepare(f"INSERT INTO {table} (pk, ck, user_ids) VALUES (?, ?, ?) USING TIMEOUT 5m")
|
|
# Default large data threshold for cells is 1 mb, for rows it is 10 mb.
|
|
# Take 10 mb cell to trigger large data reporting code both for
|
|
# static cells and static rows simultaneously.
|
|
large_set = {'x' * 1024 * 1024 * 10}
|
|
cql.execute(insert_stmt, [1, 1, large_set])
|
|
|
|
# REST API endpoint address for test scylla node
|
|
node_address = f'http://{cql.cluster.contact_points[0]}:10000'
|
|
# Execute force flush of test table to persistent storage, which is necessary to trigger
|
|
# `large_data_handler` execution.
|
|
table_without_ks = table[table.find('.') + 1:] # strip keyspace part from the table name
|
|
requests.post(f'{node_address}/storage_service/keyspace_flush/{test_keyspace}', params={'cf' : table_without_ks})
|
|
# No need to check that the Scylla server is running here, since the test will
|
|
# fail automatically in case Scylla crashes. |