Files
scylladb/test/cluster/test_tablets_cql.py
Dawid Mędrek dbb8835fdf test/cluster: Adjust simple tests to RF-rack-validity
We adjust all of the simple cases of cluster tests so they work
with `rf_rack_valid_keyspaces: true`. It boils down to assigning
nodes to multiple racks. For most of the changes, we do that by:

* Using `pytest.mark.prepare_3_racks_cluster` instead of
  `pytest.mark.prepare_3_nodes_cluster`.
* Using an additional argument -- `auto_rack_dc` -- when calling
  `ManagerClient::servers_add()`.

In some cases, we need to assign the racks manually, which may be
less obvious, but in every such situation, the tests didn't rely
on that assignment, so that doesn't affect them or what they verify.
2025-05-10 16:30:18 +02:00

134 lines
7.0 KiB
Python

#
# Copyright (C) 2024-present ScyllaDB
#
# SPDX-License-Identifier: LicenseRef-ScyllaDB-Source-Available-1.0
#
import asyncio
import logging
import pytest
from cassandra.protocol import InvalidRequest
from test.pylib.manager_client import ManagerClient
from test.pylib.rest_client import inject_error_one_shot
from test.cluster.conftest import skip_mode
from test.cluster.util import disable_schema_agreement_wait, create_new_test_keyspace, new_test_keyspace
logger = logging.getLogger(__name__)
@pytest.mark.asyncio
@skip_mode('release', 'error injections are not supported in release mode')
async def test_alter_dropped_tablets_keyspace(manager: ManagerClient) -> None:
config = {
'tablets_mode_for_new_keyspaces': 'enabled'
}
logger.info("starting a node (the leader)")
servers = [await manager.server_add(config=config)]
logger.info("starting a second node (the follower)")
servers += [await manager.server_add(config=config)]
ks = await create_new_test_keyspace(manager.get_cql(), "with "
"replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1} and "
"tablets = {'enabled': true}")
await manager.get_cql().run_async(f"create table {ks}.t (pk int primary key)")
logger.info(f"injecting wait-after-topology-coordinator-gets-event into the leader node {servers[0]}")
injection_handler = await inject_error_one_shot(manager.api, servers[0].ip_addr,
'wait-after-topology-coordinator-gets-event')
async def alter_tablets_ks_without_waiting_to_complete():
res = await manager.get_cql().run_async("select data_center from system.local")
# ALTER tablets KS only accepts a specific DC, it rejects the generic 'replication_factor' tag
this_dc = res[0].data_center
await manager.get_cql().run_async(f"alter keyspace {ks} "
f"with replication = {{'class': 'NetworkTopologyStrategy', '{this_dc}': 1}}")
# by creating a task this way we ensure it's immediately executed, but we won't wait until it's completed
task = asyncio.create_task(alter_tablets_ks_without_waiting_to_complete())
logger.info(f"waiting for the leader node {servers[0]} to start handling the keyspace-rf-change request")
leader_log_file = await manager.server_open_log(servers[0].server_id)
await leader_log_file.wait_for("wait-after-topology-coordinator-gets-event: waiting", timeout=10)
logger.info(f"dropping KS from the follower node {servers[1]} so that the leader, which hangs on injected sleep, "
f"wakes up with the drop applied")
host = manager.get_cql().cluster.metadata.get_host(servers[1].ip_addr)
await manager.get_cql().run_async(f"drop keyspace {ks}", host=host)
logger.info("Waking up the leader to continue processing ALTER with KS that doesn't exist (has been just dropped)")
await injection_handler.message()
matches = await leader_log_file.grep("topology change coordinator fiber got error "
f"data_dictionary::no_such_keyspace \(Can't find a keyspace {ks}\)")
assert not matches
with pytest.raises(InvalidRequest, match=f"Can't ALTER keyspace {ks}, keyspace doesn't exist|Can't find a keyspace {ks}") as e:
await task
@pytest.mark.asyncio
@skip_mode('release', 'error injections are not supported in release mode')
async def test_alter_tablets_keyspace_concurrent_modification(manager: ManagerClient) -> None:
config = {
'tablets_mode_for_new_keyspaces': 'enabled'
}
logger.info("starting a node (the leader)")
servers = [await manager.server_add(config=config, property_file={"dc": "dc1", "rack": "r1"})]
logger.info("starting a second node (the follower)")
servers += [await manager.server_add(config=config, property_file={"dc": "dc1", "rack": "r2"})]
async with new_test_keyspace(manager, "with "
"replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1} and "
"tablets = {'initial': 2}") as ks:
await manager.get_cql().run_async(f"create table {ks}.t (pk int primary key)")
logger.info(f"injecting wait-before-committing-rf-change-event into the leader node {servers[0]}")
injection_handler = await inject_error_one_shot(manager.api, servers[0].ip_addr,
'wait-before-committing-rf-change-event')
# ALTER tablets KS only accepts a specific DC, it rejects the generic 'replication_factor' tag
res = await manager.get_cql().run_async("select data_center from system.local")
this_dc = res[0].data_center
async def alter_tablets_ks_without_waiting_to_complete():
logger.info("scheduling ALTER KS to change the RF from 1 to 2")
await manager.get_cql().run_async(f"alter keyspace {ks} "
f"with replication = {{'class': 'NetworkTopologyStrategy', '{this_dc}': 2}}")
# by creating a task this way we ensure it's immediately executed,
# but we don't want to wait until the task is completed here,
# because we want to do something else in the meantime
task = asyncio.create_task(alter_tablets_ks_without_waiting_to_complete())
logger.info(f"waiting for the leader node {servers[0]} to start handling the keyspace-rf-change request")
leader_log_file = await manager.server_open_log(servers[0].server_id)
await leader_log_file.wait_for("wait-before-committing-rf-change-event: waiting", timeout=10)
logger.info(f"creating another keyspace from the follower node {servers[1]} so that the leader, which hangs on injected sleep, "
f"wakes up with a changed schema")
host = manager.get_cql().cluster.metadata.get_host(servers[1].ip_addr)
with disable_schema_agreement_wait(manager.get_cql()):
ks2 = await create_new_test_keyspace(manager.get_cql(), "with "
"replication = {'class': 'NetworkTopologyStrategy', 'replication_factor': 1} "
"and tablets = {'enabled': true}", host=host)
logger.info("waking up the leader to continue processing ALTER on a changed schema, which should cause a retry")
await injection_handler.message()
logger.info("waiting for ALTER to complete")
await task
# ensure that the concurrent modification error really did take place
matches = await leader_log_file.grep("topology change coordinator fiber got group0_concurrent_modification")
assert matches
# ensure that the ALTER has eventually succeeded and we changed RF from 1 to 2
res = manager.get_cql().execute(f"SELECT * FROM system_schema.keyspaces WHERE keyspace_name = '{ks}'")
assert res[0].replication[this_dc] == '2'
await manager.get_cql().run_async(f"drop keyspace {ks2}")