Files
scylladb/docs/design-notes/system_keyspace.md
dgarcia360 fd5f0c3034 docs: add organization
Closes #7818
2020-12-22 15:33:31 +02:00

149 lines
4.5 KiB
Markdown

# System keyspace layout
This section describes layouts and usage of system.* tables.
## The system.large\_* tables
Scylla performs better if partitions, rows, or cells are not too
large. To help diagnose cases where these grow too large, scylla keeps
3 tables that record large partitions, rows, and cells, respectively.
The meaning of an entry in each of these tables is similar. It means
that there is a particular sstable with a large partition, row, or
cell. In particular, this implies that:
* There is no entry until compaction aggregates enough data in a
single sstable.
* The entry stays around until the sstable is deleted.
In addition, the entries also have a TTL of 30 days.
## system.large\_partitions
Large partition table can be used to trace largest partitions in a cluster.
Schema:
~~~
CREATE TABLE system.large_partitions (
keyspace_name text,
table_name text,
sstable_name text,
partition_size bigint,
partition_key text,
compaction_time timestamp,
PRIMARY KEY ((keyspace_name, table_name), sstable_name, partition_size, partition_key)
) WITH CLUSTERING ORDER BY (sstable_name ASC, partition_size DESC, partition_key ASC);
~~~
### Example usage
#### Extracting large partitions info
~~~
SELECT * FROM system.large_partitions;
~~~
#### Extracting large partitions info for a single table
~~~
SELECT * FROM system.large_partitions WHERE keyspace_name = 'ks1' and table_name = 'standard1';
~~~
## system.large\_rows
Large row table can be used to trace large clustering and static rows in a cluster.
This table is currently only used with the MC format (issue #4868).
Schema:
~~~
CREATE TABLE system.large_rows (
keyspace_name text,
table_name text,
sstable_name text,
row_size bigint,
partition_key text,
clustering_key text,
compaction_time timestamp,
PRIMARY KEY ((keyspace_name, table_name), sstable_name, row_size, partition_key, clustering_key)
) WITH CLUSTERING ORDER BY (sstable_name ASC, row_size DESC, partition_key ASC, clustering_key ASC);
~~~
### Example usage
#### Extracting large row info
~~~
SELECT * FROM system.large_rows;
~~~
#### Extracting large rows info for a single table
~~~
SELECT * FROM system.large_rows WHERE keyspace_name = 'ks1' and table_name = 'standard1';
~~~
## system.large\_cells
Large cell table can be used to trace large cells in a cluster.
This table is currently only used with the MC format (issue #4868).
Schema:
~~~
CREATE TABLE system.large_cells (
keyspace_name text,
table_name text,
sstable_name text,
cell_size bigint,
partition_key text,
clustering_key text,
column_name text,
compaction_time timestamp,
PRIMARY KEY ((keyspace_name, table_name), sstable_name, cell_size, partition_key, clustering_key, column_name)
) WITH CLUSTERING ORDER BY (sstable_name ASC, cell_size DESC, partition_key ASC, clustering_key ASC, column_name ASC)
~~~
Note that a collection is just one cell. There is no information about
the size of each collection element.
### Example usage
#### Extracting large cells info
~~~
SELECT * FROM system.large_cells;
~~~
#### Extracting large cells info for a single table
~~~
SELECT * FROM system.large_cells WHERE keyspace_name = 'ks1' and table_name = 'standard1';
~~~
## system.truncated
Holds truncation replay positions per table and shard
Schema:
~~~
CREATE TABLE system.truncated (
table_uuid uuid, # id of truncated table
shard int, # shard
position int, # replay position
segment_id bigint, # replay segment
truncated_at timestamp static, # truncation time
PRIMARY KEY (table_uuid, shard)
) WITH CLUSTERING ORDER BY (shard ASC)
~~~
When a table is truncated, sstables are removed and the current replay position for each
shard (last mutation to be committed to either sstable or memtable) is collected.
These are then inserted into the above table, using shard as clustering.
When doing commitlog replay (in case of a crash), the data is read from the above
table and mutations are filtered based on the replay positions to ensure
truncated data is not resurrected.
Note that until the above table was added, truncation records where kept in the
`truncated_at` map column in the `system.local` table. When booting up, scylla will
merge the data in the legacy store with data the `truncated` table. Until the whole
cluster agrees on the feature `TRUNCATION_TABLE` truncation will write both new and
legacy records. When the feature is agreed upon the legacy map is removed.
## TODO: the rest