Skip to content

Commit 72e47bf

Browse files
committed
Capitalized Block and Chunk words when nouns
Signed-off-by: Marco Pracucci <[email protected]>
1 parent d213a61 commit 72e47bf

File tree

2 files changed

+17
-17
lines changed

2 files changed

+17
-17
lines changed

docs/architecture.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -28,15 +28,15 @@ The two engines mostly share the same Cortex architecture with few differences o
2828

2929
### Chunks storage (default)
3030

31-
The chunks storage stores each single time series into a separate object called _chunk_. Each chunk contains the timestamp-value pairs for a given period (defaults to 12 hours) of the series. chunks are then indexed by time range and labels, in order to provide a fast lookup across many (over millions) chunks.
31+
The chunks storage stores each single time series into a separate object called _Chunk_. Each Chunk contains the timestamp-value pairs for a given period (defaults to 12 hours) of the series. Chunks are then indexed by time range and labels, in order to provide a fast lookup across many (over millions) Chunks.
3232

3333
For this reason, the chunks storage consists of:
3434

35-
* An index for the chunks. This index can be backed by:
35+
* An index for the Chunks. This index can be backed by:
3636
* [Amazon DynamoDB](https://aws.amazon.com/dynamodb)
3737
* [Google Bigtable](https://cloud.google.com/bigtable)
3838
* [Apache Cassandra](https://cassandra.apache.org)
39-
* An object store for the chunk data itself, which can be:
39+
* An object store for the Chunk data itself, which can be:
4040
* [Amazon DynamoDB](https://aws.amazon.com/dynamodb)
4141
* [Google Bigtable](https://cloud.google.com/bigtable)
4242
* [Apache Cassandra](https://cassandra.apache.org)
@@ -51,11 +51,11 @@ The current schema recommendation is the **v10 schema** (v11 is still experiment
5151

5252
### Blocks storage (experimental)
5353

54-
The blocks storage is based on [Prometheus TSDB](https://prometheus.io/docs/prometheus/latest/storage/): it stores each tenant's time series into their own TSDB which write out their series to a on-disk block (defaults to 2h block range periods). Each block is composed by few files storing the chunks and the block index.
54+
The blocks storage is based on [Prometheus TSDB](https://prometheus.io/docs/prometheus/latest/storage/): it stores each tenant's time series into their own TSDB which write out their series to a on-disk Block (defaults to 2h block range periods). Each Block is composed by few files storing the chunks and the block index.
5555

56-
The TSDB chunk files contain the timestamp-value pairs for multiple series. The series inside the chunks are then indexed by a per-block index, which indexes metric names and labels to time series in the chunk files.
56+
The TSDB chunk files contain the timestamp-value pairs for multiple series. The series inside the Chunks are then indexed by a per-block index, which indexes metric names and labels to time series in the chunk files.
5757

58-
The blocks storage doesn't require a dedicated storage backend for the index. The only requirement is an object store for the block files, which can be:
58+
The blocks storage doesn't require a dedicated storage backend for the index. The only requirement is an object store for the Block files, which can be:
5959

6060
* [Amazon S3](https://aws.amazon.com/s3)
6161
* [Google Cloud Storage](https://cloud.google.com/storage/)

docs/operations/blocks-storage.md

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ weight: 1
55
slug: blocks-storage
66
---
77

8-
The blocks storage is an **experimental** Cortex storage engine based on [Prometheus TSDB](https://prometheus.io/docs/prometheus/latest/storage/): it stores all the time series for a given tenant and period (defaults to 2 hours) into a single TSDB block. Each block is composed by chunk files - containing the timestamp-value pairs for multiple series - and an index, which indexes metric names and labels to time series in the chunk files.
8+
The blocks storage is an **experimental** Cortex storage engine based on [Prometheus TSDB](https://prometheus.io/docs/prometheus/latest/storage/): it stores each tenant's time series into their own TSDB which write out their series to a on-disk Block (defaults to 2h block range periods). Each Block is composed by chunk files - containing the timestamp-value pairs for multiple series - and an index, which indexes metric names and labels to time series in the chunk files.
99

1010
The supported backends for the blocks storage are:
1111

@@ -18,24 +18,24 @@ The rest of the document assumes you have read the [Cortex architecture](../arch
1818

1919
## How it works
2020

21-
When the blocks storage is used, each **ingester** creates a per-tenant TSDB and ships the TSDB blocks - which by default are cut every 2 hours - to the long-term storage.
21+
When the blocks storage is used, each **ingester** creates a per-tenant TSDB and ships the TSDB Blocks - which by default are cut every 2 hours - to the long-term storage.
2222

23-
**Queriers** periodically iterate over the storage bucket to discover recently uploaded blocks and - for each block - download a subset of the block index - called "index header" - which is kept in memory and used to provide fast lookups.
23+
**Queriers** periodically iterate over the storage bucket to discover recently uploaded Blocks and - for each Block - download a subset of the block index - called "index header" - which is kept in memory and used to provide fast lookups.
2424

2525
### The write path
2626

2727
**Ingesters** receive incoming samples from the distributors. Each push request belongs to a tenant, and the ingester append the received samples to the specific per-tenant TSDB. The received samples are both kept in-memory and written to a write-ahead log (WAL) stored on the local disk and used to recover the in-memory series in case the ingester abruptly terminates. The per-tenant TSDB is lazily created in each ingester upon the first push request is received for that tenant.
2828

29-
The in-memory samples are periodically flushed to disk - and the WAL truncated - when a new TSDB block is cut, which by default occurs every 2 hours. Each new block cut is then uploaded to the long-term storage and kept in the ingester for some more time, in order to give queriers enough time to discover the new block from the storage and download its index header.
29+
The in-memory samples are periodically flushed to disk - and the WAL truncated - when a new TSDB Block is cut, which by default occurs every 2 hours. Each new Block cut is then uploaded to the long-term storage and kept in the ingester for some more time, in order to give queriers enough time to discover the new Block from the storage and download its index header.
3030

3131

3232
In order to effectively use the **WAL** and being able to recover the in-memory series upon ingester abruptly termination, the WAL needs to be stored to a persistent local disk which can survive in the event of an ingester failure (ie. AWS EBS volume or GCP persistent disk when running in the cloud). For example, if you're running the Cortex cluster in Kubernetes, you may use a StatefulSet with a persistent volume claim for the ingesters.
3333

3434
### The read path
3535

36-
**Queriers** - at startup - iterate over the entire storage bucket to discover all tenants blocks and - for each of them - download the index header. During this initial synchronization phase, a querier is not ready to handle incoming queries yet and its `/ready` readiness probe endpoint will fail.
36+
**Queriers** - at startup - iterate over the entire storage bucket to discover all tenants Blocks and - for each of them - download the index header. During this initial synchronization phase, a querier is not ready to handle incoming queries yet and its `/ready` readiness probe endpoint will fail.
3737

38-
Queriers also periodically re-iterate over the storage bucket to discover newly uploaded blocks (by the ingesters) and find out blocks deleted in the meanwhile, as effect of an optional retention policy.
38+
Queriers also periodically re-iterate over the storage bucket to discover newly uploaded Blocks (by the ingesters) and find out Blocks deleted in the meanwhile, as effect of an optional retention policy.
3939

4040
The blocks chunks and the entire index is never fully downloaded by the queriers. In the read path, a querier lookups the series label names and values using the in-memory index header and then download the required segments of the index and chunks for the matching series directly from the long-term storage using byte-range requests.
4141

@@ -49,14 +49,14 @@ It's important to note that - differently than the [chunks storage](../architect
4949

5050
### Compactor
5151

52-
The **compactor** is an optional - but highly recommended - service which compacts multiple blocks of a given tenant into a single optimized larger block. The compactor has two main benefits:
52+
The **compactor** is an optional - but highly recommended - service which compacts multiple Blocks of a given tenant into a single optimized larger Block. The compactor has two main benefits:
5353

54-
1. Vertically compact blocks uploaded by all ingesters for the same time range
55-
2. Horizontally compact blocks with small time ranges into a single larger block
54+
1. Vertically compact Blocks uploaded by all ingesters for the same time range
55+
2. Horizontally compact Blocks with small time ranges into a single larger Block
5656

57-
The **vertical compaction** compacts all the blocks of a tenant uploaded by any ingester for the same block range period (defaults to 2 hours) into a single block, de-duplicating samples that are originally written to N blocks as effect of the replication.
57+
The **vertical compaction** compacts all the Blocks of a tenant uploaded by any ingester for the same Block range period (defaults to 2 hours) into a single Block, de-duplicating samples that are originally written to N Blocks as effect of the replication.
5858

59-
The **horizontal compaction** triggers after the vertical compaction and compacts several blocks belonging to adjacent small range periods (2 hours) into a single larger block. Despite the total block chunks size doesn't change after this compaction, it may have a significative impact on the reduction of the index size and its index header kept in memory by queriers.
59+
The **horizontal compaction** triggers after the vertical compaction and compacts several Blocks belonging to adjacent small range periods (2 hours) into a single larger Block. Despite the total block chunks size doesn't change after this compaction, it may have a significative impact on the reduction of the index size and its index header kept in memory by queriers.
6060

6161
The compactor is **stateless**.
6262

0 commit comments

Comments
 (0)