You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When you click **Test connection**, Quix runs a short round-trip check to make sure your details are correct and that the platform can both see and use your storage.
43
43
@@ -55,7 +55,7 @@ Each step is shown in the dialog. Successful steps are marked with a ✓, and yo
55
55
**Failure**
56
56
If a step fails, you’ll see ✗ next to it along with the reason (for example, “Access denied” or “Wrong region”). This makes it easy to fix permissions or update your settings.
Copy file name to clipboardExpand all lines: docs/quix-cloud/managed-services/replay.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,10 @@
1
1
---
2
2
3
-
title: Quix Lake Replay (managed)
3
+
title: Quix Lake - Replay
4
4
description: Managed service that replays persisted datasets from Quix Lake back into Kafka with full fidelity.
5
5
---
6
6
7
-
# Quix Lake Replay
7
+
# Quix Lake - Replay
8
8
9
9
Quix Lake Replay is a managed service that streams persisted datasets from **Quix Lake** back into **Kafka**, preserving timestamps, partitions, offsets, headers, and gaps for high-fidelity re-runs and simulations.
10
10
@@ -19,17 +19,17 @@ You can launch a replay from multiple places in the Portal:
19
19
Use the pipeline canvas to add a Replay tile and start it in context of your flow. The screenshot shows the entry point on the canvas.
20
20

21
21
22
-
### From the Data Catalog
22
+
### From the Data Lake UI
23
23
24
-
Open **Quix Lake → Catalog**, select the dataset (topic + time window/keys/partitions), and start a replay directly from the catalog. The screenshot highlights the replay action on a selected dataset.
25
-

24
+
Open **Data Lake**, select the dataset (topic + time window/keys/partitions), and start a replay directly from the catalog. The screenshot highlights the replay action on a selected dataset.
25
+

Copy file name to clipboardExpand all lines: docs/quix-cloud/managed-services/sink.md
+14-14Lines changed: 14 additions & 14 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,9 +1,9 @@
1
1
---
2
-
title: Quix Lake Sink
2
+
title: Quix Lake - Sink
3
3
description: Connector that persists Kafka data into Quix Lake.
4
4
---
5
5
6
-
# Quix Lake Sink
6
+
# Quix Lake - Sink
7
7
8
8
The Quix Lake Sink writes Kafka topic data to your blob storage in **Avro** (raw messages) and **Parquet** (index and optional custom metadata), enabling fast discovery and high-fidelity **Replay**.
9
9
@@ -44,7 +44,7 @@ Metadata:
44
44
## How to run (UI)
45
45
46
46
1. Create or log in to your Quix account.
47
-
2. Go to **Connectors → Add connector → Quix quixlake Sink**.
47
+
2. Go to **Connectors → Add connector → Quix Lake - Sink**.
48
48
3. Click **Set up connector**, fill the parameters below, then **Test connection & deploy**.
49
49
50
50
!!! info "Managed service"
@@ -78,7 +78,7 @@ You can configure the sink in **YAML** or via the **Quix Cloud UI**.
78
78
*`consumerGroup` - Kafka consumer group ID (default: `quixstreams-default`)
79
79
*`autoOffsetReset` - `latest` or `earliest` (default: `latest`)
80
80
81
-
#### Quix Lake settings
81
+
#### Sink settings
82
82
83
83
*`avroCompression` - `snappy` or `gzip` (default: `snappy`)
84
84
*`maxWorkers` - threads for uploading (default: `5`)
@@ -90,8 +90,8 @@ You can configure the sink in **YAML** or via the **Quix Cloud UI**.
90
90
91
91
```yaml
92
92
deployments:
93
-
- name: Quix Lake Sink
94
-
application: quixlake.Sink
93
+
- name: Quix Lake - Sink
94
+
application: DataLake.Sink
95
95
version: latest
96
96
deploymentType: Managed
97
97
resources:
@@ -132,7 +132,7 @@ deployments:
132
132
* **Parquet (Custom metadata, optional)**
133
133
Your key–value annotations (`Topic`, `Key`, `MetadataKey`, `MetadataValue`, `UpdatedUtc`) used for search and grouping in the Catalog.
134
134
135
-
See **Open format** for full schemas and layout.
135
+
See [Open format](../quixlake/open-format.md) for full schemas and layout.
136
136
137
137
## Operational behavior
138
138
@@ -151,20 +151,20 @@ See **Open format** for full schemas and layout.
151
151
152
152
* **Logs**: per-segment lifecycle (rolling, upload, index write), retries, and timings
153
153
* **Metrics**: records persisted, bytes uploaded, active uploads, average upload speed
154
-
* **Catalog**: new datasets appear as index files land; use **Refresh** if you need to surface them sooner
154
+
* **Data Lake UI*: new datasets appear as index files land; use **Refresh** if you need to surface them sooner
155
155
156
156
## Security
157
157
158
158
* Uses the **cluster’s** blob storage connection (scoped credentials; one bucket/container per connection)
159
159
* Honor your cloud controls: IAM roles, key rotation, server-side encryption, access logs, retention
160
-
* The sink does not delete raw data; deletion flows through **Catalog** with soft-delete and trash retention
160
+
* The sink does not delete raw data; deletion flows through **Data Lake API** with soft-delete and trash retention
161
161
162
162
## Troubleshooting
163
163
164
164
* **Access denied**
165
165
Verify the blob connection’s permissions: list, read, write, and delete on the bucket/container.
166
-
* **Nothing appears in Catalog**
167
-
Check sink logs for successful index writes; click **Refresh** in Catalog; ensure time filters include the new data.
166
+
* **Nothing appears in Data Lake UI**
167
+
Check sink logs for successful index writes; click **Refresh** in the UI; ensure time filters include the new data.
168
168
* **Small-file explosion**
169
169
Increase `rollBytes` and/or `rollSeconds`, or add a replica to smooth throughput.
170
170
* **Slow uploads**
@@ -173,6 +173,6 @@ See **Open format** for full schemas and layout.
173
173
## See also
174
174
175
175
* [Open format](../quixlake/open-format.md)
176
-
* [Quix Lake User Interface](../quixlake/user-interface.md)
Copy file name to clipboardExpand all lines: docs/quix-cloud/quixlake/api.md
+60-9Lines changed: 60 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,7 +1,7 @@
1
1
---
2
2
3
-
title: Quix Lake API
4
-
description: Programmatic access to Quix Lake for search, metadata, file discovery, and lifecycle operations. Backend for the Quix Lake UI; Metadata endpoints are the primary integration surface for your applications.
3
+
title: Quix Lake - API
4
+
description: Programmatic access to Quix Lake for search, metadata, file discovery, and lifecycle operations. Backend for the Quix Lake UI; Metadata endpoints are the primary integration surface for your external re-indexing applications.
5
5
---
6
6
7
7
# Overview
@@ -17,9 +17,19 @@ See the UI page: [Quix Lake User Interface](./user-interface.md).
These routes back the catalog’s topic/key lists, facets, and search results.
21
+
## Catalog endpoints
22
+
23
+
The **Catalog** endpoints powers discovery features across workspaces and topics. These endpoints let you search indexed stream metadata, explore available topics and keys, and manage cache refreshes to ensure results are up to date.
24
+
25
+
Use these routes to:
26
+
27
+
* Perform flexible searches (text, fuzzy, prefix/suffix, and time-bound queries).
28
+
* Enumerate available topics, keys, and metadata facets for filtering.
29
+
* Keep catalog results fresh with cache refresh operations at the workspace or topic level.
30
+
* Retrieve workspace and global metadata keys to support consistent UI-driven filtering.
31
+
32
+
Together, these endpoints back the catalog’s search grid, topic/key lists, and filtering facets.
23
33
24
34
### Search stream metadata
25
35
@@ -74,9 +84,22 @@ Returns workspace identifiers that have discoverable data for the caller.
74
84
!!! note
75
85
The search response includes a total-count header so clients can page results consistently with the UI.
76
86
77
-
## Data
78
87
79
-
Programmatic visibility into raw objects and time bounds—useful for exports, verification, and operational tooling.
88
+
Here’s a refined introduction for the **Data** section that aligns in tone and clarity with the improved **Catalog** section:
89
+
90
+
91
+
92
+
## Data endpoints
93
+
94
+
The **Data** endpoints provides direct visibility into raw storage objects and their temporal ranges. These endpoints are designed for operational use cases such as exports, audits, verification, and impact analysis.
95
+
96
+
Use these routes to:
97
+
98
+
* Enumerate the exact Avro segment files that make up a selection.
99
+
* Identify all files associated with a key to preview or audit deletions.
100
+
* Compute temporal bounds (min/max timestamps and partitions) for sets of keys.
101
+
102
+
Together, these endpoints give precise, programmatic insight into how cataloged data is physically stored and bounded in time.
80
103
81
104
### Get timestamped file descriptors
82
105
@@ -96,9 +119,18 @@ Reports minimum/maximum timestamps and observed partitions for a set of keys. Us
96
119
!!! tip
97
120
A common flow is to use **search** to find candidate keys, then use **files** to enumerate exact object paths.
98
121
122
+
123
+
99
124
## Data Deletion
100
125
101
-
Safe lifecycle operations. Defaults are **soft delete** to protect data; hard delete removes both metadata and files.
126
+
The **Data Deletion API** supports safe lifecycle management of cataloged data. By default, deletions are **soft**, preserving underlying files and enabling recovery. When required, **hard deletes** can permanently remove both metadata and storage objects.
127
+
128
+
Use these routes to:
129
+
130
+
* Delete metadata and files for a single key or multiple keys (soft or hard).
131
+
* Restore streams that were previously soft-deleted, individually or in batches.
132
+
133
+
These operations ensure controlled data cleanup while supporting compliance and recovery workflows.
102
134
103
135
### Delete metadata/files for a single key
104
136
@@ -123,11 +155,28 @@ Clears soft-delete markers for multiple keys.
123
155
!!! warning
124
156
Use hard delete only when retention and compliance requirements allow it.
125
157
158
+
159
+
126
160
## Metadata
127
161
128
-
Attach custom **key/value properties** to datasets and query by those properties. This is intended for your applications to enrich datasets created by the [Quix Lake Sink (managed)](../managed-services/sink.md), so they’re easy to group, filter, and audit across API and UI.
162
+
The **Metadata API** lets you enrich datasets with custom **key/value properties** and query them later for grouping, filtering, lineage, and auditing. This is especially useful when working with datasets produced by the [Quix Lake Sink (managed)](../managed-services/sink.md), enabling your applications to attach meaningful business or operational context.
163
+
164
+
Typical metadata examples include:
165
+
166
+
* Machine or device identifiers
167
+
* Sensor calibration ranges
168
+
* Driver, batch, or run identifiers
169
+
* Experiment tags or quality tiers
170
+
* Flattened JSON values
171
+
172
+
Use these routes to:
173
+
174
+
* Add or update metadata properties for a key.
175
+
* Retrieve all metadata associated with a key.
176
+
* Soft-delete all metadata for a key, or selectively remove specific properties.
177
+
178
+
These operations provide a lightweight but powerful mechanism for managing dataset context across both API and UI.
129
179
130
-
**Good examples of properties:** machine id, sensor range, driver, concrete batch, simple JSON flattened to strings, experiment or run identifiers, quality tiers.
131
180
132
181
### Upsert metadata entries
133
182
@@ -152,6 +201,8 @@ Removes only the listed property names.
152
201
!!! tip
153
202
When searching, you can request the full tag set per result to drive rules without extra reads.
Copy file name to clipboardExpand all lines: docs/quix-cloud/quixlake/open-format.md
+11-9Lines changed: 11 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,10 +1,10 @@
1
1
---
2
2
3
-
title: Open format
3
+
title: Quix Lake - Open format
4
4
description: How Quix Lake structures data in Avro and Parquet for portability and performance.
5
5
---
6
6
7
-
# Open format
7
+
# Quix Lake - Open format
8
8
9
9
Quix Lake stores Kafka messages and metadata as **open files** in your blob storage (S3, GCS, Azure Blob, MinIO). The layout favors portability, fast discovery, and full-fidelity replay.
10
10
@@ -92,10 +92,10 @@ Quix Lake stores Kafka messages and metadata as **open files** in your blob stor
92
92
93
93
## How it flows
94
94
95
-
* Ingest: write Avro into partitioned folders
96
-
* Index: write Parquet descriptors alongside
97
-
* Discover: Catalog and APIs read Parquet to list and filter quickly
98
-
* Use: Replay to Kafka, or query with your engines of choice
95
+
* Ingest: write Avro into partitioned folders ([Quix Lake - Sink](../managed-services/sink.md))
96
+
* Index: write Parquet descriptors alongside ([Quix Lake - Sink](../managed-services/sink.md))
97
+
* Discover: UI and APIs read Parquet to list and filter your datasets ([Quix Lake - API](./api.md))
98
+
* Use: Replay to Kafka, or query with your engines of choice ([Quix Lake - Replay](../managed-services/replay.md))
99
99
100
100
## Guarantees
101
101
@@ -106,6 +106,8 @@ Quix Lake stores Kafka messages and metadata as **open files** in your blob stor
106
106
107
107
## See also
108
108
109
-
*[Quix Lake User Interface](./user-interface.md) - discover datasets programmatically
110
-
*[Quix Lake Replay (managed)](../managed-services/replay.md) - send datasets back to Kafka
111
-
*[Blob storage connections](../../deploy/blob-storage.md) - wire up your bucket or container
109
+
*[Quix Lake - Sink](../managed-services/sink.md) - persist data from Kafka to your Blob Storage
110
+
*[Quix Lake - API](./api.md) - discover datasets programmatically
111
+
*[Quix Lake - User Interface](./user-interface.md) - discover datasets using Quix Cloud user interface
112
+
*[Quix Lake - Replay](../managed-services/replay.md) - send datasets back to Kafka
113
+
*[Blob storage connections](../managed-services/blob-storage.md) - wire up your bucket or container
Copy file name to clipboardExpand all lines: docs/quix-cloud/quixlake/overview.md
+9-8Lines changed: 9 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -5,7 +5,7 @@ description: The central storage layer of Quix Cloud for capturing and managing
5
5
6
6
# What is Quix Lake
7
7
8
-
Quix Lake is the central storage layer of Quix Cloud. It captures, organizes, and manages Kafka topic data in an open, file-based format on blob storage systems such as Amazon S3, Azure Blob, Google Cloud Storage, or MinIO.
8
+
**Quix Lake** is the central storage layer of **Quix Cloud**. It captures, organizes, and manages Kafka topic data in an open, file-based format on blob storage systems such as Amazon S3, Azure Blob, Google Cloud Storage, or MinIO.
9
9
10
10
Instead of relying on proprietary databases, Quix Lake uses open formats and Hive-style partitioning so your data stays:
11
11
@@ -23,7 +23,7 @@ Earlier persistence options were tied to specific databases and SDKs, which limi
23
23
24
24
* Kafka messages are persisted exactly as they arrive, including timestamps, headers, partitions, offsets, and idle gaps
25
25
* Metadata is indexed alongside raw data to enable fast discovery without scanning Avro
26
-
* Services like **Catalog**, **Replay**, and sinks operate directly on the open files in your bucket
26
+
* Services like **API**, **Replay**, and **Sink** operate directly on the open files in your bucket
27
27
* You keep full control of storage, security, and lifecycle in your own cloud account
28
28
29
29
## Where your data lives
@@ -47,7 +47,7 @@ See **[open format](./open-format.md)** for the full layout and schemas.
47
47
48
48
## What you can do
49
49
50
-
***Explore datasets** with the **Quix Lake Catalog** UI or API
50
+
***Explore datasets** with the **Quix Lake** UI or API
51
51
***Replay** persisted datasets back into Kafka with full fidelity
52
52
***Search and filter** by time ranges, topics, keys, and custom metadata
53
53
***Query externally** using DuckDB, Spark, Trino, Athena, or BigQuery over Avro and Parquet
@@ -59,7 +59,7 @@ See **[open format](./open-format.md)** for the full layout and schemas.
59
59
60
60
1.**Ingest**: a sink writes raw Kafka messages to Avro files in your storage
61
61
2.**Index**: Parquet index files summarize time, partition, offsets, and sizes
62
-
3.**Discover**: the Catalog and APIs read the index to list and filter quickly
62
+
3.**Discover**: the UI and APIs read the index to list and filter quickly
63
63
4.**Replay**: any discovered dataset can be streamed back to Kafka with original order and timing preserved or simulated
64
64
5.**Use**: build pipelines that mix historical data with live streams, and run queries over Parquet
65
65
@@ -71,7 +71,8 @@ See **[open format](./open-format.md)** for the full layout and schemas.
71
71
## See also
72
72
73
73
*[Open format](./open-format.md)
74
-
*[Quix Lake User Interface](./user-interface.md)
75
-
*[Quix Lake Replay (managed)](../managed-services/replay.md)
76
-
*[Quix Lake Sink (managed)](../managed-services/sink.md)
0 commit comments