Logs: add Pipelines as Logpush destination#29469
Logs: add Pipelines as Logpush destination#29469Marcinthecloud wants to merge 1 commit intoproductionfrom
Conversation
|
This pull request requires reviews from CODEOWNERS as it changes files that match the following patterns:
|
|
/bonk review this, ensuring that:
fix trivial examples, and make suggestions on the PR otherwise. |
| WHERE EdgeResponseStatus >= 400; | ||
| ``` | ||
|
|
||
| Pipelines SQL supports string functions, regex, hashing (SHA-256, MD5), JSON extraction, timestamp conversion, conditional expressions, and more. For the full list, refer to the [Pipelines SQL reference](/pipelines/sql-reference/). |
There was a problem hiding this comment.
Per the style guide, use 'Refer to' instead of 'See':
| Pipelines SQL supports string functions, regex, hashing (SHA-256, MD5), JSON extraction, timestamp conversion, conditional expressions, and more. For the full list, refer to the [Pipelines SQL reference](/pipelines/sql-reference/). | |
| Pipelines SQL supports string functions, regex, hashing (SHA-256, MD5), JSON extraction, timestamp conversion, conditional expressions, and more. For the full list, refer to the [Pipelines SQL reference](/pipelines/sql-reference/). |
| - Choose the file format: Raw Parquet, JSON, or [R2 Data Catalog (Apache Iceberg)](/r2/data-catalog/). | ||
| - Optionally, you can edit advanced settings such as rolling policy, partitioning, and compression. For more information about these settings, refer to the [Pipelines Sinks documentation](/pipelines/sinks/). | ||
|
|
||
| 7. Select **Generate token** to automatically create a token that will be used for authentication to R2, R2 Data Catalog, and the Pipelines Stream. |
There was a problem hiding this comment.
Bold the UI element:
| 7. Select **Generate token** to automatically create a token that will be used for authentication to R2, R2 Data Catalog, and the Pipelines Stream. | |
| 7. Select **Generate token** to automatically create a token that will be used for authentication to R2, R2 Data Catalog, and the Pipelines Stream. |
|
Posted 2 inline suggestions on PR #29469:
The documentation is well-structured with proper frontmatter, correct component usage ( |
|
|
||
| 3. Select **Pipelines** as the destination. | ||
|
|
||
| 4. Select the dataset from the dropdown. The schema for the respective dataset will automatically be applied to the Pipeline that is created. |
There was a problem hiding this comment.
Should we mention that the Pipelines creation wizard would open on clicking the create button?
|
|
||
| For a full list of fields available in each dataset, refer to [Datasets](/logs/logpush/logpush-job/datasets/). | ||
|
|
||
| ## Manage via the Cloudflare dashboard |
There was a problem hiding this comment.
Would it help to have some screenshots here?
Adds documentation for using Cloudflare Pipelines as a native Logpush destination, enabling SQL transformation of logs before storage in R2.