Skip to content

Commit 2b24d7f

Browse files
[mq] working branch - merge 99e2abd on top of master at f9b9f21
{"baseBranch":"master","baseCommit":"f9b9f2192fdd175c31898dd8fc37903629e61ab0","createdAt":"2025-05-16T20:41:43.223981Z","headSha":"99e2abda33308e4f7133f6711a776e7644ec07f5","id":"aa9913b0-5e37-4740-aeca-e5ed6894b078","priority":"200","pullRequestNumber":"29384","queuedAt":"2025-05-16T20:41:43.223338Z","status":"STATUS_QUEUED"}
2 parents 95c4feb + 99e2abd commit 2b24d7f

File tree

2 files changed

+9
-0
lines changed

2 files changed

+9
-0
lines changed

content/en/data_jobs/databricks.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -79,9 +79,12 @@ Optionally, you can add tags to your Databricks cluster and Spark performance me
7979
|--------------------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------|
8080
| DD_TAGS | Add tags to Databricks cluster and Spark performance metrics. Comma or space separated key:value pairs. Follow [Datadog tag conventions][1]. Example: `env:staging,team:data_engineering` |
8181
| DD_ENV | Set the `env` environment tag on metrics, traces, and logs from this cluster. |
82+
| DD_LOGS_CONFIG_PROCESSING_RULES | Filter the logs collected with processing rules. See [Advanced Log Collection][3] for more details. |
83+
8284

8385
[1]: /getting_started/tagging/
8486
[2]: https://docs.databricks.com/api/workspace/clusters/edit#spark_env_vars
87+
[3]: /agent/logs/advanced_log_collection/?tab=environmentvariable#global-processing-rules
8588

8689
{{% /tab %}}
8790

@@ -136,11 +139,13 @@ Optionally, you can also set other init script parameters and Datadog environmen
136139
| WORKER_LOGS_ENABLED | Collect spark workers logs in Datadog. | false |
137140
| DD_TAGS | Add tags to Databricks cluster and Spark performance metrics. Comma or space separated key:value pairs. Follow [Datadog tag conventions][4]. Example: `env:staging,team:data_engineering` | |
138141
| DD_ENV | Set the `env` environment tag on metrics, traces, and logs from this cluster. | |
142+
| DD_LOGS_CONFIG_PROCESSING_RULES | Filter the logs collected with processing rules. See [Advanced Log Collection][5] for more details. | |
139143

140144
[1]: https://app.datadoghq.com/organization-settings/api-keys
141145
[2]: /getting_started/site/
142146
[3]: https://github.com/DataDog/datadog-agent/blob/main/pkg/fleet/installer/setup/djm/databricks.go
143147
[4]: /getting_started/tagging/
148+
[5]: /agent/logs/advanced_log_collection/?tab=environmentvariable#global-processing-rules
144149

145150
{{% /tab %}}
146151

@@ -191,12 +196,14 @@ Optionally, you can also set other init script parameters and Datadog environmen
191196
| WORKER_LOGS_ENABLED | Collect spark workers logs in Datadog. | false |
192197
| DD_TAGS | Add tags to Databricks cluster and Spark performance metrics. Comma or space separated key:value pairs. Follow [Datadog tag conventions][4]. Example: `env:staging,team:data_engineering` | |
193198
| DD_ENV | Set the `env` environment tag on metrics, traces, and logs from this cluster. | |
199+
| DD_LOGS_CONFIG_PROCESSING_RULES | Filter the logs collected with processing rules. See [Advanced Log Collection][5] for more details. | |
194200

195201

196202
[1]: https://app.datadoghq.com/organization-settings/api-keys
197203
[2]: /getting_started/site/
198204
[3]: https://github.com/DataDog/datadog-agent/blob/main/pkg/fleet/installer/setup/djm/databricks.go
199205
[4]: /getting_started/tagging/
206+
[5]: /agent/logs/advanced_log_collection/?tab=environmentvariable#global-processing-rules
200207

201208
3. Click **Confirm**.
202209

content/en/data_jobs/emr.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -125,8 +125,10 @@ When you create a new EMR cluster in the [Amazon EMR console][4], add a bootstra
125125
| DD_TAGS | Add tags to EMR cluster and Spark performance metrics. Comma or space separated key:value pairs. Follow [Datadog tag conventions][15]. Example: `env:staging,team:data_engineering` | |
126126
| DD_ENV | Set the `env` environment tag on metrics, traces, and logs from this cluster. | |
127127
| DD_EMR_LOGS_ENABLED | Send Spark driver and worker logs to Datadog. | false |
128+
| DD_LOGS_CONFIG_PROCESSING_RULES | Filter the logs collected with processing rules. See [Advanced Log Collection][16] for more details. | |
128129

129130
[15]: /getting_started/tagging/
131+
[16]: /agent/logs/advanced_log_collection/?tab=environmentvariable#global-processing-rules
130132

131133
1. On the **Create Cluster** page, find the **Bootstrap actions** section. Click **Add** to bring up the **Add bootstrap action** dialog.
132134
{{< img src="data_jobs/emr/add_bootstrap_action_without_arguments.png" alt="Amazon EMR console, Create Cluster, Add Bootstrap Action dialog. Text fields for name, script location, and arguments." style="width:80%;" >}}

0 commit comments

Comments
 (0)