No DAG visualization information to display for job 6252
| Stage Id ▾ | Pool Name | Description | Submitted | Duration | Tasks: Succeeded/Total | Input | Output | Shuffle Read | Shuffle Write |
|---|---|---|---|---|---|---|---|---|---|
| 115253 | tenants-pool-780 | replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD $anonfun$recordDeltaOperationInternal$1 at DatabricksLogging.scala:128
RDD: Delta Table State #1 - hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-product-productplant/_delta_logDelta Table State with Stats #1 - hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-product-productplant/_delta_log
org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139) com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128) com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117) org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87) org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138) org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128) org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118) org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87) org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265) org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204) org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87) org.apache.spark.sql.delta.stats.PreparedDeltaFileIndex.matchingFiles(PrepareDeltaScan.scala:389) org.apache.spark.sql.delta.files.TahoeFileIndex.listAddFiles(TahoeFileIndex.scala:111) org.apache.spark.sql.delta.files.TahoeFileIndex.listFiles(TahoeFileIndex.scala:103) org.apache.spark.sql.execution.FileSourceScanLike.selectedPartitions(DataSourceScanExec.scala:297) org.apache.spark.sql.execution.FileSourceScanLike.selectedPartitions$(DataSourceScanExec.scala:288) org.apache.spark.sql.execution.FileSourceScanExec.selectedPartitions$lzycompute(DataSourceScanExec.scala:607) org.apache.spark.sql.execution.FileSourceScanExec.selectedPartitions(DataSourceScanExec.scala:607) org.apache.spark.sql.execution.FileSourceScanLike.dynamicallySelectedPartitions(DataSourceScanExec.scala:330) org.apache.spark.sql.execution.FileSourceScanLike.dynamicallySelectedPartitions$(DataSourceScanExec.scala:309) | 2026/03/21 01:07:52 | 53 ms |
50/50
| 2.4 KiB |
| Stage Id ▾ | Pool Name | Description | Submitted | Duration | Tasks: Succeeded/Total | Input | Output | Shuffle Read | Shuffle Write |
|---|---|---|---|---|---|---|---|---|---|
| 115252 | Unknown Unknown | Unknown | Unknown |
0/0
|