|
3244
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storeOGRDocumentInformation(FileStorageAdapterImpl.java:347)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.TraceabilityDataAccessImpl.storeOGRDocumentInformation(TraceabilityDataAccessImpl.java:40)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.OpenGoodsMovementsTracerImpl.publish(OpenGoodsMovementsTracerImpl.java:51)
java.base/java.lang.Iterable.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:63)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/03/20 20:08:19
|
9 ms
|
|
|
|
3509
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storeOGRDocumentInformation(FileStorageAdapterImpl.java:347)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.TraceabilityDataAccessImpl.storeOGRDocumentInformation(TraceabilityDataAccessImpl.java:40)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.OpenGoodsMovementsTracerImpl.publish(OpenGoodsMovementsTracerImpl.java:51)
java.base/java.lang.Iterable.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:63)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
jdk.internal.reflect.GeneratedMethodAccessor1659.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/03/21 01:08:17
|
9 ms
|
|
|
|
2834
|
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0
+details
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:04:52
|
10 ms
|
|
|
|
2970
|
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0
+details
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:57:06
|
10 ms
|
|
|
|
3191
|
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0
+details
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:06:48
|
10 ms
|
|
|
|
3456
|
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0
+details
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
jdk.internal.reflect.GeneratedMethodAccessor1635.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:06:20
|
10 ms
|
|
|
|
2887
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.storeOGRDocumentInformation(FileStorageAdapterImpl.java:347)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.TraceabilityDataAccessImpl.storeOGRDocumentInformation(TraceabilityDataAccessImpl.java:40)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.OpenGoodsMovementsTracerImpl.publish(OpenGoodsMovementsTracerImpl.java:51)
java.base/java.lang.Iterable.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:63)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/03/20 01:06:40
|
15 ms
|
|
|
|
3004
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:17
|
17 ms
|
|
|
|
3010
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:20
|
18 ms
|
|
|
|
3001
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:15
|
19 ms
|
|
|
|
2849
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:10
|
20 ms
|
|
|
|
3471
|
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:48
|
20 ms
|
|
|
|
2969
|
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0
+details
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:57:06
|
23 ms
|
|
|
|
2987
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:06
|
23 ms
|
|
|
|
2833
|
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0
+details
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:04:52
|
27 ms
|
|
|
|
3455
|
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0
+details
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
jdk.internal.reflect.GeneratedMethodAccessor1635.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:06:20
|
27 ms
|
|
|
|
3190
|
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0
+details
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:06:48
|
28 ms
|
|
|
|
3206
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:07:52
|
47 ms
|
|
|
|
2960
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = PrepareStreams activityId = c03cb5ed-47aa-37f4-b7d5-c50c37bda41b workflowType = CalculateOrderProposalsWorkflow workflowId = 0020c3ec-5ab0-324c-997b-beeff7b311d9 attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = PrepareStreams activityId = c03cb5ed-47aa-37f4-b7d5-c50c37bda41b workflowType = CalculateOrderProposalsWorkflow workflowId = 0020c3ec-5ab0-324c-997b-beeff7b311d9 attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.Dataset.javaRDD(Dataset.scala:3270)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.CachePortImpl.<init>(CachePortImpl.java:66)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:38)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:11)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl$LifecycleControlledObjectHolder.<init>(CalculationObjectProviderImpl.java:23)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lambda$lookupOrCreate$0(CalculationObjectProviderImpl.java:83)
java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lookupOrCreate(CalculationObjectProviderImpl.java:83)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.get(CalculationObjectProviderImpl.java:57)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
|
2026/03/20 05:56:50
|
53 ms
|
|
|
|
3181
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = ffc7fa2c-d8df-38fa-88f0-fd3124a069ad workflowType = CalculateOrderProposalsWorkflow workflowId = f1005e7f-1925-3940-8274-10d50a8b3333 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = ffc7fa2c-d8df-38fa-88f0-fd3124a069ad workflowType = CalculateOrderProposalsWorkflow workflowId = f1005e7f-1925-3940-8274-10d50a8b3333 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.Dataset.javaRDD(Dataset.scala:3270)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.CachePortImpl.<init>(CachePortImpl.java:66)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:38)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:11)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl$LifecycleControlledObjectHolder.<init>(CalculationObjectProviderImpl.java:23)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lambda$lookupOrCreate$0(CalculationObjectProviderImpl.java:83)
java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lookupOrCreate(CalculationObjectProviderImpl.java:83)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.get(CalculationObjectProviderImpl.java:57)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
|
2026/03/20 20:06:37
|
54 ms
|
|
|
|
2824
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = 9e1ec4a1-ddd5-3bbd-a75a-1c206f456548 workflowType = CalculateOrderProposalsWorkflow workflowId = 6776ab81-8705-353e-bb18-6ea8447555d8 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = 9e1ec4a1-ddd5-3bbd-a75a-1c206f456548 workflowType = CalculateOrderProposalsWorkflow workflowId = 6776ab81-8705-353e-bb18-6ea8447555d8 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.Dataset.javaRDD(Dataset.scala:3270)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.CachePortImpl.<init>(CachePortImpl.java:66)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:38)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:11)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl$LifecycleControlledObjectHolder.<init>(CalculationObjectProviderImpl.java:23)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lambda$lookupOrCreate$0(CalculationObjectProviderImpl.java:83)
java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lookupOrCreate(CalculationObjectProviderImpl.java:83)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.get(CalculationObjectProviderImpl.java:57)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
|
2026/03/20 01:04:43
|
58 ms
|
|
|
|
3446
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = 272f2b55-32ed-342f-81a4-d6e310c1b3aa workflowType = CalculateOrderProposalsWorkflow workflowId = 2c47fb86-b292-3663-a679-aa9cf2542963 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = PrepareStreams activityId = 272f2b55-32ed-342f-81a4-d6e310c1b3aa workflowType = CalculateOrderProposalsWorkflow workflowId = 2c47fb86-b292-3663-a679-aa9cf2542963 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.Dataset.javaRDD(Dataset.scala:3270)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.CachePortImpl.<init>(CachePortImpl.java:66)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:38)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.OrderQuantityKpiProviderImpl.create(OrderQuantityKpiProviderImpl.java:11)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl$LifecycleControlledObjectHolder.<init>(CalculationObjectProviderImpl.java:23)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lambda$lookupOrCreate$0(CalculationObjectProviderImpl.java:83)
java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.lookupOrCreate(CalculationObjectProviderImpl.java:83)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.cachelifecycle.CalculationObjectProviderImpl.get(CalculationObjectProviderImpl.java:57)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.prepareStreams(KpiCalculationStreamingActivityImpl.java:47)
jdk.internal.reflect.GeneratedMethodAccessor1609.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
|
2026/03/21 01:06:11
|
59 ms
|
|
|
|
3062
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.classic.DataFrameWriter.save(DataFrameWriter.scala:126)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.KafkaAdapter.publishMessageToBTPKafkaWithEncryption(KafkaAdapter.java:173)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.KafkaAdapter.notifyTraceabilityDataWritten(KafkaAdapter.java:72)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.TraceabilityDataAccessImpl.lambda$storeOGRDocumentInformation$0(TraceabilityDataAccessImpl.java:41)
java.base/java.util.Optional.ifPresent(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.TraceabilityDataAccessImpl.storeOGRDocumentInformation(TraceabilityDataAccessImpl.java:41)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.OpenGoodsMovementsTracerImpl.publish(OpenGoodsMovementsTracerImpl.java:51)
java.base/java.lang.Iterable.forEach(Unknown Source)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:63)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
|
2026/03/20 06:00:17
|
60 ms
|
[5613]
|
|
|
3189
|
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0
+details
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:06:48
|
67 ms
|
|
|
|
2968
|
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0
+details
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:57:06
|
78 ms
|
|
|
|
3201
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 20:07:48
|
78 ms
|
[5815]
|
|
|
3454
|
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0
+details
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
jdk.internal.reflect.GeneratedMethodAccessor1635.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:06:19
|
78 ms
|
|
|
|
2832
|
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0
+details
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:04:52
|
80 ms
|
|
|
|
3466
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/21 01:07:44
|
83 ms
|
[6233]
|
|
|
2841
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 01:06:05
|
89 ms
|
[5110]
|
|
|
3008
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:19
|
91 ms
|
[5382]
|
|
|
2999
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:14
|
94 ms
|
[5374]
|
|
|
2844
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 01:06:06
|
0.1 s
|
[5113]
|
|
|
2981
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:59:01
|
0.1 s
|
[5343]
|
|
|
3204
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:07:51
|
0.2 s
|
[5820]
|
|
|
2996
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:59:13
|
0.2 s
|
[5371]
|
|
|
2847
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:09
|
0.2 s
|
[5118]
|
|
|
3214
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 20:07:57
|
0.2 s
|
[5839]
|
|
|
3479
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/21 01:07:53
|
0.2 s
|
[6257]
|
|
|
3469
|
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:47
|
0.2 s
|
[6238]
|
|
|
3464
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:07:44
|
0.2 s
|
[6231]
|
|
|
2985
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:05
|
0.2 s
|
[5349]
|
|
|
3009
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:19
|
0.2 s
|
[5383]
|
|
|
2852
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:81)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/03/20 01:06:13
|
0.2 s
|
[5129][5130]
|
|
|
2979
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:59:01
|
0.2 s
|
[5341]
|
|
|
2857
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 01:06:15
|
0.2 s
|
[5137]
|
|
|
2966
|
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0
+details
stream-10000000678-1
id = 577d104d-3939-4d5b-b4cf-e36fe46192c3
runId = 1bf8bf4d-82dd-4e25-8e77-025080b5dea7
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:57:03
|
0.2 s
|
|
|
|
2830
|
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0
+details
stream-10000000348-1
id = fd7f73ea-5a48-4bae-a4e0-157d03281ce8
runId = 470bb194-9c35-4eb7-8c20-404052af6c51
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:04:49
|
0.2 s
|
|
|
|
2995
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:59:13
|
0.2 s
|
[5370]
|
|
|
3452
|
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0
+details
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
jdk.internal.reflect.GeneratedMethodAccessor1635.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:06:17
|
0.2 s
|
|
|
|
2998
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:14
|
0.2 s
|
[5373]
|
|
|
3463
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/21 01:07:43
|
0.2 s
|
[6230]
|
|
|
2978
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:59:00
|
0.3 s
|
[5340]
|
|
|
3540
|
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 3
+details
stream-10000000350-1
id = dc642002-2b52-4744-bf90-967729f330bc
runId = 81faa882-918e-4a51-b5e3-3a01a5c048df
batch = 3 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
jdk.internal.reflect.GeneratedMethodAccessor1635.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:10:50
|
0.3 s
|
[6377]
|
|
|
2842
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:06:06
|
0.3 s
|
[5111]
|
|
|
3209
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:81)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/03/20 20:07:55
|
0.3 s
|
[5831][5832]
|
|
|
3006
|
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 8
+details
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 8 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:18
|
0.3 s
|
[5379]
|
|
|
3003
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:17
|
0.3 s
|
[5377]
|
|
|
3197
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 20:07:46
|
0.3 s
|
[5811]
|
|
|
3474
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:81)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
jdk.internal.reflect.GeneratedMethodAccessor1659.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
|
2026/03/21 01:07:51
|
0.3 s
|
[6249][6250]
|
|
|
3205
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:07:51
|
0.3 s
|
[5821]
|
|
|
2982
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:59:01
|
0.3 s
|
[5344]
|
|
|
3465
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/21 01:07:44
|
0.3 s
|
[6232]
|
|
|
2980
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:59:01
|
0.3 s
|
[5342]
|
|
|
2976
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:58:59
|
0.3 s
|
[5338]
|
|
|
3011
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:20
|
0.3 s
|
[5384]
|
|
|
3199
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.util.threads.DeltaThreadPool.$anonfun$submit$1(DeltaThreadPool.scala:39)
java.base/java.util.concurrent.FutureTask.run(Unknown Source)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.$anonfun$run$1(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.scala:18)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:77)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingHelper.runWithCaptured$(SparkThreadLocalForwardingThreadPoolExecutor.scala:60)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.runWithCaptured(SparkThreadLocalForwardingThreadPoolExecutor.scala:116)
org.apache.spark.sql.delta.util.threads.SparkThreadLocalCapturingRunnable.run(SparkThreadLocalForwardingThreadPoolExecutor.scala:119)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:07:47
|
0.3 s
|
[5813]
|
|
|
2848
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:09
|
0.3 s
|
[5119]
|
|
|
3470
|
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:47
|
0.4 s
|
[6239]
|
|
|
2750
|
stream-10000000347-1
id = 58eebb84-cfee-4087-b7b1-eb2de12d70e3
runId = 08a59b22-7060-442f-a3da-bada2bf2f5df
batch = 10
+details
stream-10000000347-1
id = 58eebb84-cfee-4087-b7b1-eb2de12d70e3
runId = 08a59b22-7060-442f-a3da-bada2bf2f5df
batch = 10 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/19 20:13:50
|
0.4 s
|
[4983]
|
|
|
2843
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 01:06:06
|
0.4 s
|
[5112]
|
|
|
2990
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.classic.Dataset.count(Dataset.scala:1499)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.InsightHandlerDataAccessImpl.mapExternalProductPlantIds(InsightHandlerDataAccessImpl.java:81)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.insights.InsightHandlerImpl.createUnknownDemandInsight(InsightHandlerImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.attachRawDemand(DemandSnapshotEnhancementImpl.java:65)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DemandSnapshotEnhancementImpl.withDemand(DemandSnapshotEnhancementImpl.java:35)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:56)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
|
2026/03/20 05:59:10
|
0.4 s
|
[5362][5363]
|
|
|
3472
|
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:48
|
0.4 s
|
[6240]
|
|
|
2997
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:14
|
0.4 s
|
[5372]
|
|
|
3462
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/21 01:07:42
|
0.4 s
|
[6229]
|
|
|
3198
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 20:07:46
|
0.4 s
|
[5812]
|
|
|
3187
|
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0
+details
stream-10000000349-1
id = 93409909-6965-46ba-a5f4-cc04bff91a3e
runId = 1785dd42-a3bd-40d2-8e91-765a57162f64
batch = 0 org.apache.spark.sql.classic.DataStreamWriter.start(DataStreamWriter.scala:136)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.storageaccess.FileStorageAdapterImpl.triggerOrderProposalUpdateStream(FileStorageAdapterImpl.java:171)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.KpiCalculationStreamingActivityImpl.runStreamWithId(KpiCalculationStreamingActivityImpl.java:63)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:06:45
|
0.4 s
|
|
|
|
3200
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 20:07:47
|
0.4 s
|
[5814]
|
|
|
2840
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 01:06:04
|
0.5 s
|
[5109]
|
|
|
2850
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:10
|
0.5 s
|
[5120]
|
|
|
3222
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:54)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:08:01
|
0.5 s
|
[5850][5851]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3223
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:08:01
|
0.1 s
|
[5849]
|
|
|
3005
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:17
|
0.5 s
|
[5378]
|
|
|
2977
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.DeltaLog$.recordOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.DeltaLog$.recordDeltaOperation(DeltaLog.scala:693)
org.apache.spark.sql.delta.DeltaLog$.createDeltaLog$1(DeltaLog.scala:972)
org.apache.spark.sql.delta.DeltaLog$.$anonfun$apply$5(DeltaLog.scala:996)
com.google.common.cache.LocalCache$LocalManualCache$1.load(LocalCache.java:4903)
com.google.common.cache.LocalCache$LoadingValueReference.loadFuture(LocalCache.java:3574)
com.google.common.cache.LocalCache$Segment.loadSync(LocalCache.java:2316)
com.google.common.cache.LocalCache$Segment.lockedGetOrLoad(LocalCache.java:2190)
com.google.common.cache.LocalCache$Segment.get(LocalCache.java:2080)
com.google.common.cache.LocalCache.get(LocalCache.java:4017)
com.google.common.cache.LocalCache$LocalManualCache.get(LocalCache.java:4898)
org.apache.spark.sql.delta.DeltaLog$.getDeltaLogFromCache$1(DeltaLog.scala:995)
org.apache.spark.sql.delta.DeltaLog$.initializeDeltaLog$1(DeltaLog.scala:1006)
org.apache.spark.sql.delta.DeltaLog$.apply(DeltaLog.scala:1017)
|
2026/03/20 05:58:59
|
0.5 s
|
[5339]
|
|
|
3207
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:07:52
|
0.5 s
|
[5822]
|
|
|
3220
|
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD
+details
replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:53)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 20:08:00
|
0.5 s
|
[5845][5846][5847][5848]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3221
|
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:08:00
|
0.2 s
|
[5844]
|
|
|
2986
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:05
|
0.6 s
|
[5350]
|
|
|
3014
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:53)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:59:21
|
0.6 s
|
[5389][5390][5391][5392]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3015
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:22
|
0.1 s
|
[5388]
|
|
|
2983
|
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 11
+details
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 11 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:02
|
0.6 s
|
[5346]
|
|
|
3016
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:54)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:59:22
|
0.7 s
|
[5394][5395]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3017
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:22
|
0.1 s
|
[5393]
|
|
|
2845
|
Delta: Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1
+details
Delta: Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:07
|
0.7 s
|
[5115]
|
|
|
3000
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:15
|
0.7 s
|
[5375]
|
|
|
2865
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:54)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:06:20
|
0.8 s
|
[5148][5149]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
2866
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:20
|
0.2 s
|
[5147]
|
|
|
2988
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:06
|
0.8 s
|
[5351]
|
|
|
2863
|
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.classic.Dataset.isEmpty(Dataset.scala:559)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:53)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 01:06:19
|
0.8 s
|
[5143][5144][5145][5146]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
2864
|
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000348 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = cef5d1bf-9c10-3645-a9ca-4bab0d7169c0 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = 141c3efe-47f0-32a8-981f-09028ec902c6 attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 01:06:19
|
0.2 s
|
[5142]
|
|
|
3467
|
Delta: Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1
+details
Delta: Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:45
|
0.8 s
|
[6235]
|
|
|
3202
|
Delta: Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1
+details
Delta: Delta: replenishmentRunId = 10000000349 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = 9ce6dde1-bee3-3e54-97ac-988588623651 workflowType = KpiPrepareDataSnapshotWorkflow workflowId = d6e60974-0b16-3a8e-9769-26b7657438e2 attempt = 1 cornerstoneTenantId = 8469 marketUnit = STEPHI_MU scenario = STANDARD: Filtering files for query: Compute snapshot for version: 1 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 20:07:49
|
0.8 s
|
[5817]
|
|
|
3487
|
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD
+details
replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.muListingCheck(ListingSnapshotEnhancementImpl.java:54)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ListingSnapshotEnhancementImpl.withListing(ListingSnapshotEnhancementImpl.java:34)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:59)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
jdk.internal.reflect.GeneratedMethodAccessor1659.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/21 01:07:58
|
0.9 s
|
[6268][6269]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3488
|
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000350 tenantId = 7233423560970044043 activityType = BufferDataSnapShot activityId = d11927ad-3153-358a-914c-ba06a15db40a workflowType = KpiPrepareDataSnapshotWorkflow workflowId = ae4415f9-75c3-3ec1-ab81-34528dde231c attempt = 1 cornerstoneTenantId = 8469 marketUnit = AUTO_ALL_ProdLoc scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/21 01:07:58
|
0.1 s
|
[6267]
|
|
|
3012
|
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD
+details
replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD org.apache.spark.sql.Dataset.first(Dataset.scala:2687)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ConfigurationSnapshotEnhancementImpl.fetchEnableLostSalesStockPeriod(ConfigurationSnapshotEnhancementImpl.java:41)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.ConfigurationSnapshotEnhancementImpl.withConfiguration(ConfigurationSnapshotEnhancementImpl.java:33)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.business.datasnapshot.DataSnapshotGenerationImpl.prepareDataSnapshot(DataSnapshotGenerationImpl.java:58)
com.sap.s4hana.eureka.business.crporderquantitykpiservice.core.controller.CreateDataSnapshotForShardActivityImpl.bufferDataSnapShot(CreateDataSnapshotForShardActivityImpl.java:44)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
java.base/java.lang.reflect.Method.invoke(Unknown Source)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor$POJOActivityInboundCallsInterceptor.executeActivity(RootActivityInboundCallsInterceptor.java:44)
io.temporal.internal.activity.RootActivityInboundCallsInterceptor.execute(RootActivityInboundCallsInterceptor.java:23)
io.temporal.internal.activity.ActivityTaskExecutors$BaseActivityTaskExecutor.execute(ActivityTaskExecutors.java:88)
io.temporal.internal.activity.ActivityTaskHandlerImpl.handle(ActivityTaskHandlerImpl.java:105)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handleActivity(ActivityWorker.java:294)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:258)
io.temporal.internal.worker.ActivityWorker$TaskHandlerImpl.handle(ActivityWorker.java:221)
io.temporal.internal.worker.PollTaskExecutor.lambda$process$1(PollTaskExecutor.java:76)
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
java.base/java.lang.Thread.run(Unknown Source)
|
2026/03/20 05:59:20
|
1.0 s
|
[5386][5387]
|
+details
|
|
|
| ID | Description | Submitted | Duration | Job IDs |
|
3013
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:21
|
0.3 s
|
[5385]
|
|
|
3007
|
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 8
+details
Delta: Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query: Compute snapshot for version: 8 org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1207)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:18
|
1 s
|
[5380][5381]
|
|
|
3002
|
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query
+details
Delta: replenishmentRunId = 10000000678 tenantId = 6951945722030153353 activityType = BufferDataSnapShot activityId = 372339e3-e78f-3202-9d96-da74ae3e269f workflowType = KpiPrepareDataSnapshotWorkflow workflowId = df824559-c1c8-3a48-92fe-1082cc951e9b attempt = 1 cornerstoneTenantId = 8468 marketUnit = IW_MU_CRP-125444_1 scenario = STANDARD: Filtering files for query org.apache.spark.sql.delta.metering.DeltaLogging.$anonfun$recordDeltaOperationInternal$1(DeltaLogging.scala:139)
com.databricks.spark.util.DatabricksLogging.recordOperation(DatabricksLogging.scala:128)
com.databricks.spark.util.DatabricksLogging.recordOperation$(DatabricksLogging.scala:117)
org.apache.spark.sql.delta.Snapshot.recordOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperationInternal(DeltaLogging.scala:138)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation(DeltaLogging.scala:128)
org.apache.spark.sql.delta.metering.DeltaLogging.recordDeltaOperation$(DeltaLogging.scala:118)
org.apache.spark.sql.delta.Snapshot.recordDeltaOperation(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan(DataSkippingReader.scala:1265)
org.apache.spark.sql.delta.stats.DataSkippingReaderBase.filesForScan$(DataSkippingReader.scala:1204)
org.apache.spark.sql.delta.Snapshot.filesForScan(Snapshot.scala:87)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.$anonfun$filesForScan$1(PrepareDeltaScan.scala:134)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withJobDescription(DeltaProgressReporter.scala:56)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode(DeltaProgressReporter.scala:35)
org.apache.spark.sql.delta.util.DeltaProgressReporter.withStatusCode$(DeltaProgressReporter.scala:29)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.withStatusCode(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan(PrepareDeltaScan.scala:119)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase.filesForScan$(PrepareDeltaScan.scala:114)
org.apache.spark.sql.delta.stats.PrepareDeltaScan.filesForScan(PrepareDeltaScan.scala:308)
org.apache.spark.sql.delta.stats.PrepareDeltaScanBase$$anonfun$prepareDeltaScan$1.$anonfun$applyOrElse$1(PrepareDeltaScan.scala:152)
|
2026/03/20 05:59:15
|
1 s
|
[5376]
|
|