digraph G {
0 [id="node0" labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>" tooltip="AdaptiveSparkPlan isFinalPlan=true"];
subgraph cluster1 {
isCluster="true";
id="cluster1";
label="WholeStageCodegen (4)";
tooltip="WholeStageCodegen (4)";
2 [id="node2" labelType="html" label="<br><b>SerializeFromObject</b><br><br>" tooltip="SerializeFromObject [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#946871, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#946872, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#946873, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#946874, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#946875, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#946876, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#946877, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#946878, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#946879, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#946880]"];
}
3 [id="node3" labelType="html" label="<br><b>MapPartitions</b><br><br>" tooltip="MapPartitions org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@764cc110, obj#946870: org.apache.spark.sql.delta.actions.SingleAction"];
4 [id="node4" labelType="html" label="<br><b>DeserializeToObject</b><br><br>" tooltip="DeserializeToObject newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#946869: org.apache.spark.sql.delta.actions.SingleAction"];
subgraph cluster5 {
isCluster="true";
id="cluster5";
label="WholeStageCodegen (3)";
tooltip="WholeStageCodegen (3)";
6 [id="node6" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#946781, CASE WHEN isnotnull(_extract_path#946892) THEN struct(path, add_path_canonical#946805, partitionValues, _extract_partitionValues#946893, size, _extract_size#946894L, modificationTime, _extract_modificationTime#946895L, dataChange, _extract_dataChange#946896, stats, add_stats_to_use#946803, tags, _extract_tags#946897, deletionVector, _extract_deletionVector#946898, baseRowId, _extract_baseRowId#946899L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#946900L, clusteringProvider, _extract_clusteringProvider#946901) END AS add#946825, CASE WHEN isnotnull(remove#946783.path) THEN if (isnull(remove#946783)) null else named_struct(path, remove_path_canonical#946811, deletionTimestamp, remove#946783.deletionTimestamp, dataChange, remove#946783.dataChange, extendedFileMetadata, remove#946783.extendedFileMetadata, partitionValues, remove#946783.partitionValues, size, remove#946783.size, tags, remove#946783.tags, deletionVector, remove#946783.deletionVector, baseRowId, remove#946783.baseRowId, defaultRowCommitVersion, remove#946783.defaultRowCommitVersion, stats, remove#946783.stats) END AS remove#946836, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790]"];
7 [id="node7" labelType="html" label="<br><b>Sort</b><br><br>" tooltip="Sort [version#946791L ASC NULLS FIRST], false, 0"];
}
8 [id="node8" labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 24<br>data size total (min, med, max (stageId: taskId))<br>21.9 KiB (3.3 KiB, 3.3 KiB, 8.5 KiB (stage 100833.0: task 324341))<br>shuffle write time total (min, med, max (stageId: taskId))<br>3 ms (0 ms, 0 ms, 0 ms (stage 100833.0: task 324345))<br>number of partitions: 50<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>11.0 KiB (1970.0 B, 1971.0 B, 3.3 KiB (stage 100833.0: task 324341))" tooltip="Exchange hashpartitioning(coalesce(add_path_canonical#946805, remove_path_canonical#946811), 50), REPARTITION_BY_NUM, [plan_id=1457992]"];
9 [id="node9" labelType="html" label="<br><b>Union</b><br><br>" tooltip="Union"];
subgraph cluster10 {
isCluster="true";
id="cluster10";
label="WholeStageCodegen (1)\n \nduration: 309 ms";
tooltip="WholeStageCodegen (1)";
11 [id="node11" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#946781, add#946782.path AS _extract_path#946892, add#946782.partitionValues AS _extract_partitionValues#946893, add#946782.size AS _extract_size#946894L, add#946782.modificationTime AS _extract_modificationTime#946895L, add#946782.dataChange AS _extract_dataChange#946896, add#946782.tags AS _extract_tags#946897, add#946782.deletionVector AS _extract_deletionVector#946898, add#946782.baseRowId AS _extract_baseRowId#946899L, add#946782.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946900L, add#946782.clusteringProvider AS _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add#946782.stats AS add_stats_to_use#946803, CASE WHEN isnotnull(add#946782.path) THEN UDF(add#946782.path) END AS add_path_canonical#946805, CASE WHEN isnotnull(remove#946783.path) THEN UDF(remove#946783.path) END AS remove_path_canonical#946811]"];
12 [id="node12" labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 12<br>number of input batches: 1" tooltip="ColumnarToRow"];
}
13 [id="node13" labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 179 ms<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 20.0 KiB<br>number of output rows: 12<br>number of partitions read: 1" tooltip="FileScan parquet [txn#946781,add#946782,remove#946783,metaData#946784,protocol#946785,cdc#946786,checkpointMetadata#946787,sidecar#946788,domainMetadata#946789,commitInfo#946790,version#946791L] Batched: true, DataFilters: [], Format: Parquet, Location: DeltaLogFileIndex(1 paths)[hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
subgraph cluster14 {
isCluster="true";
id="cluster14";
label="WholeStageCodegen (2)\n \nduration: total (min, med, max (stageId: taskId))\n832 ms (147 ms, 161 ms, 377 ms (stage 100833.0: task 324344))";
tooltip="WholeStageCodegen (2)";
15 [id="node15" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#946792, add#946793.path AS _extract_path#946902, add#946793.partitionValues AS _extract_partitionValues#946903, add#946793.size AS _extract_size#946904L, add#946793.modificationTime AS _extract_modificationTime#946905L, add#946793.dataChange AS _extract_dataChange#946906, add#946793.tags AS _extract_tags#946907, add#946793.deletionVector AS _extract_deletionVector#946908, add#946793.baseRowId AS _extract_baseRowId#946909L, add#946793.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946910L, add#946793.clusteringProvider AS _extract_clusteringProvider#946911, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L, add#946793.stats AS add_stats_to_use#946881, CASE WHEN isnotnull(add#946793.path) THEN UDF(add#946793.path) END AS add_path_canonical#946912, CASE WHEN isnotnull(remove#946794.path) THEN UDF(remove#946794.path) END AS remove_path_canonical#946913]"];
}
16 [id="node16" labelType="html" label="<b>Scan json </b><br><br>number of files read: 4<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 9.0 KiB<br>number of output rows: 12<br>number of partitions read: 4" tooltip="FileScan json [txn#946792,add#946793,remove#946794,metaData#946795,protocol#946796,cdc#946797,checkpointMetadata#946798,sidecar#946799,domainMetadata#946800,commitInfo#946801,version#946802L] Batched: false, DataFilters: [], Format: JSON, Location: DeltaLogFileIndex(4 paths)[hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
2->0;
3->2;
4->3;
6->4;
7->6;
8->7;
9->8;
11->9;
12->11;
13->12;
15->9;
16->15;
}
== Physical Plan ==
AdaptiveSparkPlan (24)
+- == Final Plan ==
ResultQueryStage (14), Statistics(sizeInBytes=8.0 EiB)
+- * SerializeFromObject (13)
+- MapPartitions (12)
+- DeserializeToObject (11)
+- * Project (10)
+- * Sort (9)
+- ShuffleQueryStage (8), Statistics(sizeInBytes=21.9 KiB, rowCount=24)
+- Exchange (7)
+- Union (6)
:- * Project (3)
: +- * ColumnarToRow (2)
: +- Scan parquet (1)
+- * Project (5)
+- Scan json (4)
+- == Initial Plan ==
SerializeFromObject (23)
+- MapPartitions (22)
+- DeserializeToObject (21)
+- Project (20)
+- Sort (19)
+- Exchange (18)
+- Union (17)
:- Project (15)
: +- Scan parquet (1)
+- Project (16)
+- Scan json (4)
(1) Scan parquet
Output [11]: [txn#946781, add#946782, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L]
Batched: true
Location: DeltaLogFileIndex [hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-inventory-inventorysnapshot-v2/_delta_log/00000000000000000010.checkpoint.parquet]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(2) ColumnarToRow [codegen id : 1]
Input [11]: [txn#946781, add#946782, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L]
(3) Project [codegen id : 1]
Output [23]: [txn#946781, add#946782.path AS _extract_path#946892, add#946782.partitionValues AS _extract_partitionValues#946893, add#946782.size AS _extract_size#946894L, add#946782.modificationTime AS _extract_modificationTime#946895L, add#946782.dataChange AS _extract_dataChange#946896, add#946782.tags AS _extract_tags#946897, add#946782.deletionVector AS _extract_deletionVector#946898, add#946782.baseRowId AS _extract_baseRowId#946899L, add#946782.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946900L, add#946782.clusteringProvider AS _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add#946782.stats AS add_stats_to_use#946803, CASE WHEN isnotnull(add#946782.path) THEN UDF(add#946782.path) END AS add_path_canonical#946805, CASE WHEN isnotnull(remove#946783.path) THEN UDF(remove#946783.path) END AS remove_path_canonical#946811]
Input [11]: [txn#946781, add#946782, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L]
(4) Scan json
Output [11]: [txn#946792, add#946793, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-inventory-inventorysnapshot-v2/_delta_log/00000000000000000011.json, ... 3 entries]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(5) Project [codegen id : 2]
Output [23]: [txn#946792, add#946793.path AS _extract_path#946902, add#946793.partitionValues AS _extract_partitionValues#946903, add#946793.size AS _extract_size#946904L, add#946793.modificationTime AS _extract_modificationTime#946905L, add#946793.dataChange AS _extract_dataChange#946906, add#946793.tags AS _extract_tags#946907, add#946793.deletionVector AS _extract_deletionVector#946908, add#946793.baseRowId AS _extract_baseRowId#946909L, add#946793.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946910L, add#946793.clusteringProvider AS _extract_clusteringProvider#946911, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L, add#946793.stats AS add_stats_to_use#946881, CASE WHEN isnotnull(add#946793.path) THEN UDF(add#946793.path) END AS add_path_canonical#946912, CASE WHEN isnotnull(remove#946794.path) THEN UDF(remove#946794.path) END AS remove_path_canonical#946913]
Input [11]: [txn#946792, add#946793, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L]
(6) Union
(7) Exchange
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
Arguments: hashpartitioning(coalesce(add_path_canonical#946805, remove_path_canonical#946811), 50), REPARTITION_BY_NUM, [plan_id=1457992]
(8) ShuffleQueryStage
Output [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
Arguments: 0
(9) Sort [codegen id : 3]
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
Arguments: [version#946791L ASC NULLS FIRST], false, 0
(10) Project [codegen id : 3]
Output [10]: [txn#946781, CASE WHEN isnotnull(_extract_path#946892) THEN struct(path, add_path_canonical#946805, partitionValues, _extract_partitionValues#946893, size, _extract_size#946894L, modificationTime, _extract_modificationTime#946895L, dataChange, _extract_dataChange#946896, stats, add_stats_to_use#946803, tags, _extract_tags#946897, deletionVector, _extract_deletionVector#946898, baseRowId, _extract_baseRowId#946899L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#946900L, clusteringProvider, _extract_clusteringProvider#946901) END AS add#946825, CASE WHEN isnotnull(remove#946783.path) THEN if (isnull(remove#946783)) null else named_struct(path, remove_path_canonical#946811, deletionTimestamp, remove#946783.deletionTimestamp, dataChange, remove#946783.dataChange, extendedFileMetadata, remove#946783.extendedFileMetadata, partitionValues, remove#946783.partitionValues, size, remove#946783.size, tags, remove#946783.tags, deletionVector, remove#946783.deletionVector, baseRowId, remove#946783.baseRowId, defaultRowCommitVersion, remove#946783.defaultRowCommitVersion, stats, remove#946783.stats) END AS remove#946836, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790]
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
(11) DeserializeToObject
Input [10]: [txn#946781, add#946825, remove#946836, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#946869: org.apache.spark.sql.delta.actions.SingleAction
(12) MapPartitions
Input [1]: [obj#946869]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@764cc110, obj#946870: org.apache.spark.sql.delta.actions.SingleAction
(13) SerializeFromObject [codegen id : 4]
Input [1]: [obj#946870]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#946871, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#946872, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#946873, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#946874, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#946875, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#946876, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#946877, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#946878, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#946879, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#946880]
(14) ResultQueryStage
Output [10]: [txn#946871, add#946872, remove#946873, metaData#946874, protocol#946875, cdc#946876, checkpointMetadata#946877, sidecar#946878, domainMetadata#946879, commitInfo#946880]
Arguments: 1
(15) Project
Output [23]: [txn#946781, add#946782.path AS _extract_path#946892, add#946782.partitionValues AS _extract_partitionValues#946893, add#946782.size AS _extract_size#946894L, add#946782.modificationTime AS _extract_modificationTime#946895L, add#946782.dataChange AS _extract_dataChange#946896, add#946782.tags AS _extract_tags#946897, add#946782.deletionVector AS _extract_deletionVector#946898, add#946782.baseRowId AS _extract_baseRowId#946899L, add#946782.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946900L, add#946782.clusteringProvider AS _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add#946782.stats AS add_stats_to_use#946803, CASE WHEN isnotnull(add#946782.path) THEN UDF(add#946782.path) END AS add_path_canonical#946805, CASE WHEN isnotnull(remove#946783.path) THEN UDF(remove#946783.path) END AS remove_path_canonical#946811]
Input [11]: [txn#946781, add#946782, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L]
(16) Project
Output [23]: [txn#946792, add#946793.path AS _extract_path#946902, add#946793.partitionValues AS _extract_partitionValues#946903, add#946793.size AS _extract_size#946904L, add#946793.modificationTime AS _extract_modificationTime#946905L, add#946793.dataChange AS _extract_dataChange#946906, add#946793.tags AS _extract_tags#946907, add#946793.deletionVector AS _extract_deletionVector#946908, add#946793.baseRowId AS _extract_baseRowId#946909L, add#946793.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#946910L, add#946793.clusteringProvider AS _extract_clusteringProvider#946911, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L, add#946793.stats AS add_stats_to_use#946881, CASE WHEN isnotnull(add#946793.path) THEN UDF(add#946793.path) END AS add_path_canonical#946912, CASE WHEN isnotnull(remove#946794.path) THEN UDF(remove#946794.path) END AS remove_path_canonical#946913]
Input [11]: [txn#946792, add#946793, remove#946794, metaData#946795, protocol#946796, cdc#946797, checkpointMetadata#946798, sidecar#946799, domainMetadata#946800, commitInfo#946801, version#946802L]
(17) Union
(18) Exchange
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
Arguments: hashpartitioning(coalesce(add_path_canonical#946805, remove_path_canonical#946811), 50), REPARTITION_BY_NUM, [plan_id=1457964]
(19) Sort
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
Arguments: [version#946791L ASC NULLS FIRST], false, 0
(20) Project
Output [10]: [txn#946781, CASE WHEN isnotnull(_extract_path#946892) THEN struct(path, add_path_canonical#946805, partitionValues, _extract_partitionValues#946893, size, _extract_size#946894L, modificationTime, _extract_modificationTime#946895L, dataChange, _extract_dataChange#946896, stats, add_stats_to_use#946803, tags, _extract_tags#946897, deletionVector, _extract_deletionVector#946898, baseRowId, _extract_baseRowId#946899L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#946900L, clusteringProvider, _extract_clusteringProvider#946901) END AS add#946825, CASE WHEN isnotnull(remove#946783.path) THEN if (isnull(remove#946783)) null else named_struct(path, remove_path_canonical#946811, deletionTimestamp, remove#946783.deletionTimestamp, dataChange, remove#946783.dataChange, extendedFileMetadata, remove#946783.extendedFileMetadata, partitionValues, remove#946783.partitionValues, size, remove#946783.size, tags, remove#946783.tags, deletionVector, remove#946783.deletionVector, baseRowId, remove#946783.baseRowId, defaultRowCommitVersion, remove#946783.defaultRowCommitVersion, stats, remove#946783.stats) END AS remove#946836, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790]
Input [23]: [txn#946781, _extract_path#946892, _extract_partitionValues#946893, _extract_size#946894L, _extract_modificationTime#946895L, _extract_dataChange#946896, _extract_tags#946897, _extract_deletionVector#946898, _extract_baseRowId#946899L, _extract_defaultRowCommitVersion#946900L, _extract_clusteringProvider#946901, remove#946783, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790, version#946791L, add_stats_to_use#946803, add_path_canonical#946805, remove_path_canonical#946811]
(21) DeserializeToObject
Input [10]: [txn#946781, add#946825, remove#946836, metaData#946784, protocol#946785, cdc#946786, checkpointMetadata#946787, sidecar#946788, domainMetadata#946789, commitInfo#946790]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#946869: org.apache.spark.sql.delta.actions.SingleAction
(22) MapPartitions
Input [1]: [obj#946869]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@764cc110, obj#946870: org.apache.spark.sql.delta.actions.SingleAction
(23) SerializeFromObject
Input [1]: [obj#946870]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#946871, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#946872, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#946873, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#946874, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#946875, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#946876, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#946877, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#946878, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#946879, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#946880]
(24) AdaptiveSparkPlan
Output [10]: [txn#946871, add#946872, remove#946873, metaData#946874, protocol#946875, cdc#946876, checkpointMetadata#946877, sidecar#946878, domainMetadata#946879, commitInfo#946880]
Arguments: isFinalPlan=true