digraph G {
0 [id="node0" labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>" tooltip="AdaptiveSparkPlan isFinalPlan=true"];
subgraph cluster1 {
isCluster="true";
id="cluster1";
label="WholeStageCodegen (4)";
tooltip="WholeStageCodegen (4)";
2 [id="node2" labelType="html" label="<br><b>SerializeFromObject</b><br><br>" tooltip="SerializeFromObject [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1104442, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1104443, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1104444, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1104445, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1104446, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1104447, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1104448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1104449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1104450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1104451]"];
}
3 [id="node3" labelType="html" label="<br><b>MapPartitions</b><br><br>" tooltip="MapPartitions org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@7825081d, obj#1104441: org.apache.spark.sql.delta.actions.SingleAction"];
4 [id="node4" labelType="html" label="<br><b>DeserializeToObject</b><br><br>" tooltip="DeserializeToObject newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1104440: org.apache.spark.sql.delta.actions.SingleAction"];
subgraph cluster5 {
isCluster="true";
id="cluster5";
label="WholeStageCodegen (3)";
tooltip="WholeStageCodegen (3)";
6 [id="node6" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1104352, CASE WHEN isnotnull(_extract_path#1104463) THEN struct(path, add_path_canonical#1104376, partitionValues, _extract_partitionValues#1104464, size, _extract_size#1104465L, modificationTime, _extract_modificationTime#1104466L, dataChange, _extract_dataChange#1104467, stats, add_stats_to_use#1104374, tags, _extract_tags#1104468, deletionVector, _extract_deletionVector#1104469, baseRowId, _extract_baseRowId#1104470L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1104471L, clusteringProvider, _extract_clusteringProvider#1104472) END AS add#1104396, CASE WHEN isnotnull(remove#1104354.path) THEN if (isnull(remove#1104354)) null else named_struct(path, remove_path_canonical#1104382, deletionTimestamp, remove#1104354.deletionTimestamp, dataChange, remove#1104354.dataChange, extendedFileMetadata, remove#1104354.extendedFileMetadata, partitionValues, remove#1104354.partitionValues, size, remove#1104354.size, tags, remove#1104354.tags, deletionVector, remove#1104354.deletionVector, baseRowId, remove#1104354.baseRowId, defaultRowCommitVersion, remove#1104354.defaultRowCommitVersion, stats, remove#1104354.stats) END AS remove#1104407, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361]"];
7 [id="node7" labelType="html" label="<br><b>Sort</b><br><br>" tooltip="Sort [version#1104362L ASC NULLS FIRST], false, 0"];
}
8 [id="node8" labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 100<br>data size total (min, med, max (stageId: taskId))<br>220.1 KiB (2.2 KiB, 4.5 KiB, 193.4 KiB (stage 115263.0: task 378982))<br>shuffle write time total (min, med, max (stageId: taskId))<br>7 ms (0 ms, 0 ms, 1 ms (stage 115263.0: task 378982))<br>number of partitions: 50<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>59.3 KiB (1240.0 B, 2002.0 B, 46.6 KiB (stage 115263.0: task 378982))" tooltip="Exchange hashpartitioning(coalesce(add_path_canonical#1104376, remove_path_canonical#1104382), 50), REPARTITION_BY_NUM, [plan_id=1700911]"];
9 [id="node9" labelType="html" label="<br><b>Union</b><br><br>" tooltip="Union"];
subgraph cluster10 {
isCluster="true";
id="cluster10";
label="WholeStageCodegen (1)\n \nduration: 156 ms";
tooltip="WholeStageCodegen (1)";
11 [id="node11" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1104352, add#1104353.path AS _extract_path#1104463, add#1104353.partitionValues AS _extract_partitionValues#1104464, add#1104353.size AS _extract_size#1104465L, add#1104353.modificationTime AS _extract_modificationTime#1104466L, add#1104353.dataChange AS _extract_dataChange#1104467, add#1104353.tags AS _extract_tags#1104468, add#1104353.deletionVector AS _extract_deletionVector#1104469, add#1104353.baseRowId AS _extract_baseRowId#1104470L, add#1104353.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104471L, add#1104353.clusteringProvider AS _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add#1104353.stats AS add_stats_to_use#1104374, CASE WHEN isnotnull(add#1104353.path) THEN UDF(add#1104353.path) END AS add_path_canonical#1104376, CASE WHEN isnotnull(remove#1104354.path) THEN UDF(remove#1104354.path) END AS remove_path_canonical#1104382]"];
12 [id="node12" labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 84<br>number of input batches: 1" tooltip="ColumnarToRow"];
}
13 [id="node13" labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 136 ms<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 32.5 KiB<br>number of output rows: 84<br>number of partitions read: 1" tooltip="FileScan parquet [txn#1104352,add#1104353,remove#1104354,metaData#1104355,protocol#1104356,cdc#1104357,checkpointMetadata#1104358,sidecar#1104359,domainMetadata#1104360,commitInfo#1104361,version#1104362L] Batched: true, DataFilters: [], Format: Parquet, Location: DeltaLogFileIndex(1 paths)[hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
subgraph cluster14 {
isCluster="true";
id="cluster14";
label="WholeStageCodegen (2)\n \nduration: total (min, med, max (stageId: taskId))\n1.8 s (143 ms, 153 ms, 339 ms (stage 115263.0: task 378990))";
tooltip="WholeStageCodegen (2)";
15 [id="node15" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1104363, add#1104364.path AS _extract_path#1104473, add#1104364.partitionValues AS _extract_partitionValues#1104474, add#1104364.size AS _extract_size#1104475L, add#1104364.modificationTime AS _extract_modificationTime#1104476L, add#1104364.dataChange AS _extract_dataChange#1104477, add#1104364.tags AS _extract_tags#1104478, add#1104364.deletionVector AS _extract_deletionVector#1104479, add#1104364.baseRowId AS _extract_baseRowId#1104480L, add#1104364.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104481L, add#1104364.clusteringProvider AS _extract_clusteringProvider#1104482, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L, add#1104364.stats AS add_stats_to_use#1104452, CASE WHEN isnotnull(add#1104364.path) THEN UDF(add#1104364.path) END AS add_path_canonical#1104483, CASE WHEN isnotnull(remove#1104365.path) THEN UDF(remove#1104365.path) END AS remove_path_canonical#1104484]"];
}
16 [id="node16" labelType="html" label="<b>Scan json </b><br><br>number of files read: 8<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 20.9 KiB<br>number of output rows: 16<br>number of partitions read: 8" tooltip="FileScan json [txn#1104363,add#1104364,remove#1104365,metaData#1104366,protocol#1104367,cdc#1104368,checkpointMetadata#1104369,sidecar#1104370,domainMetadata#1104371,commitInfo#1104372,version#1104373L] Batched: false, DataFilters: [], Format: JSON, Location: DeltaLogFileIndex(8 paths)[hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
2->0;
3->2;
4->3;
6->4;
7->6;
8->7;
9->8;
11->9;
12->11;
13->12;
15->9;
16->15;
}
== Physical Plan ==
AdaptiveSparkPlan (24)
+- == Final Plan ==
ResultQueryStage (14), Statistics(sizeInBytes=8.0 EiB)
+- * SerializeFromObject (13)
+- MapPartitions (12)
+- DeserializeToObject (11)
+- * Project (10)
+- * Sort (9)
+- ShuffleQueryStage (8), Statistics(sizeInBytes=220.1 KiB, rowCount=100)
+- Exchange (7)
+- Union (6)
:- * Project (3)
: +- * ColumnarToRow (2)
: +- Scan parquet (1)
+- * Project (5)
+- Scan json (4)
+- == Initial Plan ==
SerializeFromObject (23)
+- MapPartitions (22)
+- DeserializeToObject (21)
+- Project (20)
+- Sort (19)
+- Exchange (18)
+- Union (17)
:- Project (15)
: +- Scan parquet (1)
+- Project (16)
+- Scan json (4)
(1) Scan parquet
Output [11]: [txn#1104352, add#1104353, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L]
Batched: true
Location: DeltaLogFileIndex [hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/market-unit/_delta_log/00000000000000000300.checkpoint.parquet]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(2) ColumnarToRow [codegen id : 1]
Input [11]: [txn#1104352, add#1104353, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L]
(3) Project [codegen id : 1]
Output [23]: [txn#1104352, add#1104353.path AS _extract_path#1104463, add#1104353.partitionValues AS _extract_partitionValues#1104464, add#1104353.size AS _extract_size#1104465L, add#1104353.modificationTime AS _extract_modificationTime#1104466L, add#1104353.dataChange AS _extract_dataChange#1104467, add#1104353.tags AS _extract_tags#1104468, add#1104353.deletionVector AS _extract_deletionVector#1104469, add#1104353.baseRowId AS _extract_baseRowId#1104470L, add#1104353.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104471L, add#1104353.clusteringProvider AS _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add#1104353.stats AS add_stats_to_use#1104374, CASE WHEN isnotnull(add#1104353.path) THEN UDF(add#1104353.path) END AS add_path_canonical#1104376, CASE WHEN isnotnull(remove#1104354.path) THEN UDF(remove#1104354.path) END AS remove_path_canonical#1104382]
Input [11]: [txn#1104352, add#1104353, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L]
(4) Scan json
Output [11]: [txn#1104363, add#1104364, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/market-unit/_delta_log/00000000000000000301.json, ... 7 entries]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(5) Project [codegen id : 2]
Output [23]: [txn#1104363, add#1104364.path AS _extract_path#1104473, add#1104364.partitionValues AS _extract_partitionValues#1104474, add#1104364.size AS _extract_size#1104475L, add#1104364.modificationTime AS _extract_modificationTime#1104476L, add#1104364.dataChange AS _extract_dataChange#1104477, add#1104364.tags AS _extract_tags#1104478, add#1104364.deletionVector AS _extract_deletionVector#1104479, add#1104364.baseRowId AS _extract_baseRowId#1104480L, add#1104364.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104481L, add#1104364.clusteringProvider AS _extract_clusteringProvider#1104482, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L, add#1104364.stats AS add_stats_to_use#1104452, CASE WHEN isnotnull(add#1104364.path) THEN UDF(add#1104364.path) END AS add_path_canonical#1104483, CASE WHEN isnotnull(remove#1104365.path) THEN UDF(remove#1104365.path) END AS remove_path_canonical#1104484]
Input [11]: [txn#1104363, add#1104364, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L]
(6) Union
(7) Exchange
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
Arguments: hashpartitioning(coalesce(add_path_canonical#1104376, remove_path_canonical#1104382), 50), REPARTITION_BY_NUM, [plan_id=1700911]
(8) ShuffleQueryStage
Output [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
Arguments: 0
(9) Sort [codegen id : 3]
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
Arguments: [version#1104362L ASC NULLS FIRST], false, 0
(10) Project [codegen id : 3]
Output [10]: [txn#1104352, CASE WHEN isnotnull(_extract_path#1104463) THEN struct(path, add_path_canonical#1104376, partitionValues, _extract_partitionValues#1104464, size, _extract_size#1104465L, modificationTime, _extract_modificationTime#1104466L, dataChange, _extract_dataChange#1104467, stats, add_stats_to_use#1104374, tags, _extract_tags#1104468, deletionVector, _extract_deletionVector#1104469, baseRowId, _extract_baseRowId#1104470L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1104471L, clusteringProvider, _extract_clusteringProvider#1104472) END AS add#1104396, CASE WHEN isnotnull(remove#1104354.path) THEN if (isnull(remove#1104354)) null else named_struct(path, remove_path_canonical#1104382, deletionTimestamp, remove#1104354.deletionTimestamp, dataChange, remove#1104354.dataChange, extendedFileMetadata, remove#1104354.extendedFileMetadata, partitionValues, remove#1104354.partitionValues, size, remove#1104354.size, tags, remove#1104354.tags, deletionVector, remove#1104354.deletionVector, baseRowId, remove#1104354.baseRowId, defaultRowCommitVersion, remove#1104354.defaultRowCommitVersion, stats, remove#1104354.stats) END AS remove#1104407, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361]
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
(11) DeserializeToObject
Input [10]: [txn#1104352, add#1104396, remove#1104407, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1104440: org.apache.spark.sql.delta.actions.SingleAction
(12) MapPartitions
Input [1]: [obj#1104440]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@7825081d, obj#1104441: org.apache.spark.sql.delta.actions.SingleAction
(13) SerializeFromObject [codegen id : 4]
Input [1]: [obj#1104441]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1104442, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1104443, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1104444, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1104445, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1104446, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1104447, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1104448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1104449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1104450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1104451]
(14) ResultQueryStage
Output [10]: [txn#1104442, add#1104443, remove#1104444, metaData#1104445, protocol#1104446, cdc#1104447, checkpointMetadata#1104448, sidecar#1104449, domainMetadata#1104450, commitInfo#1104451]
Arguments: 1
(15) Project
Output [23]: [txn#1104352, add#1104353.path AS _extract_path#1104463, add#1104353.partitionValues AS _extract_partitionValues#1104464, add#1104353.size AS _extract_size#1104465L, add#1104353.modificationTime AS _extract_modificationTime#1104466L, add#1104353.dataChange AS _extract_dataChange#1104467, add#1104353.tags AS _extract_tags#1104468, add#1104353.deletionVector AS _extract_deletionVector#1104469, add#1104353.baseRowId AS _extract_baseRowId#1104470L, add#1104353.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104471L, add#1104353.clusteringProvider AS _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add#1104353.stats AS add_stats_to_use#1104374, CASE WHEN isnotnull(add#1104353.path) THEN UDF(add#1104353.path) END AS add_path_canonical#1104376, CASE WHEN isnotnull(remove#1104354.path) THEN UDF(remove#1104354.path) END AS remove_path_canonical#1104382]
Input [11]: [txn#1104352, add#1104353, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L]
(16) Project
Output [23]: [txn#1104363, add#1104364.path AS _extract_path#1104473, add#1104364.partitionValues AS _extract_partitionValues#1104474, add#1104364.size AS _extract_size#1104475L, add#1104364.modificationTime AS _extract_modificationTime#1104476L, add#1104364.dataChange AS _extract_dataChange#1104477, add#1104364.tags AS _extract_tags#1104478, add#1104364.deletionVector AS _extract_deletionVector#1104479, add#1104364.baseRowId AS _extract_baseRowId#1104480L, add#1104364.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1104481L, add#1104364.clusteringProvider AS _extract_clusteringProvider#1104482, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L, add#1104364.stats AS add_stats_to_use#1104452, CASE WHEN isnotnull(add#1104364.path) THEN UDF(add#1104364.path) END AS add_path_canonical#1104483, CASE WHEN isnotnull(remove#1104365.path) THEN UDF(remove#1104365.path) END AS remove_path_canonical#1104484]
Input [11]: [txn#1104363, add#1104364, remove#1104365, metaData#1104366, protocol#1104367, cdc#1104368, checkpointMetadata#1104369, sidecar#1104370, domainMetadata#1104371, commitInfo#1104372, version#1104373L]
(17) Union
(18) Exchange
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
Arguments: hashpartitioning(coalesce(add_path_canonical#1104376, remove_path_canonical#1104382), 50), REPARTITION_BY_NUM, [plan_id=1700883]
(19) Sort
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
Arguments: [version#1104362L ASC NULLS FIRST], false, 0
(20) Project
Output [10]: [txn#1104352, CASE WHEN isnotnull(_extract_path#1104463) THEN struct(path, add_path_canonical#1104376, partitionValues, _extract_partitionValues#1104464, size, _extract_size#1104465L, modificationTime, _extract_modificationTime#1104466L, dataChange, _extract_dataChange#1104467, stats, add_stats_to_use#1104374, tags, _extract_tags#1104468, deletionVector, _extract_deletionVector#1104469, baseRowId, _extract_baseRowId#1104470L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1104471L, clusteringProvider, _extract_clusteringProvider#1104472) END AS add#1104396, CASE WHEN isnotnull(remove#1104354.path) THEN if (isnull(remove#1104354)) null else named_struct(path, remove_path_canonical#1104382, deletionTimestamp, remove#1104354.deletionTimestamp, dataChange, remove#1104354.dataChange, extendedFileMetadata, remove#1104354.extendedFileMetadata, partitionValues, remove#1104354.partitionValues, size, remove#1104354.size, tags, remove#1104354.tags, deletionVector, remove#1104354.deletionVector, baseRowId, remove#1104354.baseRowId, defaultRowCommitVersion, remove#1104354.defaultRowCommitVersion, stats, remove#1104354.stats) END AS remove#1104407, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361]
Input [23]: [txn#1104352, _extract_path#1104463, _extract_partitionValues#1104464, _extract_size#1104465L, _extract_modificationTime#1104466L, _extract_dataChange#1104467, _extract_tags#1104468, _extract_deletionVector#1104469, _extract_baseRowId#1104470L, _extract_defaultRowCommitVersion#1104471L, _extract_clusteringProvider#1104472, remove#1104354, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361, version#1104362L, add_stats_to_use#1104374, add_path_canonical#1104376, remove_path_canonical#1104382]
(21) DeserializeToObject
Input [10]: [txn#1104352, add#1104396, remove#1104407, metaData#1104355, protocol#1104356, cdc#1104357, checkpointMetadata#1104358, sidecar#1104359, domainMetadata#1104360, commitInfo#1104361]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1104440: org.apache.spark.sql.delta.actions.SingleAction
(22) MapPartitions
Input [1]: [obj#1104440]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@7825081d, obj#1104441: org.apache.spark.sql.delta.actions.SingleAction
(23) SerializeFromObject
Input [1]: [obj#1104441]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1104442, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1104443, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1104444, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1104445, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1104446, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1104447, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1104448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1104449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1104450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1104451]
(24) AdaptiveSparkPlan
Output [10]: [txn#1104442, add#1104443, remove#1104444, metaData#1104445, protocol#1104446, cdc#1104447, checkpointMetadata#1104448, sidecar#1104449, domainMetadata#1104450, commitInfo#1104451]
Arguments: isFinalPlan=true