digraph G {
0 [id="node0" labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>" tooltip="AdaptiveSparkPlan isFinalPlan=true"];
subgraph cluster1 {
isCluster="true";
id="cluster1";
label="WholeStageCodegen (4)";
tooltip="WholeStageCodegen (4)";
2 [id="node2" labelType="html" label="<br><b>SerializeFromObject</b><br><br>" tooltip="SerializeFromObject [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#947128, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#947129, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#947130, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#947131, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#947132, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#947133, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#947134, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#947135, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#947136, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#947137]"];
}
3 [id="node3" labelType="html" label="<br><b>MapPartitions</b><br><br>" tooltip="MapPartitions org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@2d539f81, obj#947127: org.apache.spark.sql.delta.actions.SingleAction"];
4 [id="node4" labelType="html" label="<br><b>DeserializeToObject</b><br><br>" tooltip="DeserializeToObject newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#947126: org.apache.spark.sql.delta.actions.SingleAction"];
subgraph cluster5 {
isCluster="true";
id="cluster5";
label="WholeStageCodegen (3)";
tooltip="WholeStageCodegen (3)";
6 [id="node6" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#947038, CASE WHEN isnotnull(_extract_path#947149) THEN struct(path, add_path_canonical#947062, partitionValues, _extract_partitionValues#947150, size, _extract_size#947151L, modificationTime, _extract_modificationTime#947152L, dataChange, _extract_dataChange#947153, stats, add_stats_to_use#947060, tags, _extract_tags#947154, deletionVector, _extract_deletionVector#947155, baseRowId, _extract_baseRowId#947156L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#947157L, clusteringProvider, _extract_clusteringProvider#947158) END AS add#947082, CASE WHEN isnotnull(remove#947040.path) THEN if (isnull(remove#947040)) null else named_struct(path, remove_path_canonical#947068, deletionTimestamp, remove#947040.deletionTimestamp, dataChange, remove#947040.dataChange, extendedFileMetadata, remove#947040.extendedFileMetadata, partitionValues, remove#947040.partitionValues, size, remove#947040.size, tags, remove#947040.tags, deletionVector, remove#947040.deletionVector, baseRowId, remove#947040.baseRowId, defaultRowCommitVersion, remove#947040.defaultRowCommitVersion, stats, remove#947040.stats) END AS remove#947093, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047]"];
7 [id="node7" labelType="html" label="<br><b>Sort</b><br><br>" tooltip="Sort [version#947048L ASC NULLS FIRST], false, 0"];
}
8 [id="node8" labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 50<br>data size total (min, med, max (stageId: taskId))<br>117.3 KiB (5.0 KiB, 5.0 KiB, 97.4 KiB (stage 100836.0: task 324396))<br>shuffle write time total (min, med, max (stageId: taskId))<br>5 ms (0 ms, 1 ms, 1 ms (stage 100836.0: task 324397))<br>number of partitions: 50<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>34.3 KiB (2.2 KiB, 2.3 KiB, 25.1 KiB (stage 100836.0: task 324396))" tooltip="Exchange hashpartitioning(coalesce(add_path_canonical#947062, remove_path_canonical#947068), 50), REPARTITION_BY_NUM, [plan_id=1458123]"];
9 [id="node9" labelType="html" label="<br><b>Union</b><br><br>" tooltip="Union"];
subgraph cluster10 {
isCluster="true";
id="cluster10";
label="WholeStageCodegen (1)\n \nduration: 94 ms";
tooltip="WholeStageCodegen (1)";
11 [id="node11" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#947038, add#947039.path AS _extract_path#947149, add#947039.partitionValues AS _extract_partitionValues#947150, add#947039.size AS _extract_size#947151L, add#947039.modificationTime AS _extract_modificationTime#947152L, add#947039.dataChange AS _extract_dataChange#947153, add#947039.tags AS _extract_tags#947154, add#947039.deletionVector AS _extract_deletionVector#947155, add#947039.baseRowId AS _extract_baseRowId#947156L, add#947039.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947157L, add#947039.clusteringProvider AS _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add#947039.stats AS add_stats_to_use#947060, CASE WHEN isnotnull(add#947039.path) THEN UDF(add#947039.path) END AS add_path_canonical#947062, CASE WHEN isnotnull(remove#947040.path) THEN UDF(remove#947040.path) END AS remove_path_canonical#947068]"];
12 [id="node12" labelType="html" label="<b>ColumnarToRow</b><br><br>number of output rows: 38<br>number of input batches: 1" tooltip="ColumnarToRow"];
}
13 [id="node13" labelType="html" label="<b>Scan parquet </b><br><br>number of files read: 1<br>scan time: 90 ms<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 24.2 KiB<br>number of output rows: 38<br>number of partitions read: 1" tooltip="FileScan parquet [txn#947038,add#947039,remove#947040,metaData#947041,protocol#947042,cdc#947043,checkpointMetadata#947044,sidecar#947045,domainMetadata#947046,commitInfo#947047,version#947048L] Batched: true, DataFilters: [], Format: Parquet, Location: DeltaLogFileIndex(1 paths)[hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
subgraph cluster14 {
isCluster="true";
id="cluster14";
label="WholeStageCodegen (2)\n \nduration: total (min, med, max (stageId: taskId))\n159 ms (36 ms, 38 ms, 47 ms (stage 100836.0: task 324397))";
tooltip="WholeStageCodegen (2)";
15 [id="node15" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#947049, add#947050.path AS _extract_path#947159, add#947050.partitionValues AS _extract_partitionValues#947160, add#947050.size AS _extract_size#947161L, add#947050.modificationTime AS _extract_modificationTime#947162L, add#947050.dataChange AS _extract_dataChange#947163, add#947050.tags AS _extract_tags#947164, add#947050.deletionVector AS _extract_deletionVector#947165, add#947050.baseRowId AS _extract_baseRowId#947166L, add#947050.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947167L, add#947050.clusteringProvider AS _extract_clusteringProvider#947168, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L, add#947050.stats AS add_stats_to_use#947138, CASE WHEN isnotnull(add#947050.path) THEN UDF(add#947050.path) END AS add_path_canonical#947169, CASE WHEN isnotnull(remove#947051.path) THEN UDF(remove#947051.path) END AS remove_path_canonical#947170]"];
}
16 [id="node16" labelType="html" label="<b>Scan json </b><br><br>number of files read: 4<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 16.6 KiB<br>number of output rows: 12<br>number of partitions read: 4" tooltip="FileScan json [txn#947049,add#947050,remove#947051,metaData#947052,protocol#947053,cdc#947054,checkpointMetadata#947055,sidecar#947056,domainMetadata#947057,commitInfo#947058,version#947059L] Batched: false, DataFilters: [], Format: JSON, Location: DeltaLogFileIndex(4 paths)[hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
2->0;
3->2;
4->3;
6->4;
7->6;
8->7;
9->8;
11->9;
12->11;
13->12;
15->9;
16->15;
}
== Physical Plan ==
AdaptiveSparkPlan (24)
+- == Final Plan ==
ResultQueryStage (14), Statistics(sizeInBytes=8.0 EiB)
+- * SerializeFromObject (13)
+- MapPartitions (12)
+- DeserializeToObject (11)
+- * Project (10)
+- * Sort (9)
+- ShuffleQueryStage (8), Statistics(sizeInBytes=117.3 KiB, rowCount=50)
+- Exchange (7)
+- Union (6)
:- * Project (3)
: +- * ColumnarToRow (2)
: +- Scan parquet (1)
+- * Project (5)
+- Scan json (4)
+- == Initial Plan ==
SerializeFromObject (23)
+- MapPartitions (22)
+- DeserializeToObject (21)
+- Project (20)
+- Sort (19)
+- Exchange (18)
+- Union (17)
:- Project (15)
: +- Scan parquet (1)
+- Project (16)
+- Scan json (4)
(1) Scan parquet
Output [11]: [txn#947038, add#947039, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L]
Batched: true
Location: DeltaLogFileIndex [hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/market-unit/_delta_log/00000000000000000110.checkpoint.parquet]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(2) ColumnarToRow [codegen id : 1]
Input [11]: [txn#947038, add#947039, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L]
(3) Project [codegen id : 1]
Output [23]: [txn#947038, add#947039.path AS _extract_path#947149, add#947039.partitionValues AS _extract_partitionValues#947150, add#947039.size AS _extract_size#947151L, add#947039.modificationTime AS _extract_modificationTime#947152L, add#947039.dataChange AS _extract_dataChange#947153, add#947039.tags AS _extract_tags#947154, add#947039.deletionVector AS _extract_deletionVector#947155, add#947039.baseRowId AS _extract_baseRowId#947156L, add#947039.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947157L, add#947039.clusteringProvider AS _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add#947039.stats AS add_stats_to_use#947060, CASE WHEN isnotnull(add#947039.path) THEN UDF(add#947039.path) END AS add_path_canonical#947062, CASE WHEN isnotnull(remove#947040.path) THEN UDF(remove#947040.path) END AS remove_path_canonical#947068]
Input [11]: [txn#947038, add#947039, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L]
(4) Scan json
Output [11]: [txn#947049, add#947050, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://770f0770-bd0c-456f-b83b-9649926e0857.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/prp/market-unit/_delta_log/00000000000000000111.json, ... 3 entries]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(5) Project [codegen id : 2]
Output [23]: [txn#947049, add#947050.path AS _extract_path#947159, add#947050.partitionValues AS _extract_partitionValues#947160, add#947050.size AS _extract_size#947161L, add#947050.modificationTime AS _extract_modificationTime#947162L, add#947050.dataChange AS _extract_dataChange#947163, add#947050.tags AS _extract_tags#947164, add#947050.deletionVector AS _extract_deletionVector#947165, add#947050.baseRowId AS _extract_baseRowId#947166L, add#947050.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947167L, add#947050.clusteringProvider AS _extract_clusteringProvider#947168, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L, add#947050.stats AS add_stats_to_use#947138, CASE WHEN isnotnull(add#947050.path) THEN UDF(add#947050.path) END AS add_path_canonical#947169, CASE WHEN isnotnull(remove#947051.path) THEN UDF(remove#947051.path) END AS remove_path_canonical#947170]
Input [11]: [txn#947049, add#947050, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L]
(6) Union
(7) Exchange
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
Arguments: hashpartitioning(coalesce(add_path_canonical#947062, remove_path_canonical#947068), 50), REPARTITION_BY_NUM, [plan_id=1458123]
(8) ShuffleQueryStage
Output [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
Arguments: 0
(9) Sort [codegen id : 3]
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
Arguments: [version#947048L ASC NULLS FIRST], false, 0
(10) Project [codegen id : 3]
Output [10]: [txn#947038, CASE WHEN isnotnull(_extract_path#947149) THEN struct(path, add_path_canonical#947062, partitionValues, _extract_partitionValues#947150, size, _extract_size#947151L, modificationTime, _extract_modificationTime#947152L, dataChange, _extract_dataChange#947153, stats, add_stats_to_use#947060, tags, _extract_tags#947154, deletionVector, _extract_deletionVector#947155, baseRowId, _extract_baseRowId#947156L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#947157L, clusteringProvider, _extract_clusteringProvider#947158) END AS add#947082, CASE WHEN isnotnull(remove#947040.path) THEN if (isnull(remove#947040)) null else named_struct(path, remove_path_canonical#947068, deletionTimestamp, remove#947040.deletionTimestamp, dataChange, remove#947040.dataChange, extendedFileMetadata, remove#947040.extendedFileMetadata, partitionValues, remove#947040.partitionValues, size, remove#947040.size, tags, remove#947040.tags, deletionVector, remove#947040.deletionVector, baseRowId, remove#947040.baseRowId, defaultRowCommitVersion, remove#947040.defaultRowCommitVersion, stats, remove#947040.stats) END AS remove#947093, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047]
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
(11) DeserializeToObject
Input [10]: [txn#947038, add#947082, remove#947093, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#947126: org.apache.spark.sql.delta.actions.SingleAction
(12) MapPartitions
Input [1]: [obj#947126]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@2d539f81, obj#947127: org.apache.spark.sql.delta.actions.SingleAction
(13) SerializeFromObject [codegen id : 4]
Input [1]: [obj#947127]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#947128, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#947129, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#947130, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#947131, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#947132, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#947133, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#947134, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#947135, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#947136, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#947137]
(14) ResultQueryStage
Output [10]: [txn#947128, add#947129, remove#947130, metaData#947131, protocol#947132, cdc#947133, checkpointMetadata#947134, sidecar#947135, domainMetadata#947136, commitInfo#947137]
Arguments: 1
(15) Project
Output [23]: [txn#947038, add#947039.path AS _extract_path#947149, add#947039.partitionValues AS _extract_partitionValues#947150, add#947039.size AS _extract_size#947151L, add#947039.modificationTime AS _extract_modificationTime#947152L, add#947039.dataChange AS _extract_dataChange#947153, add#947039.tags AS _extract_tags#947154, add#947039.deletionVector AS _extract_deletionVector#947155, add#947039.baseRowId AS _extract_baseRowId#947156L, add#947039.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947157L, add#947039.clusteringProvider AS _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add#947039.stats AS add_stats_to_use#947060, CASE WHEN isnotnull(add#947039.path) THEN UDF(add#947039.path) END AS add_path_canonical#947062, CASE WHEN isnotnull(remove#947040.path) THEN UDF(remove#947040.path) END AS remove_path_canonical#947068]
Input [11]: [txn#947038, add#947039, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L]
(16) Project
Output [23]: [txn#947049, add#947050.path AS _extract_path#947159, add#947050.partitionValues AS _extract_partitionValues#947160, add#947050.size AS _extract_size#947161L, add#947050.modificationTime AS _extract_modificationTime#947162L, add#947050.dataChange AS _extract_dataChange#947163, add#947050.tags AS _extract_tags#947164, add#947050.deletionVector AS _extract_deletionVector#947165, add#947050.baseRowId AS _extract_baseRowId#947166L, add#947050.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#947167L, add#947050.clusteringProvider AS _extract_clusteringProvider#947168, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L, add#947050.stats AS add_stats_to_use#947138, CASE WHEN isnotnull(add#947050.path) THEN UDF(add#947050.path) END AS add_path_canonical#947169, CASE WHEN isnotnull(remove#947051.path) THEN UDF(remove#947051.path) END AS remove_path_canonical#947170]
Input [11]: [txn#947049, add#947050, remove#947051, metaData#947052, protocol#947053, cdc#947054, checkpointMetadata#947055, sidecar#947056, domainMetadata#947057, commitInfo#947058, version#947059L]
(17) Union
(18) Exchange
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
Arguments: hashpartitioning(coalesce(add_path_canonical#947062, remove_path_canonical#947068), 50), REPARTITION_BY_NUM, [plan_id=1458095]
(19) Sort
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
Arguments: [version#947048L ASC NULLS FIRST], false, 0
(20) Project
Output [10]: [txn#947038, CASE WHEN isnotnull(_extract_path#947149) THEN struct(path, add_path_canonical#947062, partitionValues, _extract_partitionValues#947150, size, _extract_size#947151L, modificationTime, _extract_modificationTime#947152L, dataChange, _extract_dataChange#947153, stats, add_stats_to_use#947060, tags, _extract_tags#947154, deletionVector, _extract_deletionVector#947155, baseRowId, _extract_baseRowId#947156L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#947157L, clusteringProvider, _extract_clusteringProvider#947158) END AS add#947082, CASE WHEN isnotnull(remove#947040.path) THEN if (isnull(remove#947040)) null else named_struct(path, remove_path_canonical#947068, deletionTimestamp, remove#947040.deletionTimestamp, dataChange, remove#947040.dataChange, extendedFileMetadata, remove#947040.extendedFileMetadata, partitionValues, remove#947040.partitionValues, size, remove#947040.size, tags, remove#947040.tags, deletionVector, remove#947040.deletionVector, baseRowId, remove#947040.baseRowId, defaultRowCommitVersion, remove#947040.defaultRowCommitVersion, stats, remove#947040.stats) END AS remove#947093, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047]
Input [23]: [txn#947038, _extract_path#947149, _extract_partitionValues#947150, _extract_size#947151L, _extract_modificationTime#947152L, _extract_dataChange#947153, _extract_tags#947154, _extract_deletionVector#947155, _extract_baseRowId#947156L, _extract_defaultRowCommitVersion#947157L, _extract_clusteringProvider#947158, remove#947040, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047, version#947048L, add_stats_to_use#947060, add_path_canonical#947062, remove_path_canonical#947068]
(21) DeserializeToObject
Input [10]: [txn#947038, add#947082, remove#947093, metaData#947041, protocol#947042, cdc#947043, checkpointMetadata#947044, sidecar#947045, domainMetadata#947046, commitInfo#947047]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#947126: org.apache.spark.sql.delta.actions.SingleAction
(22) MapPartitions
Input [1]: [obj#947126]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@2d539f81, obj#947127: org.apache.spark.sql.delta.actions.SingleAction
(23) SerializeFromObject
Input [1]: [obj#947127]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#947128, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#947129, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#947130, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#947131, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#947132, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#947133, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#947134, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#947135, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#947136, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#947137]
(24) AdaptiveSparkPlan
Output [10]: [txn#947128, add#947129, remove#947130, metaData#947131, protocol#947132, cdc#947133, checkpointMetadata#947134, sidecar#947135, domainMetadata#947136, commitInfo#947137]
Arguments: isFinalPlan=true