digraph G {
0 [id="node0" labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>" tooltip="AdaptiveSparkPlan isFinalPlan=true"];
subgraph cluster1 {
isCluster="true";
id="cluster1";
label="WholeStageCodegen (3)";
tooltip="WholeStageCodegen (3)";
2 [id="node2" labelType="html" label="<br><b>SerializeFromObject</b><br><br>" tooltip="SerializeFromObject [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1101782, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1101783, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1101784, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1101785, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1101786, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1101787, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1101788, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1101789, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1101790, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1101791]"];
}
3 [id="node3" labelType="html" label="<br><b>MapPartitions</b><br><br>" tooltip="MapPartitions org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@41a9b44d, obj#1101781: org.apache.spark.sql.delta.actions.SingleAction"];
4 [id="node4" labelType="html" label="<br><b>DeserializeToObject</b><br><br>" tooltip="DeserializeToObject newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1101780: org.apache.spark.sql.delta.actions.SingleAction"];
subgraph cluster5 {
isCluster="true";
id="cluster5";
label="WholeStageCodegen (2)";
tooltip="WholeStageCodegen (2)";
6 [id="node6" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1101702, CASE WHEN isnotnull(_extract_path#1101802) THEN struct(path, add_path_canonical#1101716, partitionValues, _extract_partitionValues#1101803, size, _extract_size#1101804L, modificationTime, _extract_modificationTime#1101805L, dataChange, _extract_dataChange#1101806, stats, add_stats_to_use#1101714, tags, _extract_tags#1101807, deletionVector, _extract_deletionVector#1101808, baseRowId, _extract_baseRowId#1101809L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1101810L, clusteringProvider, _extract_clusteringProvider#1101811) END AS add#1101736, CASE WHEN isnotnull(remove#1101704.path) THEN if (isnull(remove#1101704)) null else named_struct(path, remove_path_canonical#1101722, deletionTimestamp, remove#1101704.deletionTimestamp, dataChange, remove#1101704.dataChange, extendedFileMetadata, remove#1101704.extendedFileMetadata, partitionValues, remove#1101704.partitionValues, size, remove#1101704.size, tags, remove#1101704.tags, deletionVector, remove#1101704.deletionVector, baseRowId, remove#1101704.baseRowId, defaultRowCommitVersion, remove#1101704.defaultRowCommitVersion, stats, remove#1101704.stats) END AS remove#1101747, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711]"];
7 [id="node7" labelType="html" label="<br><b>Sort</b><br><br>" tooltip="Sort [version#1101712L ASC NULLS FIRST], false, 0"];
}
8 [id="node8" labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 5<br>data size total (min, med, max (stageId: taskId))<br>9.0 KiB (4.4 KiB, 4.6 KiB, 4.6 KiB (stage 115221.0: task 378404))<br>shuffle write time total (min, med, max (stageId: taskId))<br>1 ms (0 ms, 0 ms, 0 ms (stage 115221.0: task 378403))<br>number of partitions: 50<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>3.7 KiB (1237.0 B, 2.5 KiB, 2.5 KiB (stage 115221.0: task 378404))" tooltip="Exchange hashpartitioning(coalesce(add_path_canonical#1101716, remove_path_canonical#1101722), 50), REPARTITION_BY_NUM, [plan_id=1699791]"];
subgraph cluster9 {
isCluster="true";
id="cluster9";
label="WholeStageCodegen (1)\n \nduration: total (min, med, max (stageId: taskId))\n516 ms (164 ms, 352 ms, 352 ms (stage 115221.0: task 378404))";
tooltip="WholeStageCodegen (1)";
10 [id="node10" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1101702, add#1101703.path AS _extract_path#1101802, add#1101703.partitionValues AS _extract_partitionValues#1101803, add#1101703.size AS _extract_size#1101804L, add#1101703.modificationTime AS _extract_modificationTime#1101805L, add#1101703.dataChange AS _extract_dataChange#1101806, add#1101703.tags AS _extract_tags#1101807, add#1101703.deletionVector AS _extract_deletionVector#1101808, add#1101703.baseRowId AS _extract_baseRowId#1101809L, add#1101703.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1101810L, add#1101703.clusteringProvider AS _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add#1101703.stats AS add_stats_to_use#1101714, CASE WHEN isnotnull(add#1101703.path) THEN UDF(add#1101703.path) END AS add_path_canonical#1101716, CASE WHEN isnotnull(remove#1101704.path) THEN UDF(remove#1101704.path) END AS remove_path_canonical#1101722]"];
}
11 [id="node11" labelType="html" label="<b>Scan json </b><br><br>number of files read: 2<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 8.1 KiB<br>number of output rows: 5<br>number of partitions read: 2" tooltip="FileScan json [txn#1101702,add#1101703,remove#1101704,metaData#1101705,protocol#1101706,cdc#1101707,checkpointMetadata#1101708,sidecar#1101709,domainMetadata#1101710,commitInfo#1101711,version#1101712L] Batched: false, DataFilters: [], Format: JSON, Location: DeltaLogFileIndex(2 paths)[hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
2->0;
3->2;
4->3;
6->4;
7->6;
8->7;
10->8;
11->10;
}
== Physical Plan ==
AdaptiveSparkPlan (18)
+- == Final Plan ==
ResultQueryStage (10), Statistics(sizeInBytes=8.0 EiB)
+- * SerializeFromObject (9)
+- MapPartitions (8)
+- DeserializeToObject (7)
+- * Project (6)
+- * Sort (5)
+- ShuffleQueryStage (4), Statistics(sizeInBytes=9.0 KiB, rowCount=5)
+- Exchange (3)
+- * Project (2)
+- Scan json (1)
+- == Initial Plan ==
SerializeFromObject (17)
+- MapPartitions (16)
+- DeserializeToObject (15)
+- Project (14)
+- Sort (13)
+- Exchange (12)
+- Project (11)
+- Scan json (1)
(1) Scan json
Output [11]: [txn#1101702, add#1101703, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-product-productplant/_delta_log/00000000000000000000.json, ... 1 entries]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(2) Project [codegen id : 1]
Output [23]: [txn#1101702, add#1101703.path AS _extract_path#1101802, add#1101703.partitionValues AS _extract_partitionValues#1101803, add#1101703.size AS _extract_size#1101804L, add#1101703.modificationTime AS _extract_modificationTime#1101805L, add#1101703.dataChange AS _extract_dataChange#1101806, add#1101703.tags AS _extract_tags#1101807, add#1101703.deletionVector AS _extract_deletionVector#1101808, add#1101703.baseRowId AS _extract_baseRowId#1101809L, add#1101703.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1101810L, add#1101703.clusteringProvider AS _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add#1101703.stats AS add_stats_to_use#1101714, CASE WHEN isnotnull(add#1101703.path) THEN UDF(add#1101703.path) END AS add_path_canonical#1101716, CASE WHEN isnotnull(remove#1101704.path) THEN UDF(remove#1101704.path) END AS remove_path_canonical#1101722]
Input [11]: [txn#1101702, add#1101703, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L]
(3) Exchange
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
Arguments: hashpartitioning(coalesce(add_path_canonical#1101716, remove_path_canonical#1101722), 50), REPARTITION_BY_NUM, [plan_id=1699791]
(4) ShuffleQueryStage
Output [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
Arguments: 0
(5) Sort [codegen id : 2]
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
Arguments: [version#1101712L ASC NULLS FIRST], false, 0
(6) Project [codegen id : 2]
Output [10]: [txn#1101702, CASE WHEN isnotnull(_extract_path#1101802) THEN struct(path, add_path_canonical#1101716, partitionValues, _extract_partitionValues#1101803, size, _extract_size#1101804L, modificationTime, _extract_modificationTime#1101805L, dataChange, _extract_dataChange#1101806, stats, add_stats_to_use#1101714, tags, _extract_tags#1101807, deletionVector, _extract_deletionVector#1101808, baseRowId, _extract_baseRowId#1101809L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1101810L, clusteringProvider, _extract_clusteringProvider#1101811) END AS add#1101736, CASE WHEN isnotnull(remove#1101704.path) THEN if (isnull(remove#1101704)) null else named_struct(path, remove_path_canonical#1101722, deletionTimestamp, remove#1101704.deletionTimestamp, dataChange, remove#1101704.dataChange, extendedFileMetadata, remove#1101704.extendedFileMetadata, partitionValues, remove#1101704.partitionValues, size, remove#1101704.size, tags, remove#1101704.tags, deletionVector, remove#1101704.deletionVector, baseRowId, remove#1101704.baseRowId, defaultRowCommitVersion, remove#1101704.defaultRowCommitVersion, stats, remove#1101704.stats) END AS remove#1101747, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711]
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
(7) DeserializeToObject
Input [10]: [txn#1101702, add#1101736, remove#1101747, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1101780: org.apache.spark.sql.delta.actions.SingleAction
(8) MapPartitions
Input [1]: [obj#1101780]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@41a9b44d, obj#1101781: org.apache.spark.sql.delta.actions.SingleAction
(9) SerializeFromObject [codegen id : 3]
Input [1]: [obj#1101781]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1101782, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1101783, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1101784, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1101785, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1101786, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1101787, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1101788, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1101789, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1101790, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1101791]
(10) ResultQueryStage
Output [10]: [txn#1101782, add#1101783, remove#1101784, metaData#1101785, protocol#1101786, cdc#1101787, checkpointMetadata#1101788, sidecar#1101789, domainMetadata#1101790, commitInfo#1101791]
Arguments: 1
(11) Project
Output [23]: [txn#1101702, add#1101703.path AS _extract_path#1101802, add#1101703.partitionValues AS _extract_partitionValues#1101803, add#1101703.size AS _extract_size#1101804L, add#1101703.modificationTime AS _extract_modificationTime#1101805L, add#1101703.dataChange AS _extract_dataChange#1101806, add#1101703.tags AS _extract_tags#1101807, add#1101703.deletionVector AS _extract_deletionVector#1101808, add#1101703.baseRowId AS _extract_baseRowId#1101809L, add#1101703.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1101810L, add#1101703.clusteringProvider AS _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add#1101703.stats AS add_stats_to_use#1101714, CASE WHEN isnotnull(add#1101703.path) THEN UDF(add#1101703.path) END AS add_path_canonical#1101716, CASE WHEN isnotnull(remove#1101704.path) THEN UDF(remove#1101704.path) END AS remove_path_canonical#1101722]
Input [11]: [txn#1101702, add#1101703, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L]
(12) Exchange
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
Arguments: hashpartitioning(coalesce(add_path_canonical#1101716, remove_path_canonical#1101722), 50), REPARTITION_BY_NUM, [plan_id=1699776]
(13) Sort
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
Arguments: [version#1101712L ASC NULLS FIRST], false, 0
(14) Project
Output [10]: [txn#1101702, CASE WHEN isnotnull(_extract_path#1101802) THEN struct(path, add_path_canonical#1101716, partitionValues, _extract_partitionValues#1101803, size, _extract_size#1101804L, modificationTime, _extract_modificationTime#1101805L, dataChange, _extract_dataChange#1101806, stats, add_stats_to_use#1101714, tags, _extract_tags#1101807, deletionVector, _extract_deletionVector#1101808, baseRowId, _extract_baseRowId#1101809L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1101810L, clusteringProvider, _extract_clusteringProvider#1101811) END AS add#1101736, CASE WHEN isnotnull(remove#1101704.path) THEN if (isnull(remove#1101704)) null else named_struct(path, remove_path_canonical#1101722, deletionTimestamp, remove#1101704.deletionTimestamp, dataChange, remove#1101704.dataChange, extendedFileMetadata, remove#1101704.extendedFileMetadata, partitionValues, remove#1101704.partitionValues, size, remove#1101704.size, tags, remove#1101704.tags, deletionVector, remove#1101704.deletionVector, baseRowId, remove#1101704.baseRowId, defaultRowCommitVersion, remove#1101704.defaultRowCommitVersion, stats, remove#1101704.stats) END AS remove#1101747, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711]
Input [23]: [txn#1101702, _extract_path#1101802, _extract_partitionValues#1101803, _extract_size#1101804L, _extract_modificationTime#1101805L, _extract_dataChange#1101806, _extract_tags#1101807, _extract_deletionVector#1101808, _extract_baseRowId#1101809L, _extract_defaultRowCommitVersion#1101810L, _extract_clusteringProvider#1101811, remove#1101704, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711, version#1101712L, add_stats_to_use#1101714, add_path_canonical#1101716, remove_path_canonical#1101722]
(15) DeserializeToObject
Input [10]: [txn#1101702, add#1101736, remove#1101747, metaData#1101705, protocol#1101706, cdc#1101707, checkpointMetadata#1101708, sidecar#1101709, domainMetadata#1101710, commitInfo#1101711]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1101780: org.apache.spark.sql.delta.actions.SingleAction
(16) MapPartitions
Input [1]: [obj#1101780]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@41a9b44d, obj#1101781: org.apache.spark.sql.delta.actions.SingleAction
(17) SerializeFromObject
Input [1]: [obj#1101781]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1101782, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1101783, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1101784, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1101785, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1101786, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1101787, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1101788, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1101789, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1101790, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1101791]
(18) AdaptiveSparkPlan
Output [10]: [txn#1101782, add#1101783, remove#1101784, metaData#1101785, protocol#1101786, cdc#1101787, checkpointMetadata#1101788, sidecar#1101789, domainMetadata#1101790, commitInfo#1101791]
Arguments: isFinalPlan=true