digraph G {
0 [id="node0" labelType="html" label="<br><b>AdaptiveSparkPlan</b><br><br>" tooltip="AdaptiveSparkPlan isFinalPlan=true"];
subgraph cluster1 {
isCluster="true";
id="cluster1";
label="WholeStageCodegen (3)";
tooltip="WholeStageCodegen (3)";
2 [id="node2" labelType="html" label="<br><b>SerializeFromObject</b><br><br>" tooltip="SerializeFromObject [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1013448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1013449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1013450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1013451, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1013452, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1013453, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1013454, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1013455, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1013456, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1013457]"];
}
3 [id="node3" labelType="html" label="<br><b>MapPartitions</b><br><br>" tooltip="MapPartitions org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@6ce6ac2b, obj#1013447: org.apache.spark.sql.delta.actions.SingleAction"];
4 [id="node4" labelType="html" label="<br><b>DeserializeToObject</b><br><br>" tooltip="DeserializeToObject newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1013446: org.apache.spark.sql.delta.actions.SingleAction"];
subgraph cluster5 {
isCluster="true";
id="cluster5";
label="WholeStageCodegen (2)";
tooltip="WholeStageCodegen (2)";
6 [id="node6" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1013368, CASE WHEN isnotnull(_extract_path#1013468) THEN struct(path, add_path_canonical#1013382, partitionValues, _extract_partitionValues#1013469, size, _extract_size#1013470L, modificationTime, _extract_modificationTime#1013471L, dataChange, _extract_dataChange#1013472, stats, add_stats_to_use#1013380, tags, _extract_tags#1013473, deletionVector, _extract_deletionVector#1013474, baseRowId, _extract_baseRowId#1013475L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1013476L, clusteringProvider, _extract_clusteringProvider#1013477) END AS add#1013402, CASE WHEN isnotnull(remove#1013370.path) THEN if (isnull(remove#1013370)) null else named_struct(path, remove_path_canonical#1013388, deletionTimestamp, remove#1013370.deletionTimestamp, dataChange, remove#1013370.dataChange, extendedFileMetadata, remove#1013370.extendedFileMetadata, partitionValues, remove#1013370.partitionValues, size, remove#1013370.size, tags, remove#1013370.tags, deletionVector, remove#1013370.deletionVector, baseRowId, remove#1013370.baseRowId, defaultRowCommitVersion, remove#1013370.defaultRowCommitVersion, stats, remove#1013370.stats) END AS remove#1013413, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377]"];
7 [id="node7" labelType="html" label="<br><b>Sort</b><br><br>" tooltip="Sort [version#1013378L ASC NULLS FIRST], false, 0"];
}
8 [id="node8" labelType="html" label="<b>Exchange</b><br><br>shuffle records written: 5<br>data size total (min, med, max (stageId: taskId))<br>8.5 KiB (3.6 KiB, 4.9 KiB, 4.9 KiB (stage 105738.0: task 348443))<br>shuffle write time total (min, med, max (stageId: taskId))<br>1 ms (0 ms, 0 ms, 0 ms (stage 105738.0: task 348444))<br>number of partitions: 50<br>shuffle bytes written total (min, med, max (stageId: taskId))<br>3.1 KiB (1236.0 B, 1931.0 B, 1931.0 B (stage 105738.0: task 348444))" tooltip="Exchange hashpartitioning(coalesce(add_path_canonical#1013382, remove_path_canonical#1013388), 50), REPARTITION_BY_NUM, [plan_id=1634537]"];
subgraph cluster9 {
isCluster="true";
id="cluster9";
label="WholeStageCodegen (1)\n \nduration: total (min, med, max (stageId: taskId))\n185 ms (79 ms, 106 ms, 106 ms (stage 105738.0: task 348443))";
tooltip="WholeStageCodegen (1)";
10 [id="node10" labelType="html" label="<br><b>Project</b><br><br>" tooltip="Project [txn#1013368, add#1013369.path AS _extract_path#1013468, add#1013369.partitionValues AS _extract_partitionValues#1013469, add#1013369.size AS _extract_size#1013470L, add#1013369.modificationTime AS _extract_modificationTime#1013471L, add#1013369.dataChange AS _extract_dataChange#1013472, add#1013369.tags AS _extract_tags#1013473, add#1013369.deletionVector AS _extract_deletionVector#1013474, add#1013369.baseRowId AS _extract_baseRowId#1013475L, add#1013369.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1013476L, add#1013369.clusteringProvider AS _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add#1013369.stats AS add_stats_to_use#1013380, CASE WHEN isnotnull(add#1013369.path) THEN UDF(add#1013369.path) END AS add_path_canonical#1013382, CASE WHEN isnotnull(remove#1013370.path) THEN UDF(remove#1013370.path) END AS remove_path_canonical#1013388]"];
}
11 [id="node11" labelType="html" label="<b>Scan json </b><br><br>number of files read: 2<br>dynamic partition pruning time: 0 ms<br>metadata time: 0 ms<br>size of files read: 7.6 KiB<br>number of output rows: 5<br>number of partitions read: 2" tooltip="FileScan json [txn#1013368,add#1013369,remove#1013370,metaData#1013371,protocol#1013372,cdc#1013373,checkpointMetadata#1013374,sidecar#1013375,domainMetadata#1013376,commitInfo#1013377,version#1013378L] Batched: false, DataFilters: [], Format: JSON, Location: DeltaLogFileIndex(2 paths)[hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanac..., PartitionFilters: [], PushedFilters: [], ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitio..."];
2->0;
3->2;
4->3;
6->4;
7->6;
8->7;
10->8;
11->10;
}
== Physical Plan ==
AdaptiveSparkPlan (18)
+- == Final Plan ==
ResultQueryStage (10), Statistics(sizeInBytes=8.0 EiB)
+- * SerializeFromObject (9)
+- MapPartitions (8)
+- DeserializeToObject (7)
+- * Project (6)
+- * Sort (5)
+- ShuffleQueryStage (4), Statistics(sizeInBytes=8.5 KiB, rowCount=5)
+- Exchange (3)
+- * Project (2)
+- Scan json (1)
+- == Initial Plan ==
SerializeFromObject (17)
+- MapPartitions (16)
+- DeserializeToObject (15)
+- Project (14)
+- Sort (13)
+- Exchange (12)
+- Project (11)
+- Scan json (1)
(1) Scan json
Output [11]: [txn#1013368, add#1013369, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L]
Batched: false
Location: DeltaLogFileIndex [hdlfs://ced12e6d-bca7-47af-9525-449cba9023d7.files.hdl.prod-eu20.hanacloud.ondemand.com:443/crp-dl-stream-service/cornerstone/sap-cic-product-product-v2/_delta_log/00000000000000000000.json, ... 1 entries]
ReadSchema: struct<txn:struct<appId:string,version:bigint,lastUpdated:bigint>,add:struct<path:string,partitionValues:map<string,string>,size:bigint,modificationTime:bigint,dataChange:boolean,stats:string,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,clusteringProvider:string>,remove:struct<path:string,deletionTimestamp:bigint,dataChange:boolean,extendedFileMetadata:boolean,partitionValues:map<string,string>,size:bigint,tags:map<string,string>,deletionVector:struct<storageType:string,pathOrInlineDv:string,offset:int,sizeInBytes:int,cardinality:bigint,maxRowIndex:bigint>,baseRowId:bigint,defaultRowCommitVersion:bigint,stats:string>,metaData:struct<id:string,name:string,description:string,format:struct<provider:string,options:map<string,string>>,schemaString:string,partitionColumns:array<string>,configuration:map<string,string>,createdTime:bigint>,protocol:struct<minReaderVersion:int,minWriterVersion:int,readerFeatures:array<string>,writerFeatures:array<string>>,cdc:struct<path:string,partitionValues:map<string,string>,size:bigint,tags:map<string,string>>,checkpointMetadata:struct<version:bigint,tags:map<string,string>>,sidecar:struct<path:string,sizeInBytes:bigint,modificationTime:bigint,tags:map<string,string>>,domainMetadata:struct<domain:string,configuration:string,removed:boolean>,commitInfo:struct<version:bigint,inCommitTimestamp:bigint,timestamp:timestamp,userId:string,userName:string,operation:string,operationParameters:map<string,string>,job:struct<jobId:string,jobName:string,jobRunId:string,runId:string,jobOwnerId:string,triggerType:string>,notebook:struct<notebookId:string>,clusterId:string,readVersion:bigint,isolationLevel:string,isBlindAppend:boolean,operationMetrics:map<string,string>,userMetadata:string,tags:map<string,string>,engineInfo:string,txnId:string>>
(2) Project [codegen id : 1]
Output [23]: [txn#1013368, add#1013369.path AS _extract_path#1013468, add#1013369.partitionValues AS _extract_partitionValues#1013469, add#1013369.size AS _extract_size#1013470L, add#1013369.modificationTime AS _extract_modificationTime#1013471L, add#1013369.dataChange AS _extract_dataChange#1013472, add#1013369.tags AS _extract_tags#1013473, add#1013369.deletionVector AS _extract_deletionVector#1013474, add#1013369.baseRowId AS _extract_baseRowId#1013475L, add#1013369.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1013476L, add#1013369.clusteringProvider AS _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add#1013369.stats AS add_stats_to_use#1013380, CASE WHEN isnotnull(add#1013369.path) THEN UDF(add#1013369.path) END AS add_path_canonical#1013382, CASE WHEN isnotnull(remove#1013370.path) THEN UDF(remove#1013370.path) END AS remove_path_canonical#1013388]
Input [11]: [txn#1013368, add#1013369, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L]
(3) Exchange
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
Arguments: hashpartitioning(coalesce(add_path_canonical#1013382, remove_path_canonical#1013388), 50), REPARTITION_BY_NUM, [plan_id=1634537]
(4) ShuffleQueryStage
Output [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
Arguments: 0
(5) Sort [codegen id : 2]
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
Arguments: [version#1013378L ASC NULLS FIRST], false, 0
(6) Project [codegen id : 2]
Output [10]: [txn#1013368, CASE WHEN isnotnull(_extract_path#1013468) THEN struct(path, add_path_canonical#1013382, partitionValues, _extract_partitionValues#1013469, size, _extract_size#1013470L, modificationTime, _extract_modificationTime#1013471L, dataChange, _extract_dataChange#1013472, stats, add_stats_to_use#1013380, tags, _extract_tags#1013473, deletionVector, _extract_deletionVector#1013474, baseRowId, _extract_baseRowId#1013475L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1013476L, clusteringProvider, _extract_clusteringProvider#1013477) END AS add#1013402, CASE WHEN isnotnull(remove#1013370.path) THEN if (isnull(remove#1013370)) null else named_struct(path, remove_path_canonical#1013388, deletionTimestamp, remove#1013370.deletionTimestamp, dataChange, remove#1013370.dataChange, extendedFileMetadata, remove#1013370.extendedFileMetadata, partitionValues, remove#1013370.partitionValues, size, remove#1013370.size, tags, remove#1013370.tags, deletionVector, remove#1013370.deletionVector, baseRowId, remove#1013370.baseRowId, defaultRowCommitVersion, remove#1013370.defaultRowCommitVersion, stats, remove#1013370.stats) END AS remove#1013413, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377]
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
(7) DeserializeToObject
Input [10]: [txn#1013368, add#1013402, remove#1013413, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1013446: org.apache.spark.sql.delta.actions.SingleAction
(8) MapPartitions
Input [1]: [obj#1013446]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@6ce6ac2b, obj#1013447: org.apache.spark.sql.delta.actions.SingleAction
(9) SerializeFromObject [codegen id : 3]
Input [1]: [obj#1013447]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1013448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1013449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1013450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1013451, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1013452, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1013453, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1013454, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1013455, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1013456, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1013457]
(10) ResultQueryStage
Output [10]: [txn#1013448, add#1013449, remove#1013450, metaData#1013451, protocol#1013452, cdc#1013453, checkpointMetadata#1013454, sidecar#1013455, domainMetadata#1013456, commitInfo#1013457]
Arguments: 1
(11) Project
Output [23]: [txn#1013368, add#1013369.path AS _extract_path#1013468, add#1013369.partitionValues AS _extract_partitionValues#1013469, add#1013369.size AS _extract_size#1013470L, add#1013369.modificationTime AS _extract_modificationTime#1013471L, add#1013369.dataChange AS _extract_dataChange#1013472, add#1013369.tags AS _extract_tags#1013473, add#1013369.deletionVector AS _extract_deletionVector#1013474, add#1013369.baseRowId AS _extract_baseRowId#1013475L, add#1013369.defaultRowCommitVersion AS _extract_defaultRowCommitVersion#1013476L, add#1013369.clusteringProvider AS _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add#1013369.stats AS add_stats_to_use#1013380, CASE WHEN isnotnull(add#1013369.path) THEN UDF(add#1013369.path) END AS add_path_canonical#1013382, CASE WHEN isnotnull(remove#1013370.path) THEN UDF(remove#1013370.path) END AS remove_path_canonical#1013388]
Input [11]: [txn#1013368, add#1013369, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L]
(12) Exchange
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
Arguments: hashpartitioning(coalesce(add_path_canonical#1013382, remove_path_canonical#1013388), 50), REPARTITION_BY_NUM, [plan_id=1634522]
(13) Sort
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
Arguments: [version#1013378L ASC NULLS FIRST], false, 0
(14) Project
Output [10]: [txn#1013368, CASE WHEN isnotnull(_extract_path#1013468) THEN struct(path, add_path_canonical#1013382, partitionValues, _extract_partitionValues#1013469, size, _extract_size#1013470L, modificationTime, _extract_modificationTime#1013471L, dataChange, _extract_dataChange#1013472, stats, add_stats_to_use#1013380, tags, _extract_tags#1013473, deletionVector, _extract_deletionVector#1013474, baseRowId, _extract_baseRowId#1013475L, defaultRowCommitVersion, _extract_defaultRowCommitVersion#1013476L, clusteringProvider, _extract_clusteringProvider#1013477) END AS add#1013402, CASE WHEN isnotnull(remove#1013370.path) THEN if (isnull(remove#1013370)) null else named_struct(path, remove_path_canonical#1013388, deletionTimestamp, remove#1013370.deletionTimestamp, dataChange, remove#1013370.dataChange, extendedFileMetadata, remove#1013370.extendedFileMetadata, partitionValues, remove#1013370.partitionValues, size, remove#1013370.size, tags, remove#1013370.tags, deletionVector, remove#1013370.deletionVector, baseRowId, remove#1013370.baseRowId, defaultRowCommitVersion, remove#1013370.defaultRowCommitVersion, stats, remove#1013370.stats) END AS remove#1013413, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377]
Input [23]: [txn#1013368, _extract_path#1013468, _extract_partitionValues#1013469, _extract_size#1013470L, _extract_modificationTime#1013471L, _extract_dataChange#1013472, _extract_tags#1013473, _extract_deletionVector#1013474, _extract_baseRowId#1013475L, _extract_defaultRowCommitVersion#1013476L, _extract_clusteringProvider#1013477, remove#1013370, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377, version#1013378L, add_stats_to_use#1013380, add_path_canonical#1013382, remove_path_canonical#1013388]
(15) DeserializeToObject
Input [10]: [txn#1013368, add#1013402, remove#1013413, metaData#1013371, protocol#1013372, cdc#1013373, checkpointMetadata#1013374, sidecar#1013375, domainMetadata#1013376, commitInfo#1013377]
Arguments: newInstance(class org.apache.spark.sql.delta.actions.SingleAction), obj#1013446: org.apache.spark.sql.delta.actions.SingleAction
(16) MapPartitions
Input [1]: [obj#1013446]
Arguments: org.apache.spark.sql.delta.Snapshot$$Lambda/0x0000000801cc9f20@6ce6ac2b, obj#1013447: org.apache.spark.sql.delta.actions.SingleAction
(17) SerializeFromObject
Input [1]: [obj#1013447]
Arguments: [if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn()))) null else named_struct(appId, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).appId()))), version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).version()), lastUpdated, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).txn())).lastUpdated()))) AS txn#1013448, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -1), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -2), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).size()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).modificationTime()), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).dataChange()), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).stats()))), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -3), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -4), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).defaultRowCommitVersion())), clusteringProvider, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).add())).clusteringProvider()))))) AS add#1013449, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).path()))), deletionTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionTimestamp())), dataChange, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).dataChange()), extendedFileMetadata, unwrapoption(BooleanType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).extendedFileMetadata())), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -5), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -6), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).partitionValues())), size, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).size())), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -7), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -8), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).tags())), deletionVector, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector()))) null else named_struct(storageType, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).storageType()))), pathOrInlineDv, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).pathOrInlineDv()))), offset, unwrapoption(IntegerType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).offset())), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).sizeInBytes()), cardinality, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).cardinality()), maxRowIndex, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).deletionVector())).maxRowIndex()))), baseRowId, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).baseRowId())), defaultRowCommitVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).defaultRowCommitVersion())), stats, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).remove())).stats())))) AS remove#1013450, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData()))) null else named_struct(id, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).id()))), name, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).name()))), description, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).description()))), format, if (isnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format()))) null else named_struct(provider, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).provider()))), options, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -9), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -10), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).format())).options()))), schemaString, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).schemaString()))), partitionColumns, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -11), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).partitionColumns()), None), configuration, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -12), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -13), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).configuration())), createdTime, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).metaData())).createdTime()))) AS metaData#1013451, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol()))) null else named_struct(minReaderVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minReaderVersion()), minWriterVersion, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).minWriterVersion()), readerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -14), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).readerFeatures())).toSeq()), None), writerFeatures, mapobjects(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(MapObject, ObjectType(class java.lang.Object), true, -15), StringType, ObjectType(class java.lang.String)))), invoke(unwrapoption(ObjectType(interface scala.collection.immutable.Set), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).protocol())).writerFeatures())).toSeq()), None)) AS protocol#1013452, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).path()))), partitionValues, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -16), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -17), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).partitionValues())), size, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).size()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -18), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -19), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).cdc())).tags()))) AS cdc#1013453, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata()))) null else named_struct(version, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).version()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -20), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -21), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).checkpointMetadata())).tags()))) AS checkpointMetadata#1013454, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar()))) null else named_struct(path, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).path()))), sizeInBytes, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).sizeInBytes()), modificationTime, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).modificationTime()), tags, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -22), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -23), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).sidecar())).tags()))) AS sidecar#1013455, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata()))) null else named_struct(domain, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).domain()))), configuration, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).configuration()))), removed, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).domainMetadata())).removed())) AS domainMetadata#1013456, if (isnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo()))) null else named_struct(version, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).version())), inCommitTimestamp, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).inCommitTimestamp())), timestamp, static_invoke(DateTimeUtils.fromJavaTimestamp(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).timestamp()))), userId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userId())))), userName, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).userName())))), operation, static_invoke(UTF8String.fromString(invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operation()))), operationParameters, externalmaptocatalyst(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_key, ObjectType(class java.lang.Object), true, -24), StringType, ObjectType(class java.lang.String)))), lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), static_invoke(UTF8String.fromString(validateexternaltype(lambdavariable(ExternalMapToCatalyst_value, ObjectType(class java.lang.Object), true, -25), StringType, ObjectType(class java.lang.String)))), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).operationParameters())), job, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job())))) null else named_struct(jobId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobId()))), jobName, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobName()))), jobRunId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobRunId()))), runId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).runId()))), jobOwnerId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).jobOwnerId()))), triggerType, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.JobInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).job()))).triggerType())))), notebook, if (isnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook())))) null else named_struct(notebookId, static_invoke(UTF8String.fromString(invoke(knownnotnull(unwrapoption(ObjectType(class org.apache.spark.sql.delta.actions.NotebookInfo), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).notebook()))).notebookId())))), clusterId, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).clusterId())))), readVersion, unwrapoption(LongType, invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).readVersion())), isolationLevel, static_invoke(UTF8String.fromString(unwrapoption(ObjectType(class java.lang.String), invoke(knownnotnull(invoke(knownnotnull(assertnotnull(input[0, org.apache.spark.sql.delta.actions.SingleAction, true])).commitInfo())).isolationLevel())))), isBlindAppend, ... 11 more fields) AS commitInfo#1013457]
(18) AdaptiveSparkPlan
Output [10]: [txn#1013448, add#1013449, remove#1013450, metaData#1013451, protocol#1013452, cdc#1013453, checkpointMetadata#1013454, sidecar#1013455, domainMetadata#1013456, commitInfo#1013457]
Arguments: isFinalPlan=true