public class ScriptTransformationExec
extends org.apache.spark.sql.execution.SparkPlan
implements org.apache.spark.sql.execution.UnaryExecNode, scala.Product, scala.Serializable
param: input the set of expression that should be passed to the script. param: script the command that should be executed. param: output the attributes that are produced by the script.
Constructor and Description |
---|
ScriptTransformationExec(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input,
String script,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output,
org.apache.spark.sql.execution.SparkPlan child,
HiveScriptIOSchema ioschema) |
Modifier and Type | Method and Description |
---|---|
abstract static R |
apply(T1 v1,
T2 v2,
T3 v3,
T4 v4,
T5 v5) |
org.apache.spark.sql.execution.SparkPlan |
child() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
input() |
HiveScriptIOSchema |
ioschema() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
org.apache.spark.sql.catalyst.plans.physical.Partitioning |
outputPartitioning() |
org.apache.spark.sql.catalyst.expressions.AttributeSet |
producedAttributes() |
String |
script() |
static String |
toString() |
cleanupResources, doExecuteBroadcast, doExecuteColumnar, doPrepare, execute, executeBroadcast, executeCollect, executeCollectIterator, executeCollectPublic, executeColumnar, executeQuery, executeTake, executeToIterator, id, initializeForcefully, initializeLogIfNecessary, initializeLogIfNecessary, initializeLogIfNecessary$default$2, isTraceEnabled, log, logDebug, logDebug, logError, logError, LOGICAL_PLAN_INHERITED_TAG, LOGICAL_PLAN_TAG, logicalLink, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, longMetric, makeCopy, metrics, org$apache$spark$internal$Logging$$log__$eq, org$apache$spark$internal$Logging$$log_, org$apache$spark$sql$execution$SparkPlan$$runningSubqueries, outputOrdering, prepare, prepareSubqueries, requiredChildDistribution, requiredChildOrdering, resetMetrics, setLogicalLink, sparkContext, sqlContext, supportsColumnar, vectorTypes, waitForSubqueries
allAttributes, append, append$default$5, append$default$6, canonicalized, CODEGEN_ID_TAG, collectInPlanAndSubqueries, conf, doCanonicalize, expressions, findExpressionAndTrackLineageDown, innerChildren, inputSet, isCanonicalizedPlan, mapExpressions, missingInput, normalizeExpressions, normalizePredicates, OP_ID_TAG, outputSet, printSchema, references, sameResult, schema, schemaString, semanticHash, simpleString, simpleStringWithNodeId, statePrefix, subqueries, subqueriesAll, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp, verboseString, verboseStringWithOperatorId
apply, argString, asCode, children, clone, collect, collectFirst, collectLeaves, containsChild, copyTagsFrom, fastEquals, find, flatMap, foreach, foreachUp, generateTreeString, generateTreeString$default$5, generateTreeString$default$6, getTagValue, hashCode, jsonFields, map, mapChildren, mapProductIterator, nodeName, numberedTreeString, origin, otherCopyArgs, p, prettyJson, productIterator, productPrefix, setTagValue, stringArgs, toJSON, toString, transform, transformDown, transformUp, treeString, treeString, treeString, treeString$default$2, treeString$default$3, treeString$default$4, unsetTagValue, verboseStringWithSuffix, withNewChildren
$init$, children, unapply, verboseStringWithOperatorId
$init$, productArity, productElement, productIterator, productPrefix
initializeLogging
public ScriptTransformationExec(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input, String script, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output, org.apache.spark.sql.execution.SparkPlan child, HiveScriptIOSchema ioschema)
public abstract static R apply(T1 v1, T2 v2, T3 v3, T4 v4, T5 v5)
public static String toString()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> input()
public String script()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>
public org.apache.spark.sql.execution.SparkPlan child()
child
in interface org.apache.spark.sql.execution.UnaryExecNode
public HiveScriptIOSchema ioschema()
public org.apache.spark.sql.catalyst.expressions.AttributeSet producedAttributes()
producedAttributes
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>
public org.apache.spark.sql.catalyst.plans.physical.Partitioning outputPartitioning()
outputPartitioning
in class org.apache.spark.sql.execution.SparkPlan