public class AnalyzeTable extends SparkPlan implements Command, scala.Product, scala.Serializable
Right now, it only supports Hive tables and it only updates the size of a Hive table in the Hive metastore.
Constructor and Description |
---|
AnalyzeTable(String tableName) |
Modifier and Type | Method and Description |
---|---|
RDD<org.apache.spark.sql.catalyst.expressions.Row> |
execute()
Runs this query returning the result as an RDD.
|
HiveContext |
hiveContext() |
scala.collection.Seq<scala.runtime.Nothing$> |
output() |
String |
tableName() |
codegenEnabled, executeCollect, makeCopy, outputPartitioning, requiredChildDistribution
expressions, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, schema, schemaString, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, generateTreeString, getNodeNumbered, id, map, mapChildren, nextId, nodeName, numberedTreeString, otherCopyArgs, sameInstance, simpleString, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initialized, initializeIfNecessary, initializeLogging, initLock, isTraceEnabled, log_, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning
public String tableName()
public HiveContext hiveContext()
public scala.collection.Seq<scala.runtime.Nothing$> output()