public class InsertIntoParquetTable extends org.apache.spark.sql.execution.SparkPlan implements org.apache.spark.sql.execution.UnaryNode, SparkHadoopMapReduceUtil, scala.Product, scala.Serializable
WARNING: EXPERIMENTAL! InsertIntoParquetTable with overwrite=false may cause data corruption in the case that multiple users try to append to the same table simultaneously. Inserting into a table that was previously generated by other means (e.g., by creating an HDFS directory and importing Parquet files generated by other tools) may cause unpredicted behaviour and therefore results in a RuntimeException (only detected via filename pattern so will not catch all cases).
Constructor and Description |
---|
InsertIntoParquetTable(ParquetRelation relation,
org.apache.spark.sql.execution.SparkPlan child,
boolean overwrite) |
Modifier and Type | Method and Description |
---|---|
org.apache.spark.sql.execution.SparkPlan |
child() |
RDD<Row> |
execute()
Inserts all rows into the Parquet file.
|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
boolean |
overwrite() |
ParquetRelation |
relation() |
codegenEnabled, executeCollect, executeTake, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, makeCopy, newMutableProjection, newOrdering, newPredicate, newProjection, org$apache$spark$Logging$$log__$eq, org$apache$spark$Logging$$log_, outputPartitioning, requiredChildDistribution, sparkContext, sqlContext
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, foreachUp, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, origin, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
firstAvailableClass, newJobContext, newTaskAttemptContext, newTaskAttemptID
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, log_
public InsertIntoParquetTable(ParquetRelation relation, org.apache.spark.sql.execution.SparkPlan child, boolean overwrite)
public ParquetRelation relation()
public org.apache.spark.sql.execution.SparkPlan child()
child
in interface org.apache.spark.sql.catalyst.trees.UnaryNode<org.apache.spark.sql.execution.SparkPlan>
public boolean overwrite()
public RDD<Row> execute()
execute
in class org.apache.spark.sql.execution.SparkPlan
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>