the logical plan representing the table. In the future this should be a org.apache.spark.sql.catalyst.catalog.CatalogTable once we converge Hive tables and data source tables.
a map from the partition key to the partition value (optional). If the partition
value is optional, dynamic partition insert will be performed.
As an example, INSERT INTO tbl PARTITION (a=1, b=2) AS ...
would have
Map('a' -> Some('1'), 'b' -> Some('2'))
and INSERT INTO tbl PARTITION (a=1, b) AS ...
would have
Map('a' -> Some('1'), 'b' -> None)
.
the logical plan representing data to write to.
overwrite existing table or partitions.
If true, only write if the table or partition does not exist.
the logical plan representing data to write to.
the logical plan representing data to write to.
If true, only write if the table or partition does not exist.
overwrite existing table or partitions.
a map from the partition key to the partition value (optional).
a map from the partition key to the partition value (optional). If the partition
value is optional, dynamic partition insert will be performed.
As an example, INSERT INTO tbl PARTITION (a=1, b=2) AS ...
would have
Map('a' -> Some('1'), 'b' -> Some('2'))
and INSERT INTO tbl PARTITION (a=1, b) AS ...
would have
Map('a' -> Some('1'), 'b' -> None)
.
Inserts all the rows in the table into Hive.
Inserts all the rows in the table into Hive. Row objects are properly serialized with the
org.apache.hadoop.hive.serde2.SerDe
and the
org.apache.hadoop.mapred.OutputFormat
provided by the table definition.
Note: this is run once and then kept to avoid double insertions.
the logical plan representing the table.
the logical plan representing the table. In the future this should be a org.apache.spark.sql.catalyst.catalog.CatalogTable once we converge Hive tables and data source tables.
Command for writing data out to a Hive table.
This class is mostly a mess, for legacy reasons (since it evolved in organic ways and had to follow Hive's internal implementations closely, which itself was a mess too). Please don't blame Reynold for this! He was just moving code around!
In the future we should converge the write path for Hive with the normal data source write path, as defined in org.apache.spark.sql.execution.datasources.FileFormatWriter.
the logical plan representing the table. In the future this should be a org.apache.spark.sql.catalyst.catalog.CatalogTable once we converge Hive tables and data source tables.
a map from the partition key to the partition value (optional). If the partition value is optional, dynamic partition insert will be performed. As an example,
INSERT INTO tbl PARTITION (a=1, b=2) AS ...
would haveand
INSERT INTO tbl PARTITION (a=1, b) AS ...
would have.
the logical plan representing data to write to.
overwrite existing table or partitions.
If true, only write if the table or partition does not exist.