本文整理了Java中org.apache.spark.sql.DataFrameWriter.partitionBy()
方法的一些代码示例,展示了DataFrameWriter.partitionBy()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。DataFrameWriter.partitionBy()
方法的具体详情如下:
包路径:org.apache.spark.sql.DataFrameWriter
类名称:DataFrameWriter
方法名:partitionBy
暂无
代码示例来源:origin: cloudera-labs/envelope
LOG.debug("Partitioning output");
writer = writer.partitionBy(columns.toArray(new String[columns.size()]));
代码示例来源:origin: cloudera-labs/envelope
@Override
public void applyBulkMutations(List<Tuple2<MutationType, Dataset<Row>>> planned) {
for (Tuple2<MutationType, Dataset<Row>> plan : planned) {
MutationType mutationType = plan._1();
Dataset<Row> mutation = (doesAlignColumns) ? alignColumns(plan._2()) : plan._2();
DataFrameWriter<Row> writer = mutation.write();
if (partitionColumns != null) {
writer = writer.partitionBy(partitionColumns);
}
if (options != null) {
writer = writer.options(options);
}
switch (mutationType) {
case INSERT:
writer = writer.mode(SaveMode.Append);
break;
case OVERWRITE:
writer = writer.mode(SaveMode.Overwrite);
break;
default:
throw new RuntimeException("Hive output does not support mutation type: " + mutationType);
}
writer.insertInto(tableName);
}
}
代码示例来源:origin: sectong/SparkToParquet
df.write().partitionBy("ipAddress", "method", "responseCode").mode(SaveMode.Append).parquet(Flags.getInstance().getParquetFile());
代码示例来源:origin: com.cerner.bunsen/bunsen-core
.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(conceptMapTable);
代码示例来源:origin: cerner/bunsen
.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(conceptMapTable);
代码示例来源:origin: cerner/bunsen
.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(valueSetTable);
代码示例来源:origin: com.cerner.bunsen/bunsen-core
.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(valueSetTable);
内容来源于网络,如有侵权,请联系作者删除!