org.apache.spark.sql.DataFrameWriter.partitionBy()方法的使用及代码示例

x33g5p2x  于2022-01-18 转载在 其他  
字(2.0k)|赞(0)|评价(0)|浏览(265)

本文整理了Java中org.apache.spark.sql.DataFrameWriter.partitionBy()方法的一些代码示例,展示了DataFrameWriter.partitionBy()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。DataFrameWriter.partitionBy()方法的具体详情如下:
包路径:org.apache.spark.sql.DataFrameWriter
类名称:DataFrameWriter
方法名:partitionBy

DataFrameWriter.partitionBy介绍

暂无

代码示例

代码示例来源:origin: cloudera-labs/envelope

LOG.debug("Partitioning output");
writer = writer.partitionBy(columns.toArray(new String[columns.size()]));

代码示例来源:origin: cloudera-labs/envelope

@Override
public void applyBulkMutations(List<Tuple2<MutationType, Dataset<Row>>> planned) {    
 for (Tuple2<MutationType, Dataset<Row>> plan : planned) {
  MutationType mutationType = plan._1();
  Dataset<Row> mutation = (doesAlignColumns) ? alignColumns(plan._2()) : plan._2();
  DataFrameWriter<Row> writer = mutation.write();
  if (partitionColumns != null) {
   writer = writer.partitionBy(partitionColumns);
  }
  if (options != null) {
   writer = writer.options(options);
  }
  switch (mutationType) {
   case INSERT:
    writer = writer.mode(SaveMode.Append);
    break;
   case OVERWRITE:
    writer = writer.mode(SaveMode.Overwrite);
    break;
   default:
    throw new RuntimeException("Hive output does not support mutation type: " + mutationType);
  }
  writer.insertInto(tableName);
 }
}

代码示例来源:origin: sectong/SparkToParquet

df.write().partitionBy("ipAddress", "method", "responseCode").mode(SaveMode.Append).parquet(Flags.getInstance().getParquetFile());

代码示例来源:origin: com.cerner.bunsen/bunsen-core

.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(conceptMapTable);

代码示例来源:origin: cerner/bunsen

.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(conceptMapTable);

代码示例来源:origin: cerner/bunsen

.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(valueSetTable);

代码示例来源:origin: com.cerner.bunsen/bunsen-core

.write()
.format("parquet")
.partitionBy("timestamp")
.saveAsTable(valueSetTable);

相关文章