org.apache.spark.api.java.JavaRDD.aggregate()方法的使用及代码示例

x33g5p2x  于2022-01-21 转载在 其他  
字(4.0k)|赞(0)|评价(0)|浏览(129)

本文整理了Java中org.apache.spark.api.java.JavaRDD.aggregate()方法的一些代码示例,展示了JavaRDD.aggregate()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaRDD.aggregate()方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaRDD
类名称:JavaRDD
方法名:aggregate

JavaRDD.aggregate介绍

暂无

代码示例

代码示例来源:origin: databricks/learning-spark

AvgCount result = rdd.aggregate(initial, addAndCount, combine);
System.out.println(result.avg());
sc.stop();

代码示例来源:origin: databricks/learning-spark

AvgCount result = rdd.aggregate(initial, addAndCount, combine);
System.out.println(result.avg());

代码示例来源:origin: org.apache.spark/spark-core

@Test
public void aggregate() {
 JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
 int sum = rdd.aggregate(0, new AddInts(), new AddInts());
 assertEquals(10, sum);
}

代码示例来源:origin: org.apache.spark/spark-core_2.11

@Test
public void aggregate() {
 JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
 int sum = rdd.aggregate(0, new AddInts(), new AddInts());
 assertEquals(10, sum);
}

代码示例来源:origin: org.apache.spark/spark-core_2.10

@Test
public void aggregate() {
 JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
 int sum = rdd.aggregate(0, new AddInts(), new AddInts());
 assertEquals(10, sum);
}

代码示例来源:origin: DataSystemsLab/GeoSpark

StatCalculator agg = (StatCalculator) this.rawSpatialRDD.aggregate(null, seqOp, combOp);
if (agg != null) {
  this.boundaryEnvelope = agg.getBoundary();

代码示例来源:origin: org.datasyslab/geospark

StatCalculator agg = (StatCalculator) this.rawSpatialRDD.aggregate(null, seqOp, combOp);
if (agg != null) {
  this.boundaryEnvelope = agg.getBoundary();

代码示例来源:origin: org.datavec/datavec-spark_2.11

/**
 *
 * @param schema
 * @param data
 * @return
 */
public static DataQualityAnalysis analyzeQuality(final Schema schema, final JavaRDD<List<Writable>> data) {
  data.cache();
  int nColumns = schema.numColumns();
  List<ColumnType> columnTypes = schema.getColumnTypes();
  List<QualityAnalysisState> states = data.aggregate(null, new QualityAnalysisAddFunction(schema),
          new QualityAnalysisCombineFunction());
  List<ColumnQuality> list = new ArrayList<>(nColumns);
  for (QualityAnalysisState qualityState : states) {
    list.add(qualityState.getColumnQuality());
  }
  return new DataQualityAnalysis(schema, list);
}

代码示例来源:origin: org.datavec/datavec-spark

/**
 *
 * @param schema
 * @param data
 * @return
 */
public static DataQualityAnalysis analyzeQuality(final Schema schema, final JavaRDD<List<Writable>> data) {
  data.cache();
  int nColumns = schema.numColumns();
  List<ColumnType> columnTypes = schema.getColumnTypes();
  List<QualityAnalysisState> states = data.aggregate(null, new QualityAnalysisAddFunction(schema),
          new QualityAnalysisCombineFunction());
  List<ColumnQuality> list = new ArrayList<>(nColumns);
  for (QualityAnalysisState qualityState : states) {
    list.add(qualityState.getColumnQuality());
  }
  return new DataQualityAnalysis(schema, list);
}

代码示例来源:origin: ypriverol/spark-java8

AvgCount result = rdd.aggregate(initial,
    (AvgCount a, Integer x) -> {
      a.total_ += x;

代码示例来源:origin: org.datavec/datavec-spark_2.11

data.aggregate(null, new AnalysisAddFunction(schema), new AnalysisCombineFunction());
data.aggregate(null, new HistogramAddFunction(maxHistogramBuckets, schema, minsMaxes),
        new HistogramCombineFunction());

代码示例来源:origin: org.datavec/datavec-spark

data.aggregate(null, new AnalysisAddFunction(schema), new AnalysisCombineFunction());
data.aggregate(null, new HistogramAddFunction(maxHistogramBuckets, schema, minsMaxes),
        new HistogramCombineFunction());

代码示例来源:origin: org.apache.beam/beam-runners-spark

inputRDDBytes.aggregate(
  CoderHelpers.toByteArray(sparkCombineFn.zeroValue(), iterAccumCoder),
  (ab, ib) -> {

代码示例来源:origin: org.datavec/datavec-spark_2.11

seqLengths.cache();
SequenceLengthAnalysisCounter counter = new SequenceLengthAnalysisCounter();
counter = seqLengths.aggregate(counter, new SequenceLengthAnalysisAddFunction(),
        new SequenceLengthAnalysisMergeFunction());

代码示例来源:origin: org.datavec/datavec-spark

seqLengths.cache();
SequenceLengthAnalysisCounter counter = new SequenceLengthAnalysisCounter();
counter = seqLengths.aggregate(counter, new SequenceLengthAnalysisAddFunction(),
        new SequenceLengthAnalysisMergeFunction());

相关文章

微信公众号

最新文章

更多