本文整理了Java中org.apache.spark.api.java.JavaRDD.aggregate()
方法的一些代码示例,展示了JavaRDD.aggregate()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaRDD.aggregate()
方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaRDD
类名称:JavaRDD
方法名:aggregate
暂无
代码示例来源:origin: databricks/learning-spark
AvgCount result = rdd.aggregate(initial, addAndCount, combine);
System.out.println(result.avg());
sc.stop();
代码示例来源:origin: databricks/learning-spark
AvgCount result = rdd.aggregate(initial, addAndCount, combine);
System.out.println(result.avg());
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void aggregate() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.aggregate(0, new AddInts(), new AddInts());
assertEquals(10, sum);
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void aggregate() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.aggregate(0, new AddInts(), new AddInts());
assertEquals(10, sum);
}
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void aggregate() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.aggregate(0, new AddInts(), new AddInts());
assertEquals(10, sum);
}
代码示例来源:origin: DataSystemsLab/GeoSpark
StatCalculator agg = (StatCalculator) this.rawSpatialRDD.aggregate(null, seqOp, combOp);
if (agg != null) {
this.boundaryEnvelope = agg.getBoundary();
代码示例来源:origin: org.datasyslab/geospark
StatCalculator agg = (StatCalculator) this.rawSpatialRDD.aggregate(null, seqOp, combOp);
if (agg != null) {
this.boundaryEnvelope = agg.getBoundary();
代码示例来源:origin: org.datavec/datavec-spark_2.11
/**
*
* @param schema
* @param data
* @return
*/
public static DataQualityAnalysis analyzeQuality(final Schema schema, final JavaRDD<List<Writable>> data) {
data.cache();
int nColumns = schema.numColumns();
List<ColumnType> columnTypes = schema.getColumnTypes();
List<QualityAnalysisState> states = data.aggregate(null, new QualityAnalysisAddFunction(schema),
new QualityAnalysisCombineFunction());
List<ColumnQuality> list = new ArrayList<>(nColumns);
for (QualityAnalysisState qualityState : states) {
list.add(qualityState.getColumnQuality());
}
return new DataQualityAnalysis(schema, list);
}
代码示例来源:origin: org.datavec/datavec-spark
/**
*
* @param schema
* @param data
* @return
*/
public static DataQualityAnalysis analyzeQuality(final Schema schema, final JavaRDD<List<Writable>> data) {
data.cache();
int nColumns = schema.numColumns();
List<ColumnType> columnTypes = schema.getColumnTypes();
List<QualityAnalysisState> states = data.aggregate(null, new QualityAnalysisAddFunction(schema),
new QualityAnalysisCombineFunction());
List<ColumnQuality> list = new ArrayList<>(nColumns);
for (QualityAnalysisState qualityState : states) {
list.add(qualityState.getColumnQuality());
}
return new DataQualityAnalysis(schema, list);
}
代码示例来源:origin: ypriverol/spark-java8
AvgCount result = rdd.aggregate(initial,
(AvgCount a, Integer x) -> {
a.total_ += x;
代码示例来源:origin: org.datavec/datavec-spark_2.11
data.aggregate(null, new AnalysisAddFunction(schema), new AnalysisCombineFunction());
data.aggregate(null, new HistogramAddFunction(maxHistogramBuckets, schema, minsMaxes),
new HistogramCombineFunction());
代码示例来源:origin: org.datavec/datavec-spark
data.aggregate(null, new AnalysisAddFunction(schema), new AnalysisCombineFunction());
data.aggregate(null, new HistogramAddFunction(maxHistogramBuckets, schema, minsMaxes),
new HistogramCombineFunction());
代码示例来源:origin: org.apache.beam/beam-runners-spark
inputRDDBytes.aggregate(
CoderHelpers.toByteArray(sparkCombineFn.zeroValue(), iterAccumCoder),
(ab, ib) -> {
代码示例来源:origin: org.datavec/datavec-spark_2.11
seqLengths.cache();
SequenceLengthAnalysisCounter counter = new SequenceLengthAnalysisCounter();
counter = seqLengths.aggregate(counter, new SequenceLengthAnalysisAddFunction(),
new SequenceLengthAnalysisMergeFunction());
代码示例来源:origin: org.datavec/datavec-spark
seqLengths.cache();
SequenceLengthAnalysisCounter counter = new SequenceLengthAnalysisCounter();
counter = seqLengths.aggregate(counter, new SequenceLengthAnalysisAddFunction(),
new SequenceLengthAnalysisMergeFunction());
内容来源于网络,如有侵权,请联系作者删除!