本文整理了Java中org.apache.spark.api.java.JavaRDD.zipPartitions()
方法的一些代码示例,展示了JavaRDD.zipPartitions()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaRDD.zipPartitions()
方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaRDD
类名称:JavaRDD
方法名:zipPartitions
暂无
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(Iterator<Integer> i, Iterator<String> s) -> {
int sizeI = 0;
while (i.hasNext()) {
sizeI += 1;
i.next();
}
int sizeS = 0;
while (s.hasNext()) {
sizeS += 1;
s.next();
}
return Arrays.asList(sizeI, sizeS).iterator();
};
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(Iterator<Integer> i, Iterator<String> s) -> {
int sizeI = 0;
while (i.hasNext()) {
sizeI += 1;
i.next();
}
int sizeS = 0;
while (s.hasNext()) {
sizeS += 1;
s.next();
}
return Arrays.asList(sizeI, sizeS).iterator();
};
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(Iterator<Integer> i, Iterator<String> s) -> {
int sizeI = 0;
while (i.hasNext()) {
sizeI += 1;
i.next();
}
int sizeS = 0;
while (s.hasNext()) {
sizeS += 1;
s.next();
}
return Arrays.asList(sizeI, sizeS).iterator();
};
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void zipPartitions() {
JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
(i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
代码示例来源:origin: org.datasyslab/geospark
final RightIndexLookupJudgement judgement =
new RightIndexLookupJudgement(joinParams.considerBoundaryIntersection, dedupParams);
resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.indexedRDD, judgement);
resultWithDuplicates = leftRDD.indexedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
dedupParams,
buildCount, streamCount, resultCount, candidateCount);
resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
resultWithDuplicates = rightRDD.spatialPartitionedRDD.zipPartitions(leftRDD.spatialPartitionedRDD, judgement);
代码示例来源:origin: DataSystemsLab/GeoSpark
final RightIndexLookupJudgement judgement =
new RightIndexLookupJudgement(joinParams.considerBoundaryIntersection, dedupParams);
resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.indexedRDD, judgement);
resultWithDuplicates = leftRDD.indexedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
dedupParams,
buildCount, streamCount, resultCount, candidateCount);
resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
resultWithDuplicates = rightRDD.spatialPartitionedRDD.zipPartitions(leftRDD.spatialPartitionedRDD, judgement);
内容来源于网络,如有侵权,请联系作者删除!