org.apache.spark.api.java.JavaRDD.zipPartitions()方法的使用及代码示例

x33g5p2x  于2022-01-21 转载在 其他  
字(5.2k)|赞(0)|评价(0)|浏览(81)

本文整理了Java中org.apache.spark.api.java.JavaRDD.zipPartitions()方法的一些代码示例,展示了JavaRDD.zipPartitions()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaRDD.zipPartitions()方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaRDD
类名称:JavaRDD
方法名:zipPartitions

JavaRDD.zipPartitions介绍

暂无

代码示例

代码示例来源:origin: org.apache.spark/spark-core_2.10

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
  (Iterator<Integer> i, Iterator<String> s) -> {
   int sizeI = 0;
   while (i.hasNext()) {
    sizeI += 1;
    i.next();
   }
   int sizeS = 0;
   while (s.hasNext()) {
    sizeS += 1;
    s.next();
   }
   return Arrays.asList(sizeI, sizeS).iterator();
  };
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.apache.spark/spark-core_2.11

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
  (Iterator<Integer> i, Iterator<String> s) -> {
   int sizeI = 0;
   while (i.hasNext()) {
    sizeI += 1;
    i.next();
   }
   int sizeS = 0;
   while (s.hasNext()) {
    sizeS += 1;
    s.next();
   }
   return Arrays.asList(sizeI, sizeS).iterator();
  };
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.apache.spark/spark-core

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
  (Iterator<Integer> i, Iterator<String> s) -> {
   int sizeI = 0;
   while (i.hasNext()) {
    sizeI += 1;
    i.next();
   }
   int sizeS = 0;
   while (s.hasNext()) {
    sizeS += 1;
    s.next();
   }
   return Arrays.asList(sizeI, sizeS).iterator();
  };
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.apache.spark/spark-core_2.11

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
   (i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.apache.spark/spark-core_2.10

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
   (i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.apache.spark/spark-core

@Test
public void zipPartitions() {
 JavaRDD<Integer> rdd1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6), 2);
 JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("1", "2", "3", "4"), 2);
 FlatMapFunction2<Iterator<Integer>, Iterator<String>, Integer> sizesFn =
   (i, s) -> Arrays.asList(Iterators.size(i), Iterators.size(s)).iterator();
 JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
 assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}

代码示例来源:origin: org.datasyslab/geospark

final RightIndexLookupJudgement judgement =
      new RightIndexLookupJudgement(joinParams.considerBoundaryIntersection, dedupParams);
  resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.indexedRDD, judgement);
  resultWithDuplicates = leftRDD.indexedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
          dedupParams,
          buildCount, streamCount, resultCount, candidateCount);
  resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
resultWithDuplicates = rightRDD.spatialPartitionedRDD.zipPartitions(leftRDD.spatialPartitionedRDD, judgement);

代码示例来源:origin: DataSystemsLab/GeoSpark

final RightIndexLookupJudgement judgement =
      new RightIndexLookupJudgement(joinParams.considerBoundaryIntersection, dedupParams);
  resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.indexedRDD, judgement);
  resultWithDuplicates = leftRDD.indexedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
          dedupParams,
          buildCount, streamCount, resultCount, candidateCount);
  resultWithDuplicates = leftRDD.spatialPartitionedRDD.zipPartitions(rightRDD.spatialPartitionedRDD, judgement);
resultWithDuplicates = rightRDD.spatialPartitionedRDD.zipPartitions(leftRDD.spatialPartitionedRDD, judgement);

相关文章

微信公众号

最新文章

更多