org.apache.spark.api.java.JavaPairRDD.union()方法的使用及代码示例

x33g5p2x  于2022-01-21 转载在 其他  
字(6.2k)|赞(0)|评价(0)|浏览(99)

本文整理了Java中org.apache.spark.api.java.JavaPairRDD.union()方法的一些代码示例,展示了JavaPairRDD.union()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaPairRDD.union()方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaPairRDD
类名称:JavaPairRDD
方法名:union

JavaPairRDD.union介绍

暂无

代码示例

代码示例来源:origin: OryxProject/oryx

@Override
public void runUpdate(JavaSparkContext sparkContext,
           long timestamp,
           JavaPairRDD<String,String> newData,
           JavaPairRDD<String,String> pastData,
           String modelDirString,
           TopicProducer<String,String> modelUpdateTopic) throws IOException {
 JavaPairRDD<String,String> allData = pastData == null ? newData : newData.union(pastData);
 String modelString;
 try {
  modelString = new ObjectMapper().writeValueAsString(countDistinctOtherWords(allData));
 } catch (JsonProcessingException jpe) {
  throw new IOException(jpe);
 }
 modelUpdateTopic.send("MODEL", modelString);
}

代码示例来源:origin: apache/drill

rdd = prevRDD;
  } else {
   rdd = rdd.union(prevRDD);
 finalRDD = rdd;
} else {
 finalRDD = finalRDD.union(rdd);

代码示例来源:origin: apache/hive

sc.setCallSite(CallSite.apply("UnionRDD (" + rdd.name() + ", " +
          prevRDD.name() + ")", ""));
  rdd = rdd.union(prevRDD);
  rdd.setName("UnionRDD (" + rdd.getNumPartitions() + ")");
sc.setCallSite(CallSite.apply("UnionRDD (" + rdd.name() + ", " + finalRDD.name() + ")",
    ""));
finalRDD = finalRDD.union(rdd);
finalRDD.setName("UnionRDD (" + finalRDD.getNumPartitions() + ")");

代码示例来源:origin: mahmoudparsian/data-algorithms-book

JavaPairRDD<String,Tuple2<String,String>> allRDD = transactionsRDD.union(usersRDD);

代码示例来源:origin: mahmoudparsian/data-algorithms-book

JavaPairRDD<String,Tuple2<String,String>> allRDD = transactionsRDD.union(usersRDD);

代码示例来源:origin: org.rcsb/mmtf-spark

/**
 * Join two RDDs together into one larger one.
 * @param datasetOne the first {@link JavaPairRDD} to join 
 * @param datasetTwo the second {@link JavaPairRDD} to join
 * @return the {@link JavaPairRDD} of the joined datasets
 */
public static JavaPairRDD<Text, BytesWritable> joinDatasets(JavaPairRDD<Text, BytesWritable> datasetOne,
    JavaPairRDD<Text, BytesWritable> datasetTwo){
  return datasetOne.union(datasetTwo);
}

代码示例来源:origin: scipr-lab/dizk

public void union(R1CSConstraintsRDD<FieldT> inputRDD) {
  A = A.union(inputRDD.A());
  B = B.union(inputRDD.B());
  C = C.union(inputRDD.C());
}

代码示例来源:origin: com.davidbracewell/mango

@Override
public MPairStream<T, U> union(@NonNull MPairStream<? extends T, ? extends U> other) {
 return new SparkPairStream<>(rdd.union(toPairRDD(other)));
}

代码示例来源:origin: co.cask.cdap/hydrator-spark-core2

public <K, V> JavaPairRDD<K, V> createRDD(JavaSparkExecutionContext sec, JavaSparkContext jsc, String sourceName,
                     Class<K> keyClass, Class<V> valueClass) {
 Set<String> inputNames = sourceInputs.get(sourceName);
 if (inputNames == null || inputNames.isEmpty()) {
  // should never happen if validation happened correctly at pipeline configure time
  throw new IllegalArgumentException(
   sourceName + " has no input. Please check that the source calls setInput at some input.");
 }
 JavaPairRDD<K, V> inputRDD = JavaPairRDD.fromJavaRDD(jsc.<Tuple2<K, V>>emptyRDD());
 for (String inputName : inputNames) {
  inputRDD = inputRDD.union(createInputRDD(sec, jsc, inputName, keyClass, valueClass));
 }
 return inputRDD;
}

代码示例来源:origin: DataSystemsLab/GeoSpark

this.distributedBackVectorImage = this.distributedBackVectorImage.union(distributedFontImageNoHeaderFooter);
this.distributedBackVectorImage = this.distributedBackVectorImage.sortByKey();
logger.info("[GeoSparkViz][JoinImage][Stop]");

代码示例来源:origin: com.facebook.presto.hive/hive-apache

rdd = prevRDD;
  } else {
   rdd = rdd.union(prevRDD);
 finalRDD = rdd;
} else {
 finalRDD = finalRDD.union(rdd);

代码示例来源:origin: DataSystemsLab/GeoSpark

svgHeaderFooter.add(new Tuple2<Integer, String>(2, g2.getSVGFooter()));
JavaPairRDD<Integer, String> distributedSVGHeaderFooter = sparkContext.parallelizePairs(svgHeaderFooter);
this.distributedVectorImage = this.distributedVectorImage.union(distributedSVGHeaderFooter);
this.distributedVectorImage = this.distributedVectorImage.sortByKey();
if (this.parallelRenderImage == true) {

代码示例来源:origin: scipr-lab/dizk

});
oneFullAssignment = oneFullAssignment.union(newARDD).union(newBRDD);
oneFullAssignment = oneFullAssignment.union(config.sparkContext().parallelizePairs(Collections.singletonList(new Tuple2<>((long) 0, one))));

代码示例来源:origin: scipr-lab/dizk

}).union(zeroIndexedA).reduceByKey(FieldT::add);
}).union(zeroIndexedB).reduceByKey(FieldT::add);
}).union(zeroIndexedC).reduceByKey(FieldT::add);

代码示例来源:origin: scipr-lab/dizk

}).union(popularAt).union(AtCoeffs).reduceByKey(FieldT::add, numPartitions).persist(config.storageLevel());
}).union(popularBt).reduceByKey(FieldT::add, numPartitions).persist(config.storageLevel());
}).union(popularCt).reduceByKey(FieldT::add, numPartitions).persist(config.storageLevel());

代码示例来源:origin: org.nd4j/canova-spark

JavaPairRDD<String, Tuple3<String,Integer,PortableDataStream>> merged = first2.union(second2);

代码示例来源:origin: org.datavec/datavec-spark

JavaPairRDD<String, Tuple3<String, Integer, PortableDataStream>> second2 =
        second.mapToPair(new PathToKeyFunction(1, converter2));
JavaPairRDD<String, Tuple3<String, Integer, PortableDataStream>> merged = first2.union(second2);

代码示例来源:origin: org.datavec/datavec-spark_2.11

JavaPairRDD<String, Tuple3<String, Integer, PortableDataStream>> second2 =
        second.mapToPair(new PathToKeyFunction(1, converter2));
JavaPairRDD<String, Tuple3<String, Integer, PortableDataStream>> merged = first2.union(second2);

代码示例来源:origin: scipr-lab/dizk

}).union(additionalA).union(zeroIndexedA).reduceByKey(FieldT::add);
    }).union(zeroIndexedB).reduceByKey(FieldT::add);
    }).union(zeroIndexedC).reduceByKey(FieldT::add);
config.endLog("Compute evaluation of polynomials A, B, and C, on set S.");
coefficientsH = DistributedFFT
    .radix2CosetInverseFFT(coefficientsH, multiplicativeGenerator, rows, cols, fieldFactory)
    .union(config.sparkContext()
        .parallelizePairs(Collections.singletonList(new Tuple2<>(domainSize, zero))));
config.endLog("Compute coefficients of polynomial H.");

代码示例来源:origin: seznam/euphoria

.union(rightPair)
.setName(operator.getName() + "::union-inputs")
.repartitionAndSortWithinPartitions(partitioner, comparator)

相关文章

微信公众号

最新文章

更多

JavaPairRDD类方法