本文整理了Java中org.apache.spark.api.java.JavaPairRDD.id()
方法的一些代码示例,展示了JavaPairRDD.id()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaPairRDD.id()
方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaPairRDD
类名称:JavaPairRDD
方法名:id
暂无
代码示例来源:origin: apache/drill
@Override
public JavaPairRDD<HiveKey, BytesWritable> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
JavaPairRDD<HiveKey, BytesWritable> result = shuffler.shuffle(input, numOfPartitions);
if (toCache) {
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
}
return result;
}
代码示例来源:origin: apache/hive
@Override
public JavaPairRDD<WritableComparable, Writable> transform(
JavaPairRDD<WritableComparable, Writable> input) {
Preconditions.checkArgument(input == null,
"AssertionError: MapInput doesn't take any input");
JavaPairRDD<WritableComparable, Writable> result;
if (toCache) {
result = hadoopRDD.mapToPair(new CopyFunction());
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
} else {
result = hadoopRDD;
}
result.setName(this.name);
return result;
}
代码示例来源:origin: apache/hive
@Override
public JavaPairRDD<HiveKey, BytesWritable> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
JavaPairRDD<HiveKey, BytesWritable> result = shuffler.shuffle(input, numOfPartitions);
if (toCache) {
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
}
return result.setName(this.name + " (" + edge.getShuffleType() + ", " + numOfPartitions +
(toCache ? ", cached)" : ")"));
}
代码示例来源:origin: apache/drill
@Override
public JavaPairRDD<WritableComparable, Writable> transform(
JavaPairRDD<WritableComparable, Writable> input) {
Preconditions.checkArgument(input == null,
"AssertionError: MapInput doesn't take any input");
JavaPairRDD<WritableComparable, Writable> result;
if (toCache) {
result = hadoopRDD.mapToPair(new CopyFunction());
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
} else {
result = hadoopRDD;
}
return result;
}
代码示例来源:origin: apache/drill
@Override
public JavaPairRDD<HiveKey, BytesWritable> shuffle(
JavaPairRDD<HiveKey, BytesWritable> input, int numPartitions) {
JavaPairRDD<HiveKey, BytesWritable> rdd;
if (totalOrder) {
if (numPartitions > 0) {
if (numPartitions > 1 && input.getStorageLevel() == StorageLevel.NONE()) {
input.persist(StorageLevel.DISK_ONLY());
sparkPlan.addCachedRDDId(input.id());
}
rdd = input.sortByKey(true, numPartitions);
} else {
rdd = input.sortByKey(true);
}
} else {
Partitioner partitioner = new HashPartitioner(numPartitions);
rdd = input.repartitionAndSortWithinPartitions(partitioner);
}
return rdd;
}
代码示例来源:origin: apache/hive
@Override
public JavaPairRDD<HiveKey, BytesWritable> shuffle(
JavaPairRDD<HiveKey, BytesWritable> input, int numPartitions) {
JavaPairRDD<HiveKey, BytesWritable> rdd;
if (totalOrder) {
if (numPartitions > 0) {
if (numPartitions > 1 && input.getStorageLevel() == StorageLevel.NONE()) {
input.persist(StorageLevel.DISK_ONLY());
sparkPlan.addCachedRDDId(input.id());
}
rdd = input.sortByKey(true, numPartitions);
} else {
rdd = input.sortByKey(true);
}
} else {
Partitioner partitioner = new HashPartitioner(numPartitions);
rdd = input.repartitionAndSortWithinPartitions(partitioner);
}
if (shuffleSerializer != null) {
if (rdd.rdd() instanceof ShuffledRDD) {
((ShuffledRDD) rdd.rdd()).setSerializer(shuffleSerializer);
}
}
return rdd;
}
代码示例来源:origin: jaibeermalik/searchanalytics-bigdata
@Override
public Void call(JavaPairRDD<Integer, String> rdd) {
String out = "\nSpark, Top 10 entries for stream id: " + rdd.id() + "\n";
for (Tuple2<Integer, String> t : rdd.take(10)) {
out = out + t.toString() + "\n";
}
System.out.println(out);
return null;
}
});
代码示例来源:origin: com.facebook.presto.hive/hive-apache
@Override
public JavaPairRDD<HiveKey, Iterable<BytesWritable>> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
JavaPairRDD<HiveKey, Iterable<BytesWritable>> result = shuffler.shuffle(input, numOfPartitions);
if (toCache) {
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
}
return result;
}
}
代码示例来源:origin: com.facebook.presto.hive/hive-apache
@Override
public JavaPairRDD<WritableComparable, Writable> transform(
JavaPairRDD<WritableComparable, Writable> input) {
Preconditions.checkArgument(input == null,
"AssertionError: MapInput doesn't take any input");
JavaPairRDD<WritableComparable, Writable> result;
if (toCache) {
result = hadoopRDD.mapToPair(new CopyFunction());
sparkPlan.addCachedRDDId(result.id());
result = result.persist(StorageLevel.MEMORY_AND_DISK());
} else {
result = hadoopRDD;
}
return result;
}
内容来源于网络,如有侵权,请联系作者删除!