org.apache.spark.api.java.JavaPairRDD.id()方法的使用及代码示例

x33g5p2x  于2022-01-21 转载在 其他  
字(4.6k)|赞(0)|评价(0)|浏览(89)

本文整理了Java中org.apache.spark.api.java.JavaPairRDD.id()方法的一些代码示例,展示了JavaPairRDD.id()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaPairRDD.id()方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaPairRDD
类名称:JavaPairRDD
方法名:id

JavaPairRDD.id介绍

暂无

代码示例

代码示例来源:origin: apache/drill

@Override
public JavaPairRDD<HiveKey, BytesWritable> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
 JavaPairRDD<HiveKey, BytesWritable> result = shuffler.shuffle(input, numOfPartitions);
 if (toCache) {
  sparkPlan.addCachedRDDId(result.id());
  result = result.persist(StorageLevel.MEMORY_AND_DISK());
 }
 return result;
}

代码示例来源:origin: apache/hive

@Override
public JavaPairRDD<WritableComparable, Writable> transform(
  JavaPairRDD<WritableComparable, Writable> input) {
 Preconditions.checkArgument(input == null,
   "AssertionError: MapInput doesn't take any input");
 JavaPairRDD<WritableComparable, Writable> result;
 if (toCache) {
  result = hadoopRDD.mapToPair(new CopyFunction());
  sparkPlan.addCachedRDDId(result.id());
  result = result.persist(StorageLevel.MEMORY_AND_DISK());
 } else {
  result = hadoopRDD;
 }
 result.setName(this.name);
 return result;
}

代码示例来源:origin: apache/hive

@Override
public JavaPairRDD<HiveKey, BytesWritable> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
 JavaPairRDD<HiveKey, BytesWritable> result = shuffler.shuffle(input, numOfPartitions);
 if (toCache) {
  sparkPlan.addCachedRDDId(result.id());
  result = result.persist(StorageLevel.MEMORY_AND_DISK());
 }
 return result.setName(this.name + " (" + edge.getShuffleType() + ", " + numOfPartitions +
     (toCache ? ", cached)" : ")"));
}

代码示例来源:origin: apache/drill

@Override
public JavaPairRDD<WritableComparable, Writable> transform(
  JavaPairRDD<WritableComparable, Writable> input) {
 Preconditions.checkArgument(input == null,
   "AssertionError: MapInput doesn't take any input");
 JavaPairRDD<WritableComparable, Writable> result;
 if (toCache) {
  result = hadoopRDD.mapToPair(new CopyFunction());
  sparkPlan.addCachedRDDId(result.id());
  result = result.persist(StorageLevel.MEMORY_AND_DISK());
 } else {
  result = hadoopRDD;
 }
 return result;
}

代码示例来源:origin: apache/drill

@Override
public JavaPairRDD<HiveKey, BytesWritable> shuffle(
  JavaPairRDD<HiveKey, BytesWritable> input, int numPartitions) {
 JavaPairRDD<HiveKey, BytesWritable> rdd;
 if (totalOrder) {
  if (numPartitions > 0) {
   if (numPartitions > 1 && input.getStorageLevel() == StorageLevel.NONE()) {
    input.persist(StorageLevel.DISK_ONLY());
    sparkPlan.addCachedRDDId(input.id());
   }
   rdd = input.sortByKey(true, numPartitions);
  } else {
   rdd = input.sortByKey(true);
  }
 } else {
  Partitioner partitioner = new HashPartitioner(numPartitions);
  rdd = input.repartitionAndSortWithinPartitions(partitioner);
 }
 return rdd;
}

代码示例来源:origin: apache/hive

@Override
public JavaPairRDD<HiveKey, BytesWritable> shuffle(
  JavaPairRDD<HiveKey, BytesWritable> input, int numPartitions) {
 JavaPairRDD<HiveKey, BytesWritable> rdd;
 if (totalOrder) {
  if (numPartitions > 0) {
   if (numPartitions > 1 && input.getStorageLevel() == StorageLevel.NONE()) {
    input.persist(StorageLevel.DISK_ONLY());
    sparkPlan.addCachedRDDId(input.id());
   }
   rdd = input.sortByKey(true, numPartitions);
  } else {
   rdd = input.sortByKey(true);
  }
 } else {
  Partitioner partitioner = new HashPartitioner(numPartitions);
  rdd = input.repartitionAndSortWithinPartitions(partitioner);
 }
 if (shuffleSerializer != null) {
  if (rdd.rdd() instanceof ShuffledRDD) {
   ((ShuffledRDD) rdd.rdd()).setSerializer(shuffleSerializer);
  }
 }
 return rdd;
}

代码示例来源:origin: jaibeermalik/searchanalytics-bigdata

@Override
  public Void call(JavaPairRDD<Integer, String> rdd) {
    String out = "\nSpark, Top 10 entries for stream id: " + rdd.id() + "\n";
    for (Tuple2<Integer, String> t : rdd.take(10)) {
      out = out + t.toString() + "\n";
    }
    System.out.println(out);
    return null;
  }
});

代码示例来源:origin: com.facebook.presto.hive/hive-apache

@Override
 public JavaPairRDD<HiveKey, Iterable<BytesWritable>> transform(JavaPairRDD<HiveKey, BytesWritable> input) {
  JavaPairRDD<HiveKey, Iterable<BytesWritable>> result = shuffler.shuffle(input, numOfPartitions);
  if (toCache) {
   sparkPlan.addCachedRDDId(result.id());
   result = result.persist(StorageLevel.MEMORY_AND_DISK());
  }
  return result;
 }
}

代码示例来源:origin: com.facebook.presto.hive/hive-apache

@Override
public JavaPairRDD<WritableComparable, Writable> transform(
  JavaPairRDD<WritableComparable, Writable> input) {
 Preconditions.checkArgument(input == null,
   "AssertionError: MapInput doesn't take any input");
 JavaPairRDD<WritableComparable, Writable> result;
 if (toCache) {
  result = hadoopRDD.mapToPair(new CopyFunction());
  sparkPlan.addCachedRDDId(result.id());
  result = result.persist(StorageLevel.MEMORY_AND_DISK());
 } else {
  result = hadoopRDD;
 }
 return result;
}

相关文章

微信公众号

最新文章

更多

JavaPairRDD类方法