org.apache.spark.api.java.JavaSparkContext.hadoopRDD()方法的使用及代码示例

x33g5p2x  于2022-01-21 转载在 其他  
字(2.7k)|赞(0)|评价(0)|浏览(87)

本文整理了Java中org.apache.spark.api.java.JavaSparkContext.hadoopRDD()方法的一些代码示例,展示了JavaSparkContext.hadoopRDD()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaSparkContext.hadoopRDD()方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaSparkContext
类名称:JavaSparkContext
方法名:hadoopRDD

JavaSparkContext.hadoopRDD介绍

暂无

代码示例

代码示例来源:origin: apache/drill

@SuppressWarnings("unchecked")
private MapInput generateMapInput(SparkPlan sparkPlan, MapWork mapWork)
  throws Exception {
 JobConf jobConf = cloneJobConf(mapWork);
 Class ifClass = getInputFormat(jobConf, mapWork);
 JavaPairRDD<WritableComparable, Writable> hadoopRDD;
 if (mapWork.getNumMapTasks() != null) {
  jobConf.setNumMapTasks(mapWork.getNumMapTasks());
  hadoopRDD = sc.hadoopRDD(jobConf, ifClass,
    WritableComparable.class, Writable.class, mapWork.getNumMapTasks());
 } else {
  hadoopRDD = sc.hadoopRDD(jobConf, ifClass, WritableComparable.class, Writable.class);
 }
 // Caching is disabled for MapInput due to HIVE-8920
 MapInput result = new MapInput(sparkPlan, hadoopRDD, false/*cloneToWork.containsKey(mapWork)*/);
 return result;
}

代码示例来源:origin: apache/hive

@SuppressWarnings("unchecked")
private MapInput generateMapInput(SparkPlan sparkPlan, MapWork mapWork)
  throws Exception {
 JobConf jobConf = cloneJobConf(mapWork);
 Class ifClass = getInputFormat(jobConf, mapWork);
 sc.sc().setCallSite(CallSite.apply(mapWork.getName(), ""));
 JavaPairRDD<WritableComparable, Writable> hadoopRDD;
 if (mapWork.getNumMapTasks() != null) {
  jobConf.setNumMapTasks(mapWork.getNumMapTasks());
  hadoopRDD = sc.hadoopRDD(jobConf, ifClass,
    WritableComparable.class, Writable.class, mapWork.getNumMapTasks());
 } else {
  hadoopRDD = sc.hadoopRDD(jobConf, ifClass, WritableComparable.class, Writable.class);
 }
 boolean toCache = false/*cloneToWork.containsKey(mapWork)*/;
 String tables = mapWork.getAllRootOperators().stream()
     .filter(op -> op instanceof TableScanOperator)
     .map(ts -> ((TableScanDesc) ts.getConf()).getAlias())
     .collect(Collectors.joining(", "));
 String rddName = mapWork.getName() + " (" + tables + ", " + hadoopRDD.getNumPartitions() +
     (toCache ? ", cached)" : ")");
 // Caching is disabled for MapInput due to HIVE-8920
 MapInput result = new MapInput(sparkPlan, hadoopRDD, toCache, rddName, mapWork);
 return result;
}

代码示例来源:origin: com.facebook.presto.hive/hive-apache

@SuppressWarnings("unchecked")
private MapInput generateMapInput(SparkPlan sparkPlan, MapWork mapWork)
  throws Exception {
 JobConf jobConf = cloneJobConf(mapWork);
 Class ifClass = getInputFormat(jobConf, mapWork);
 JavaPairRDD<WritableComparable, Writable> hadoopRDD = sc.hadoopRDD(jobConf, ifClass,
   WritableComparable.class, Writable.class);
 // Caching is disabled for MapInput due to HIVE-8920
 MapInput result = new MapInput(sparkPlan, hadoopRDD, false/*cloneToWork.containsKey(mapWork)*/);
 return result;
}

相关文章

微信公众号

最新文章

更多