org.apache.spark.SparkContext.wholeTextFiles()方法的使用及代码示例

x33g5p2x  于2022-01-30 转载在 其他  
字(2.9k)|赞(0)|评价(0)|浏览(116)

本文整理了Java中org.apache.spark.SparkContext.wholeTextFiles()方法的一些代码示例,展示了SparkContext.wholeTextFiles()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。SparkContext.wholeTextFiles()方法的具体详情如下:
包路径:org.apache.spark.SparkContext
类名称:SparkContext
方法名:wholeTextFiles

SparkContext.wholeTextFiles介绍

暂无

代码示例

代码示例来源:origin: uber/uberscriptquery

Tuple2<String, String>[] tuples = (Tuple2<String, String>[]) sparkSession.sparkContext().wholeTextFiles(query, 1).collect();
query = tuples[0]._2();
System.out.println("Query: " + query);

代码示例来源:origin: com.cerner.bunsen/bunsen-core

/**
 * Returns an RDD of bundles loaded from the given path.
 *
 * @param spark the spark session
 * @param path a path to a directory of FHIR Bundles
 * @param minPartitions a suggested value for the minimal number of partitions
 * @return an RDD of FHIR Bundles
 */
public JavaRDD<BundleContainer> loadFromDirectory(SparkSession spark,
  String path,
  int minPartitions) {
 return spark.sparkContext()
   .wholeTextFiles(path, minPartitions)
   .toJavaRDD()
   .map(new ToBundle(fhirVersion));
}

代码示例来源:origin: cerner/bunsen

/**
 * Returns an RDD of bundles loaded from the given path.
 *
 * @param spark the spark session
 * @param path a path to a directory of FHIR Bundles
 * @param minPartitions a suggested value for the minimal number of partitions
 * @return an RDD of FHIR Bundles
 */
public JavaRDD<BundleContainer> loadFromDirectory(SparkSession spark,
  String path,
  int minPartitions) {
 return spark.sparkContext()
   .wholeTextFiles(path, minPartitions)
   .toJavaRDD()
   .map(new ToBundle(fhirVersion));
}

代码示例来源:origin: cerner/bunsen

/**
 * Returns a dataset of ValueSet from the content stored at the given directory.
 */
protected Dataset<T> valueSetDatasetFromDirectory(String path) {
 JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
   .wholeTextFiles(path, 1)
   .toJavaRDD();
 return this.spark.createDataset(fileNamesAndContents
   .map(new ToValueSet(fhirVersion))
   .rdd(), valueSetEncoder);
}

代码示例来源:origin: com.cerner.bunsen/bunsen-core

/**
 * Returns a dataset of ValueSet from the content stored at the given directory.
 */
protected Dataset<T> valueSetDatasetFromDirectory(String path) {
 JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
   .wholeTextFiles(path, 1)
   .toJavaRDD();
 return this.spark.createDataset(fileNamesAndContents
   .map(new ToValueSet(fhirVersion))
   .rdd(), valueSetEncoder);
}

代码示例来源:origin: com.cerner.bunsen/bunsen-core

private Dataset<T> conceptMapsDatasetFromDirectory(String path) {
 JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
   .wholeTextFiles(path, 1)
   .toJavaRDD();
 return this.spark.createDataset(fileNamesAndContents
   .map(new ToConceptMap(fhirVersion))
   .rdd(), conceptMapEncoder);
}

代码示例来源:origin: cerner/bunsen

private Dataset<T> conceptMapsDatasetFromDirectory(String path) {
 JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
   .wholeTextFiles(path, 1)
   .toJavaRDD();
 return this.spark.createDataset(fileNamesAndContents
   .map(new ToConceptMap(fhirVersion))
   .rdd(), conceptMapEncoder);
}

相关文章