本文整理了Java中org.apache.spark.SparkContext.wholeTextFiles()
方法的一些代码示例,展示了SparkContext.wholeTextFiles()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。SparkContext.wholeTextFiles()
方法的具体详情如下:
包路径:org.apache.spark.SparkContext
类名称:SparkContext
方法名:wholeTextFiles
暂无
代码示例来源:origin: uber/uberscriptquery
Tuple2<String, String>[] tuples = (Tuple2<String, String>[]) sparkSession.sparkContext().wholeTextFiles(query, 1).collect();
query = tuples[0]._2();
System.out.println("Query: " + query);
代码示例来源:origin: com.cerner.bunsen/bunsen-core
/**
* Returns an RDD of bundles loaded from the given path.
*
* @param spark the spark session
* @param path a path to a directory of FHIR Bundles
* @param minPartitions a suggested value for the minimal number of partitions
* @return an RDD of FHIR Bundles
*/
public JavaRDD<BundleContainer> loadFromDirectory(SparkSession spark,
String path,
int minPartitions) {
return spark.sparkContext()
.wholeTextFiles(path, minPartitions)
.toJavaRDD()
.map(new ToBundle(fhirVersion));
}
代码示例来源:origin: cerner/bunsen
/**
* Returns an RDD of bundles loaded from the given path.
*
* @param spark the spark session
* @param path a path to a directory of FHIR Bundles
* @param minPartitions a suggested value for the minimal number of partitions
* @return an RDD of FHIR Bundles
*/
public JavaRDD<BundleContainer> loadFromDirectory(SparkSession spark,
String path,
int minPartitions) {
return spark.sparkContext()
.wholeTextFiles(path, minPartitions)
.toJavaRDD()
.map(new ToBundle(fhirVersion));
}
代码示例来源:origin: cerner/bunsen
/**
* Returns a dataset of ValueSet from the content stored at the given directory.
*/
protected Dataset<T> valueSetDatasetFromDirectory(String path) {
JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
.wholeTextFiles(path, 1)
.toJavaRDD();
return this.spark.createDataset(fileNamesAndContents
.map(new ToValueSet(fhirVersion))
.rdd(), valueSetEncoder);
}
代码示例来源:origin: com.cerner.bunsen/bunsen-core
/**
* Returns a dataset of ValueSet from the content stored at the given directory.
*/
protected Dataset<T> valueSetDatasetFromDirectory(String path) {
JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
.wholeTextFiles(path, 1)
.toJavaRDD();
return this.spark.createDataset(fileNamesAndContents
.map(new ToValueSet(fhirVersion))
.rdd(), valueSetEncoder);
}
代码示例来源:origin: com.cerner.bunsen/bunsen-core
private Dataset<T> conceptMapsDatasetFromDirectory(String path) {
JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
.wholeTextFiles(path, 1)
.toJavaRDD();
return this.spark.createDataset(fileNamesAndContents
.map(new ToConceptMap(fhirVersion))
.rdd(), conceptMapEncoder);
}
代码示例来源:origin: cerner/bunsen
private Dataset<T> conceptMapsDatasetFromDirectory(String path) {
JavaRDD<Tuple2<String,String>> fileNamesAndContents = this.spark.sparkContext()
.wholeTextFiles(path, 1)
.toJavaRDD();
return this.spark.createDataset(fileNamesAndContents
.map(new ToConceptMap(fhirVersion))
.rdd(), conceptMapEncoder);
}
内容来源于网络,如有侵权,请联系作者删除!