本文整理了Java中org.apache.spark.SparkContext.runJob()
方法的一些代码示例,展示了SparkContext.runJob()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。SparkContext.runJob()
方法的具体详情如下:
包路径:org.apache.spark.SparkContext
类名称:SparkContext
方法名:runJob
暂无
代码示例来源:origin: com.basho.riak/spark-riak-connector
public void saveToRiakTS(
RiakConnector connector,
String bucketType,
String bucketName,
WriteConf writeConf,
WriteDataMapperFactory<T, RowDef> factory
) {
RiakWriter<T, RowDef> writer = RiakWriter.tsWriter(connector, bucketType, bucketName, writeConf, factory);
rdd.sparkContext().runJob(rdd, new JobFunc<>(writer), JavaApiHelper.getClassTag(Void.class));
}
代码示例来源:origin: com.basho.riak/spark-riak-connector-java
public void saveToRiakTS(
RiakConnector connector,
String bucketType,
String bucketName,
WriteConf writeConf,
WriteDataMapperFactory<T, RowDef> factory
) {
RiakWriter<T, RowDef> writer = RiakWriter.tsWriter(connector, bucketType, bucketName, writeConf, factory);
rdd.sparkContext().runJob(rdd, new JobFunc<>(writer), JavaApiHelper.getClassTag(Void.class));
}
代码示例来源:origin: org.qcri.rheem/rheem-spark
int tid = rand.nextInt(partitionSize); // sample item inside the partition
List<Integer> partitions = Collections.singletonList(pid);
Object samples = sparkContext.runJob(inputRdd.rdd(),
new PartitionSampleFunction(tid, ((tid + sampleSize))),
(scala.collection.Seq) JavaConversions.asScalaBuffer(partitions),
sparkContext.runJob(inputRdd.rdd(),
new PartitionSampleListFunction(list),
(scala.collection.Seq) JavaConversions.asScalaBuffer(partitions),
代码示例来源:origin: com.stratio.deep/deep-cassandra
mappedRDD.context().runJob(mappedRDD,
new AbstractSerializableFunction2<TaskContext, Iterator<Tuple2<Cells, Cells>>, Integer>() {
代码示例来源:origin: org.qcri.rheem/rheem-spark
pars.add(partitionID);
Object samples = sparkContext.runJob(shuffledRDD.rdd(),
new TakeSampleFunction(tupleID, tupleID + sampleSize),
(scala.collection.Seq) JavaConversions.asScalaBuffer(pars), true, scala.reflect.ClassTag$.MODULE$.apply(List.class));
代码示例来源:origin: org.wso2.carbon.analytics/org.wso2.carbon.analytics.spark.core
private void writeDataFrameToDAL(DataFrame data) {
if (this.preserveOrder) {
logDebug("Inserting data with order preserved! Each partition will be written using separate jobs.");
for (int i = 0; i < data.rdd().partitions().length; i++) {
data.sqlContext().sparkContext().runJob(data.rdd(),
new AnalyticsWritingFunction(this.tenantId, this.tableName, data.schema(),
this.globalTenantAccess, this.schemaString, this.primaryKeys, this.mergeFlag,
this.recordStore, this.recordBatchSize), CarbonScalaUtils.getNumberSeq(i, i + 1),
false, ClassTag$.MODULE$.Unit());
}
} else {
data.foreachPartition(new AnalyticsWritingFunction(this.tenantId, this.tableName, data.schema(),
this.globalTenantAccess, this.schemaString, this.primaryKeys, this.mergeFlag,
this.recordStore, this.recordBatchSize));
}
}
内容来源于网络,如有侵权,请联系作者删除!