org.apache.spark.SparkContext.runJob()方法的使用及代码示例

x33g5p2x  于2022-01-30 转载在 其他  
字(2.8k)|赞(0)|评价(0)|浏览(145)

本文整理了Java中org.apache.spark.SparkContext.runJob()方法的一些代码示例,展示了SparkContext.runJob()的具体用法。这些代码示例主要来源于Github/Stackoverflow/Maven等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。SparkContext.runJob()方法的具体详情如下:
包路径:org.apache.spark.SparkContext
类名称:SparkContext
方法名:runJob

SparkContext.runJob介绍

暂无

代码示例

代码示例来源:origin: com.basho.riak/spark-riak-connector

public void saveToRiakTS(
    RiakConnector connector,
    String bucketType,
    String bucketName,
    WriteConf writeConf,
    WriteDataMapperFactory<T, RowDef> factory
) {
  RiakWriter<T, RowDef> writer = RiakWriter.tsWriter(connector, bucketType, bucketName, writeConf, factory);
  rdd.sparkContext().runJob(rdd, new JobFunc<>(writer), JavaApiHelper.getClassTag(Void.class));
}

代码示例来源:origin: com.basho.riak/spark-riak-connector-java

public void saveToRiakTS(
    RiakConnector connector,
    String bucketType,
    String bucketName,
    WriteConf writeConf,
    WriteDataMapperFactory<T, RowDef> factory
) {
  RiakWriter<T, RowDef> writer = RiakWriter.tsWriter(connector, bucketType, bucketName, writeConf, factory);
  rdd.sparkContext().runJob(rdd, new JobFunc<>(writer), JavaApiHelper.getClassTag(Void.class));
}

代码示例来源:origin: org.qcri.rheem/rheem-spark

int tid = rand.nextInt(partitionSize); // sample item inside the partition
List<Integer> partitions = Collections.singletonList(pid);
Object samples = sparkContext.runJob(inputRdd.rdd(),
    new PartitionSampleFunction(tid, ((tid + sampleSize))),
    (scala.collection.Seq) JavaConversions.asScalaBuffer(partitions),
      sparkContext.runJob(inputRdd.rdd(),
          new PartitionSampleListFunction(list),
          (scala.collection.Seq) JavaConversions.asScalaBuffer(partitions),

代码示例来源:origin: com.stratio.deep/deep-cassandra

mappedRDD.context().runJob(mappedRDD,
    new AbstractSerializableFunction2<TaskContext, Iterator<Tuple2<Cells, Cells>>, Integer>() {

代码示例来源:origin: org.qcri.rheem/rheem-spark

pars.add(partitionID);
Object samples = sparkContext.runJob(shuffledRDD.rdd(),
    new TakeSampleFunction(tupleID, tupleID + sampleSize),
    (scala.collection.Seq) JavaConversions.asScalaBuffer(pars), true, scala.reflect.ClassTag$.MODULE$.apply(List.class));

代码示例来源:origin: org.wso2.carbon.analytics/org.wso2.carbon.analytics.spark.core

private void writeDataFrameToDAL(DataFrame data) {
  if (this.preserveOrder) {
    logDebug("Inserting data with order preserved! Each partition will be written using separate jobs.");
    for (int i = 0; i < data.rdd().partitions().length; i++) {
      data.sqlContext().sparkContext().runJob(data.rdd(),
                          new AnalyticsWritingFunction(this.tenantId, this.tableName, data.schema(),
                                         this.globalTenantAccess, this.schemaString, this.primaryKeys, this.mergeFlag,
                                         this.recordStore, this.recordBatchSize), CarbonScalaUtils.getNumberSeq(i, i + 1),
                          false, ClassTag$.MODULE$.Unit());
    }
  } else {
    data.foreachPartition(new AnalyticsWritingFunction(this.tenantId, this.tableName, data.schema(),
                              this.globalTenantAccess, this.schemaString, this.primaryKeys, this.mergeFlag,
                              this.recordStore, this.recordBatchSize));
  }
}

相关文章