本文整理了Java中org.apache.spark.SparkContext.<init>()
方法的一些代码示例,展示了SparkContext.<init>()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。SparkContext.<init>()
方法的具体详情如下:
包路径:org.apache.spark.SparkContext
类名称:SparkContext
方法名:<init>
暂无
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void scalaSparkContext() {
List<String> jars = List$.MODULE$.empty();
Map<String, String> environment = Map$.MODULE$.empty();
new SparkContext(new SparkConf().setMaster("local").setAppName("name")).stop();
new SparkContext("local", "name", new SparkConf()).stop();
new SparkContext("local", "name").stop();
new SparkContext("local", "name", "sparkHome").stop();
new SparkContext("local", "name", "sparkHome", jars).stop();
new SparkContext("local", "name", "sparkHome", jars, environment).stop();
}
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void scalaSparkContext() {
List<String> jars = List$.MODULE$.empty();
Map<String, String> environment = Map$.MODULE$.empty();
new SparkContext(new SparkConf().setMaster("local").setAppName("name")).stop();
new SparkContext("local", "name", new SparkConf()).stop();
new SparkContext("local", "name").stop();
new SparkContext("local", "name", "sparkHome").stop();
new SparkContext("local", "name", "sparkHome", jars).stop();
new SparkContext("local", "name", "sparkHome", jars, environment).stop();
}
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
public static void main(String[] args) throws Exception {
assertNotEquals(0, args.length);
assertEquals(args[0], "hello");
new SparkContext().stop();
synchronized (LOCK) {
LOCK.notifyAll();
}
}
代码示例来源:origin: org.apache.spark/spark-core
public static void main(String[] args) throws Exception {
assertNotEquals(0, args.length);
assertEquals(args[0], "hello");
new SparkContext().stop();
synchronized (LOCK) {
LOCK.notifyAll();
}
}
代码示例来源:origin: Impetus/Kundera
@Override
public void initialize(Map<String, Object> puProperties)
{
reader = new SparkEntityReader(kunderaMetadata);
setExternalProperties(puProperties);
initializePropertyReader();
PersistenceUnitMetadata pum = kunderaMetadata.getApplicationMetadata().getPersistenceUnitMetadata(
getPersistenceUnit());
sparkconf = new SparkConf(true);
configureClientProperties(pum);
sparkContext = new SparkContext(sparkconf);
sqlContext = new HiveContext(sparkContext);
}
代码示例来源:origin: stackoverflow.com
val sc = new SparkContext();
val fs = FileSystem.get(sc.hadoopConfig)
val snapshotPath = fs.createSnapshot("path to createsnapshot of","snapshot name")
.....
.....
if (condition satisfied) {
fs.deleteSnapshot(snapshotPath,"snapshot name")
}
代码示例来源:origin: stackoverflow.com
import org.apache.spark.SparkContext
import org.apache.spark.SparkContext._
import scala.io.Source
import java.io.File
val sc = new SparkContext("your_spark_master", "your_application_name")
// Read from a local dir
val sourceRecords = new File("your_dir_path").listFiles().flatMap(file => Source.fromFile(file).getLines.toList).filter(_.length > 0)
val sourceRecordsBc = sc.broadcast(sourceRecords)
val inputRDD = sc.textFile("hdfs://your_input_path")
val outputRDD = inputRDD.filter { _.length > 0 }
.map { line => (line, line.split(",")) }
.filter { _._2.size == 21 }
.flatMap {
case (line, target) => {
val pos_last_name = 0
for (
inputLine <- sourceRecordsBc.value if inputLine != line;
val source = inputLine.split(",") if source.size == 21 && source(pos_last_name) == target(pos_last_name)
) yield (inputLine, line)
}
}
outputRDD.saveAsTextFile("hdfs://your_output_path")
代码示例来源:origin: stackoverflow.com
import org.apache.spark.SparkContext
import org.apache.spark.sql.SQLContext
import org.apache.phoenix.spark._
val sc = new SparkContext("local", "phoenix-test")
val sqlContext = new SQLContext(sc)
val df = sqlContext.load(
"org.apache.phoenix.spark",
Map("table" -> "TABLE1", "zkUrl" -> "phoenix-server:2181")
)
df
.filter(df("COL1") === "test_row_1" && df("ID") === 1L)
.select(df("ID"))
.show
代码示例来源:origin: stackoverflow.com
scala.collection.immutable.Range rangeMscala =
scala.collection.immutable.Range$.MODULE$.apply(1, 10);
SparkContext sc = new SparkContext();
RDD dataSetMscala =
sc.parallelize(rangeMscala, 3, scala.reflect.ClassTag$.MODULE$.Object());
代码示例来源:origin: bhdrkn/Java-Examples
@Bean(destroyMethod = "stop")
public SparkContext sparkContext() {
final SparkConf conf = new SparkConf()
.setAppName("MLSparkService")
.setMaster("local");
return new SparkContext(conf);
}
代码示例来源:origin: com.stratio.meta/meta-core
/**
* Initialize the DeepSparkContext adding the required jars if the deployment is not local.
* @param config The {@link com.stratio.meta.core.engine.EngineConfig}
* @return A new context.
*/
private DeepSparkContext initializeDeep(EngineConfig config){
//DeepSparkContext result = new DeepSparkContext(config.getSparkMaster(), config.getJobName());
SparkConf sparkConf = new SparkConf().set("spark.driver.port",
"0")//String.valueOf(StreamingUtils.findFreePort()))
.set("spark.ui.port",
"0");//String.valueOf(StreamingUtils.findFreePort()));
DeepSparkContext result = new DeepSparkContext(new SparkContext(config.getSparkMaster(), config.getJobName(), sparkConf));
if(!config.getSparkMaster().toLowerCase().startsWith("local")){
for(String jar : config.getJars()){
result.addJar(jar);
}
}
return result;
}
代码示例来源:origin: amidst/toolbox
public static void main(String[] args) throws Exception {
String arffFile = "datasets/simulated/syntheticData.arff";
String sparkFile = "datasets/simulated/syntheticData.json";
SparkConf conf = new SparkConf().setAppName("SparkLink!").setMaster("local");
SparkContext sc = new SparkContext(conf);
SQLContext sqlContext = new SQLContext(sc);
JavaSparkContext jsc = new JavaSparkContext(sc);
ARFFtoSparkFormat(arffFile, sparkFile, "json", sqlContext, jsc);
}
代码示例来源:origin: com.holdenkarau/spark-testing-base_2.11
@Before
public void runBefore() {
initialized = (_sc != null);
if (!initialized) {
_sc = new SparkContext(conf());
_jsc = new JavaSparkContext(_sc);
beforeAllTestCasesHook();
}
}
代码示例来源:origin: org.talend.components/components-adapter-beam
public Pipeline createPipeline() {
SparkContextOptions sparkOpts = options.as(SparkContextOptions.class);
SparkConf conf = new SparkConf();
conf.setAppName(appName);
conf.setMaster("local[2]");
conf.set("spark.driver.allowMultipleContexts", "true");
JavaSparkContext jsc = new JavaSparkContext(new SparkContext(conf));
sparkOpts.setProvidedSparkContext(jsc);
sparkOpts.setUsesProvidedSparkContext(true);
sparkOpts.setRunner(SparkRunner.class);
return Pipeline.create(sparkOpts);
}
}
代码示例来源:origin: com.holdenkarau/spark-testing-base_2.10
@Before
public void runBefore() {
initialized = (_sc != null);
if (!initialized) {
_sc = new SparkContext(conf());
_jsc = new JavaSparkContext(_sc);
beforeAllTestCasesHook();
}
}
代码示例来源:origin: amidst/toolbox
public static void main(String[] args) throws Exception {
SparkConf conf = new SparkConf().setAppName("SLink!").setMaster("local");
SparkContext sc = new SparkContext(conf);
SQLContext sqlContext = new SQLContext(sc);
//Path to dataset
String path ="datasets/simulated/WI_samples.json";
//Create an AMIDST object for managing the data
DataSpark dataSpark = DataSparkLoader.open(sqlContext, path);
//Print all the instances in the dataset
dataSpark.collectDataStream()
.forEach(
dataInstance -> System.out.println(dataInstance)
);
}
}
代码示例来源:origin: Talend/components
public Pipeline createPipeline() {
SparkContextOptions sparkOpts = options.as(SparkContextOptions.class);
SparkConf conf = new SparkConf();
conf.setAppName(appName);
conf.setMaster("local[2]");
conf.set("spark.driver.allowMultipleContexts", "true");
JavaSparkContext jsc = new JavaSparkContext(new SparkContext(conf));
sparkOpts.setProvidedSparkContext(jsc);
sparkOpts.setUsesProvidedSparkContext(true);
sparkOpts.setRunner(SparkRunner.class);
return Pipeline.create(sparkOpts);
}
}
代码示例来源:origin: io.zipkin.dependencies/zipkin-dependencies-cassandra
public void run() {
long microsLower = day * 1000;
long microsUpper = (day * 1000) + TimeUnit.DAYS.toMicros(1) - 1;
log.info("Running Dependencies job for {}: {} ≤ Span.timestamp {}", dateStamp, microsLower,
microsUpper);
SparkContext sc = new SparkContext(conf);
List<DependencyLink> links = javaFunctions(sc)
.cassandraTable(keyspace, "traces")
.spanBy(ROW_TRACE_ID, Long.class)
.flatMapValues(new CassandraRowsToDependencyLinks(logInitializer, microsLower, microsUpper))
.values()
.mapToPair(LINK_TO_PAIR)
.reduceByKey(MERGE_LINK)
.values()
.collect();
sc.stop();
saveToCassandra(links);
}
代码示例来源:origin: amidst/toolbox
public static void main(String[] args) throws Exception {
SparkConf conf = new SparkConf().setAppName("SparkLink!").setMaster("local");;
SparkContext sc = new SparkContext(conf);
SQLContext sqlContext = new SQLContext(sc);
代码示例来源:origin: jgperrin/net.jgp.labs.spark
private void start() {
SparkConf conf = new SparkConf().setAppName("Checkpoint").setMaster(
"local[*]");
SparkContext sparkContext = new SparkContext(conf);
// We need to specify where Spark will save the checkpoint file. It can be
// an HDFS location.
sparkContext.setCheckpointDir("/tmp");
SparkSession spark = SparkSession.builder().appName("Checkpoint").master(
"local[*]").getOrCreate();
String filename = "data/tuple-data-file.csv";
Dataset<Row> df1 = spark.read().format("csv").option("inferSchema", "true")
.option("header", "false")
.load(filename);
System.out.println("DF #1 - step #1: simple dump of the dataframe");
df1.show();
System.out.println("DF #2 - step #2: same as DF #1 - step #1");
Dataset<Row> df2 = df1.checkpoint(false);
df2.show();
df1 = df1.withColumn("x", df1.col("_c0"));
System.out.println(
"DF #1 - step #2: new column x, which is the same as _c0");
df1.show();
System.out.println("DF #2 - step #2: no operation was done on df2");
df2.show();
}
}
内容来源于网络,如有侵权,请联系作者删除!