本文整理了Java中org.apache.spark.api.java.JavaRDD.cache()
方法的一些代码示例,展示了JavaRDD.cache()
的具体用法。这些代码示例主要来源于Github
/Stackoverflow
/Maven
等平台,是从一些精选项目中提取出来的代码,具有较强的参考意义,能在一定程度帮忙到你。JavaRDD.cache()
方法的具体详情如下:
包路径:org.apache.spark.api.java.JavaRDD
类名称:JavaRDD
方法名:cache
暂无
代码示例来源:origin: deeplearning4j/dl4j-examples
throw new IllegalStateException("No saved network parameters at " + f.getAbsolutePath());
Arrays.sort(content);
testDataPaths.cache();
代码示例来源:origin: OryxProject/oryx
newData.cache();
pastData.cache();
pastData.foreachPartition(p -> {});
代码示例来源:origin: databricks/learning-spark
trainingData.cache(); // Cache data since Logistic Regression is an iterative algorithm.
代码示例来源:origin: deeplearning4j/dl4j-examples
JavaRDD<String> trainDataPaths = SparkUtils.listPaths(sc, trainDataPathRootDir);
JavaRDD<String> testDataPaths = totalExamplesTest <= 0 ? null : listPathsSubset(sc, testDataPathRootDir, totalExamplesTest, 12345);
trainDataPaths.cache();
if(testDataPaths != null)
testDataPaths.cache();
代码示例来源:origin: OryxProject/oryx
parsedTestRDD.cache();
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void testJavaJdbcRDD() throws Exception {
JavaRDD<Integer> rdd = JdbcRDD.create(
sc,
() -> DriverManager.getConnection("jdbc:derby:target/JavaJdbcRDDSuiteDb"),
"SELECT DATA FROM FOO WHERE ? <= ID AND ID <= ?",
1, 100, 1,
r -> r.getInt(1)
).cache();
Assert.assertEquals(100, rdd.count());
Assert.assertEquals(Integer.valueOf(10100), rdd.reduce((i1, i2) -> i1 + i2));
}
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void testJavaJdbcRDD() throws Exception {
JavaRDD<Integer> rdd = JdbcRDD.create(
sc,
() -> DriverManager.getConnection("jdbc:derby:target/JavaJdbcRDDSuiteDb"),
"SELECT DATA FROM FOO WHERE ? <= ID AND ID <= ?",
1, 100, 1,
r -> r.getInt(1)
).cache();
Assert.assertEquals(100, rdd.count());
Assert.assertEquals(Integer.valueOf(10100), rdd.reduce((i1, i2) -> i1 + i2));
}
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void testJavaJdbcRDD() throws Exception {
JavaRDD<Integer> rdd = JdbcRDD.create(
sc,
() -> DriverManager.getConnection("jdbc:derby:target/JavaJdbcRDDSuiteDb"),
"SELECT DATA FROM FOO WHERE ? <= ID AND ID <= ?",
1, 100, 1,
r -> r.getInt(1)
).cache();
Assert.assertEquals(100, rdd.count());
Assert.assertEquals(Integer.valueOf(10100), rdd.reduce((i1, i2) -> i1 + i2));
}
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void testGetPersistentRDDs() {
java.util.Map<Integer, JavaRDD<?>> cachedRddsMap = sc.getPersistentRDDs();
assertTrue(cachedRddsMap.isEmpty());
JavaRDD<String> rdd1 = sc.parallelize(Arrays.asList("a", "b")).setName("RDD1").cache();
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("c", "d")).setName("RDD2").cache();
cachedRddsMap = sc.getPersistentRDDs();
assertEquals(2, cachedRddsMap.size());
assertEquals("RDD1", cachedRddsMap.get(0).name());
assertEquals("RDD2", cachedRddsMap.get(1).name());
}
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void testGetPersistentRDDs() {
java.util.Map<Integer, JavaRDD<?>> cachedRddsMap = sc.getPersistentRDDs();
assertTrue(cachedRddsMap.isEmpty());
JavaRDD<String> rdd1 = sc.parallelize(Arrays.asList("a", "b")).setName("RDD1").cache();
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("c", "d")).setName("RDD2").cache();
cachedRddsMap = sc.getPersistentRDDs();
assertEquals(2, cachedRddsMap.size());
assertEquals("RDD1", cachedRddsMap.get(0).name());
assertEquals("RDD2", cachedRddsMap.get(1).name());
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void testGetPersistentRDDs() {
java.util.Map<Integer, JavaRDD<?>> cachedRddsMap = sc.getPersistentRDDs();
assertTrue(cachedRddsMap.isEmpty());
JavaRDD<String> rdd1 = sc.parallelize(Arrays.asList("a", "b")).setName("RDD1").cache();
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("c", "d")).setName("RDD2").cache();
cachedRddsMap = sc.getPersistentRDDs();
assertEquals(2, cachedRddsMap.size());
assertEquals("RDD1", cachedRddsMap.get(0).name());
assertEquals("RDD2", cachedRddsMap.get(1).name());
}
代码示例来源:origin: OryxProject/oryx
parsedRDD.cache();
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(x -> 1.0 * x).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x))
.cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: org.apache.spark/spark-core_2.10
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(Integer::doubleValue).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x)).cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(Integer::doubleValue).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x)).cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: org.apache.spark/spark-core
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(x -> 1.0 * x).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x))
.cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(Integer::doubleValue).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x)).cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: org.apache.spark/spark-core_2.11
@Test
public void map() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
JavaDoubleRDD doubles = rdd.mapToDouble(x -> 1.0 * x).cache();
doubles.collect();
JavaPairRDD<Integer, Integer> pairs = rdd.mapToPair(x -> new Tuple2<>(x, x))
.cache();
pairs.collect();
JavaRDD<String> strings = rdd.map(Object::toString).cache();
strings.collect();
}
代码示例来源:origin: deeplearning4j/dl4j-examples
processed.cache();
代码示例来源:origin: mahmoudparsian/data-algorithms-book
static JavaRDD<Vector> getFeatureizedData(String wikiData, JavaSparkContext context) {
JavaRDD<Vector> data = context.textFile(wikiData).map(new Function<String, Vector>() {
@Override
public Vector call(String arg0) throws Exception {
return Util.buildVector(arg0, "\t");
}
}).cache();
return data;
}
内容来源于网络,如有侵权,请联系作者删除!