将Dataframe写入hdfs,连接被拒绝

xeufq47z  于 2021-05-29  发布在  Hadoop
关注(0)|答案(1)|浏览(693)

我试图运行来自精通ApacheSpark2.x的书中的示例。

scala> val df = sc.parallelize(Array(1,2,3)).toDF
df: org.apache.spark.sql.DataFrame = [value: int]

我是spark world的新手,但我想Dataframe应该保存到hdfs中

scala> df.write.json("hdfs://localhost:9000/tmp/account.json")
java.net.ConnectException: Call From miki/127.0.1.1 to localhost:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused

我和dfsadmin联系过了

hadoop dfsadmin -safemode enter
WARNING: Use of this script to execute dfsadmin is deprecated.
WARNING: Attempting to execute replacement "hdfs dfsadmin" instead.

safemode: FileSystem file:/// is not an HDFS file system

jps输出

miki@miki:~$ jps
13798 Jps
10906 SparkSubmit

如何解决这个问题?

92dk7w1h

92dk7w1h1#

根据您的jps输出,您没有运行读写hdfs(namenode、datanode、resourcemanager)所需的hadoop守护进程。确保在机器上运行start yarn和start dfs,以使hdfs正常运行。

相关问题