线程“main”org.apache.hadoop.mapred.invalidinputexception中的异常:输入路径不存在:

yyyllmsg  于 2021-06-01  发布在  Hadoop
关注(0)|答案(1)|浏览(463)

我已经创建了一个输入目录并将示例文件放入其中。我还创建了一个输出目录。但是在执行mapreduce程序时,我遇到了以下错误。下面是执行mapreduce的命令

bin/hdfs dfs -mkdir /input
bin/hdfs dfs -put /home/biswajit/sample.txt /input/
bin/hadoop jar /usr/local/hadoop/hadoop-2.9.0/share/hadoop/mapreduce/units.jar com.hadoop.ProcessUnits /input/sample.txt /output

错误是

Exception in thread "main" org.apache.hadoop.mapred.InvalidInputException: Input path does not exist:**hdfs://localhost:54310/home/biswajit/input/sample.txt**
    at org.apache.hadoop.mapred.FileInputFormat.singleThreadedListStatus(FileInputFormat.java:294)
    at org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:236)
    at org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:322)
    at org.apache.hadoop.mapreduce.JobSubmitter.writeOldSplits(JobSubmitter.java:341)
    at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:333)
    at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:202)
    at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1570)
    at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1567)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
    at org.apache.hadoop.mapreduce.Job.submit(Job.java:1567)
    at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:576)
    at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:571)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1886)
    at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:571)
    at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:562)
    at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:871)
    at com.hadoop.ProcessUnits.main(ProcessUnits.java:96)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at org.apache.hadoop.util.RunJar.run(RunJar.java:239)
    at org.apache.hadoop.util.RunJar.main(RunJar.java:153)
omhiaaxx

omhiaaxx1#

$HADOOP_HOME/input hdfs上不存在。 $HADOOP_HOME 是本地文件系统上的bash变量。
您只为创建了一个目录 /input ,所以你可以 mkdir 如果您希望该命令按原样运行,或者在运行jar文件时需要删除该变量,则使用该变量的完整路径
只要 hdfs dfs -ls /input/* 显示一些文件,否则该命令看起来很好,但我不确定java类实际期望的输入是什么
注:两者之间有区别

hdfs://localhost:54310/home/biswajit/input

以及

hdfs://localhost:54310/input

更具体地说,hdfs没有 /home 文件夹,所以看起来不是没有运行伪分布式集群,就是您自己创建了该目录。

相关问题