Hadoop Streaming - Unable to find file error

2019-01-15 13:54发布

I am trying to run a hadoop-streaming python job.

bin/hadoop jar contrib/streaming/hadoop-0.20.1-streaming.jar 
-D stream.non.zero.exit.is.failure=true 
-input /ixml 
-output /oxml 
-mapper scripts/mapper.py 
-file scripts/mapper.py 
-inputreader "StreamXmlRecordReader,begin=channel,end=/channel" 
-jobconf mapred.reduce.tasks=0 

I made sure mapper.py has all the permissions. It errors out saying

Caused by: java.io.IOException: Cannot run program "mapper.py":     
error=2, No such file or directory
    at java.lang.ProcessBuilder.start(ProcessBuilder.java:460)
    at org.apache.hadoop.streaming.PipeMapRed.configure(PipeMapRed.java:214)
... 19 more
Caused by: java.io.IOException: error=2, No such file or directory
    at java.lang.UNIXProcess.forkAndExec(Native Method)
    at java.lang.UNIXProcess.(UNIXProcess.java:53)
    at java.lang.ProcessImpl.start(ProcessImpl.java:91)
    at java.lang.ProcessBuilder.start(ProcessBuilder.java:453)

I tried copying mapper.py to hdfs and give the same hdfs://localhost/mapper.py link, that does not work too! Any thoughts on how to fix this bug?.

8条回答
Bombasti
2楼-- · 2019-01-15 14:28

I just received the same error when my mapper returns a null or empty string. So I had to do a check for the value:

try:
    # Skip over any errors

    word = words[18].strip()

        if (len(word) == 0):
            word = "UKNOWN"

    print '%s\t%s' % (word, 1)

except Value:
    pass
查看更多
乱世女痞
3楼-- · 2019-01-15 14:34

Your problem most likely is that python executable does not exist on the slaves (where TaskTracker is running). Java will give the same error message.

Install it everywhere where it's used. Un your file you can use shebang as you probably already do:

#!/usr/bin/python -O
rest
of
the
code

Make sure that the path after the shebang is the same where python is installed on the TaskTrackers.

查看更多
孤傲高冷的网名
4楼-- · 2019-01-15 14:35

One other sneaky thing can cause this. If your line-endings on the script are DOS-style, then your first line (the "shebang line") may look like this to the naked eye:

#!/usr/bin/python

...my code here...

but its bytes look like this to the kernel when it tries to execute your script:

% od -a myScript.py
0000000   #   !   /   u   s   r   /   b   i   n   /   p   y   t   h   o
0000020   n  cr  nl  cr  nl   .   .   .   m   y  sp   c   o   d   e  sp
0000040   h   e   r   e   .   .   .  cr  nl

It's looking for an executable called "/usr/bin/python\r", which it can't find, so it dies with "No such file or directory".

This bit me today, again, so I had to write it down somewhere on SO.

查看更多
劫难
5楼-- · 2019-01-15 14:37

Looking at the example on the HadoopStreaming wiki page, it seems that you should change

-mapper scripts/mapper.py 
-file scripts/mapper.py 

to

-mapper mapper.py 
-file scripts/mapper.py 

since "shipped files go to the working directory". You might also need to specify the python interpreter directly:

-mapper /path/to/python mapper.py 
-file scripts/mapper.py 
查看更多
乱世女痞
6楼-- · 2019-01-15 14:37

File not found error sometimes does not means "File not found", instead it means "Cannot execute this script".

Knowing this I solved problems like this, when you are facing with issues ( no java ) on streaming I suggest you to follow this check list:

  1. Does the scripts run? Don't start is using the interpreter i.e. python myScript.py make it executable at start it as ./myScript.py this is the way the streaming will call your script.
  2. use -verbose to see what is going into the jar which will be deployed into the container, sometime this help.
  3. Inside the containers scripts are symlink not real files.
  4. Files which are moved using -file are not in folders. -mapper folder/script.py or -reducer folder/script.py are treat as script.py
  5. Containers and anything inside them are deleted after the job completes, if you want to see what is happening into a container move it into HDFS, I.E: replacing the mapper or the reducer with a .sh script which does the work.

This checklist helped me a lot, I hope can be useful also for you.

Here follows the classic log with the ambiguous error message.

It's true, it cannot run the program.

Caused by: java.io.IOException: Cannot run program "/hadoop/yarn/local/usercache/root/appcache/application_1475243242823_0007/container_1475243242823_0007_01_000004/./reducer.py": 
error=2, No such file or directory

It's the reason the lie.

    at java.lang.ProcessBuilder.start(ProcessBuilder.java:1047)
    at org.apache.hadoop.streaming.PipeMapRed.configure(PipeMapRed.java:209)
    ... 15 more

Read this:

Caused by: java.io.IOException: error=2, No such file or directory

It's a lie, file does exists if -verbose shows it into the packaging list.

    at java.lang.UNIXProcess.forkAndExec(Native Method)
    at java.lang.UNIXProcess.<init>(UNIXProcess.java:187)
    at java.lang.ProcessImpl.start(ProcessImpl.java:130)
    at java.lang.ProcessBuilder.start(ProcessBuilder.java:1028)
查看更多
Luminary・发光体
7楼-- · 2019-01-15 14:38

Does your mapper.py have execute permission on it ? If not then you need it.

chmod a+x scripts/mapper.py

Hadoop forks and runs the the script before it writes/reads to std so you need to give it execute permission to run.

查看更多
登录 后发表回答