-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I cannot read the output of a mapreduce job #4
Comments
On Mon, May 13, 2013 at 9:34 AM, fonsoim [email protected] wrote:
This is not documented, you are not supposed to do it, it could break in
Can you post the output of traceback() called immediately after this call? Antonio
|
Sorry for submitting the same problem in different places. I do not understand why I am not supposed to do this code. The versions of rmr2 and hadoop are 2.1.0 and 2.0.0, respectively. The code: data=to.dfs(1:10) The error:
/usr/lib/hadoop-hdfs/bin/hdfs: line 24: /usr/lib/hadoop-hdfs/bin/../libexec/hdfs-config.sh: No such file or directory
Caused by: org.apache.hadoop.ipc.RemoteException(java.io.FileNotFoundException): File does not exist: /tmp/RtmpzXyC7B/file34c6342d57ed/_logs
$key $val |
On Tue, May 21, 2013 at 1:21 AM, fonsoim [email protected] wrote:
The versions of rmr2 and hadoop are 2.1.0 and 2.0.0, respectively. How about the OS? Are you running windows? If so unfortunately it's not to.dfs(1:10, output = "/tmp/ls-test") At the shell prompt try hadoop dfs -ls /tmp/ls-test The first two errors that you get point to hadoop problems independent of R Antonio The code: data=to.dfs(1:10) The error: from.dfs(res) /usr/lib/hadoop-hdfs/bin/hdfs: line 24: This is where I suspect you are running the windows version. /usr/lib/hadoop-hdfs/bin/hdfs: line 162: exec: : not found
Caused by:
$key $val — |
The OS is Ubuntu 12.04 I did your experinent: to.dfs(1:10, output = "/tmp/ls-test") It works. The HDFS contains the file located in "/tmp/ls-test". Then, I list the file at the shell prompt. |
Maybe we have two problems here. One is that you have a configuration rmr2:::part.list("/tmp/ls-test") I am not sure what the connection between the two problems could be, but Antonio On Wed, May 22, 2013 at 1:27 AM, fonsoim [email protected] wrote:
|
hi, is this resolved? I have the same problem thanks |
I cannot read the output of a mapreduce job.
The code:
data=to.dfs(1:10)
res = mapreduce(input = data, map = function(k, v) cbind(v, 2*v))
print(res())
[1] "/tmp/Rtmpr5Xv1g/file34916a6426bf"
And then....
from.dfs(res)
Exception in thread "main" java.io.FileNotFoundException: File does not exist: /tmp/Rtmpr5Xv1g/file34916a6426bf/_logs
...
...
Finally,
hdfs.ls("/tmp/Rtmpr5Xv1g/file34916a6426bf")
permission owner group size modtime
1 -rw------- daniel supergroup 0 2013-05-13 18:24
2 drwxrwxrwt daniel supergroup 0 2013-05-13 18:23
3 -rw------- daniel supergroup 448 2013-05-13 18:24
4 -rw------- daniel supergroup 122 2013-05-13 18:23
file
1 /tmp/Rtmpr5Xv1g/file34916a6426bf/_SUCCESS
2 /tmp/Rtmpr5Xv1g/file34916a6426bf/_logs
3 /tmp/Rtmpr5Xv1g/file34916a6426bf/part-00000
4 /tmp/Rtmpr5Xv1g/file34916a6426bf/part-00001
I note that /tmp/Rtmpr5Xv1g/file34916a6426bf/_logs is a directory
Why does the program search the file "_logs" when it is a directory??????
Thanks in advance
Alfonso
The text was updated successfully, but these errors were encountered: