RAb*_*ham 4 hadoop amazon-ec2 apache-spark
我试图通过EC2的spark脚本在EC2集群设置上使用distcp在Hadoop和Amazon S3之间进行复制
[root]# bin/hadoop distcp s3n://bucket/f1 hdfs:///user/root/
Run Code Online (Sandbox Code Playgroud)
我得到的错误是
INFO ipc.Client: Retrying connect to server: .. Already tried n time(s).
Copy failed: java.net.ConnectException: Call to ..my_server failed on connection excep\
tion: java.net.ConnectException: Connection refused
Run Code Online (Sandbox Code Playgroud)
Spark EC2脚本可能是故意的,不会启动JobTracker和TaskTracker服务.因此,在运行Spark EC2安装脚本后,为了启动服务,我跑了
{HADOOP_HOME}/bin/start-all.sh
Run Code Online (Sandbox Code Playgroud)
参考:感谢Brock Noland,网址为https://groups.google.com/a/cloudera.org/forum/#!topic/cdh-user/H6wAfdeLIJo
| 归档时间: |
|
| 查看次数: |
842 次 |
| 最近记录: |