搭建Spark前要做的配置

安装Spark

1.下载VM

http://sw.bos.baidu.com/sw-search-sp/software/a08321b624453/VMware_workstation_full_12.5.2.exe

2.安装vm,使用ctrl+alt切换内外鼠标

3.下载ubuntu

http://old-releases.ubuntu.com/releases/14.04.1/ubuntu-14.04.1-server-amd64.iso

4.安装ubuntu

5.下载ubuntu桌面控制见博客

http://bdxnote.blog.163.com/blog/static/84442352013222112540421/

6.安装vm的vmtools,方便鼠标切换。

7.安装SSh

sshserversudoapt-getinstallopenssh-server

详见:http://jingyan.baidu.com/article/9c69d48fb9fd7b13c8024e6b.html

8.Linux下安装jdk

wget--no-check-certificate--no-cookies--header"Cookie:oraclelicense=accept-securebackup-cookie"http://download.oracle.com/otn-pub/java/jdk/8u121-b13/e9e7ea248e2c4826b92b3f075a80e441/jdk-8u121-linux-x64.tar.gz

设置环境变量

5.下载

Hadoop

6.下载Spark

7.启动spark

相关推荐