- User running the scripts should be in the sudoers list.
- Setup hostname and fqdn
- Make sure to review/update the config files
Copy this repo:
$ git clone git://github.com/ezhaar/spark-installer
Run the install script:
$ cd spark-installer;./install
- Go grab a coffee.
- Created a dedicated group and user for hadoop (hduser:hadoop)
- Installed Jdk-1.7 and set Java Path
- Downloaded, installed and configured hadoop-2.4.0 in
/home/hduser/DataAnalaysis/hadoop
and update PATH. - Downloaded, installed and configured Scala-2.10.3.
- Downloaded, installed and configured Spark-1.1.0 with YARN.
Switch to the newly created hduser and cd to home directory:
$ sudo su hduser;cd
Update the hostname in
$HADOOP_DIR/conf/core-site.xml
:$ sed -i s/XXXX/myHostname/g $HADOOP_CONF_DIR/core-site.xml
Format hadoop's namenode:
$ hdfs namenode -format
Start HDFS processes:
$ start-dfs.sh
Start Yarn Processes:
$ start-yarn.sh
Create the initial directories:
$ hdfs dfs -mkdir /user;hdfs dfs -mkdir /user/hduser