Usage / Deployment Instruction
Step 1: Open Putty for SSH
Step 2: Open Putty and Type <instanceID>> at “Host Name”
Step 3: Open Conncetion->SSH->Auth tab from Left Side Area
Step 4: Click on browse button and select ppk file for Instance and then click on Open
Step 5: Type "ubuntu" as user name Password auto taken from PPK file
Step 6: Use following Linux command to start Hadoop
Step 6.1: sudo vi /etc/hosts
Take the Private Ip address from your machine as per the below screenshot and then replace the second line of your command screen with that Private ip address


Step 6.2: ssh-keygen -t rsa -P ""
This command is used to generate the ssh key.

Step 6.3: cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
This command is used to move the generated ssh key to the desired location

Step 6.4: ssh localhost

Step 6.5: hdfs namenode –format
You have to write “yes” when it prompts you – Are you sure you want to continue?

Step 6.6: start-all.sh

Step 6.7: After the above command executes successfully, you should check the below urls in the browser -
http://<instanceId>:8088

http://<instanceId>:50070

http://<instanceId>:50090

Step 7: Use following Linux command to start Scala and Spark
Step 7.1: cd spark-2.1.0/
Step 7.2: ./bin/spark-shell

Step 7.3: You can check the spark by going on to the following url in your browser –
http://<instanceId>:4040

Step 7.4: Now you can execute your scala programs as below –
