Steps to install Spark in local mode:
Install Java 7 or later.
To test java installation is complete, open command prompt type java
and hit enter.
If you receive a message 'Java' is not recognized as an internal or external command.
You need to configure your environment variables, JAVA_HOME
and PATH
to point to the path of jdk.
Set SCALA_HOME
in Control Panel\System and Security\System
goto "Adv System settings" and add %SCALA_HOME%\bin
in PATH variable in environment variables.
Install Python 2.6 or later from Python Download link.
Download SBT. Install it and set SBT_HOME
as an environment variable with value as <<SBT PATH>>
.
Download winutils.exe
from HortonWorks repo or git repo. Since we don't have a local Hadoop installation on Windows we have to download winutils.exe
and place it in a bin
directory under a created Hadoop
home directory.
Set HADOOP_HOME = <<Hadoop home directory>>
in environment variable.
We will be using a pre-built Spark package, so choose a Spark pre-built package for Hadoop Spark download. Download and extract it.
Set SPARK_HOME
and add %SPARK_HOME%\bin
in PATH variable in environment variables.
Run command: spark-shell
Open http://localhost:4040/
in a browser to see the SparkContext web UI.