


spark-shell – it should run scala version.then reload bash file – source ~/.bashrc.
#CAN I INSTALL SPARK ON WINDOWS UPDATE#
Update PATHS by updating file ~/.bashrc:.Rename spark-2.3.0-bin-hadoop2.7 to spark – mv spark-2.3.0-bin-hadoop2.7 spark.Unzip the tar – tar xvfz spark-2.3.0-bin-hadoop2.7.tgz.Now download proper version of Spark(First go to and then copy the link address) – wget.echo “alias python=python36” > ~/.bashrc.Setup alias for python command and update the ~/.bashrc.To install JDK8- yum install -y java-1.8.0-openjdk-devel.To install JRE8- yum install -y java-1.8.0-openjdk.Type and Enter quit() to exit the spark.If you get successful count then you succeeded in installing Spark with Python on Windows.Type and Enter myRDD= sc.textFile(“README.md”).Look for README.md or CHANGES.txt in that folder.Select environment for Windows(32 bit or 64 bit) and download 3.5 version canopy and install.Right-click Windows menu –> select Control Panel –> System and Security –> System –> Advanced System Settings –> Environment Variables.execute command – winutils.exe chmod 777 \tmp\hive from that folder.Edit the file to change log level to ERROR – for log4j.rootCategory.Rename file conf\ file to log4j.properties.Now lets unzip the tar file using WinRar or 7Z and copy the content of the unzipped folder to a new folder D:\Spark.Lets select Spark version 2.3.0 and click on the download link.Install JDK, but make sure your installation folder should not have spaces in path name e.g d:\jdk8.Select your environment ( Windows x86 or 圆4).
