
build/mvn -Pyarn -Phive -Phive-thriftserver -DskipTests clean package To enable Hive integration for Spark SQL along with its JDBC server and CLI,Īdd the -Phive and Phive-thriftserver profiles to your existing build options.īy default Spark will build with Hive 1.2.1 bindings. build/mvn -Pyarn -Phadoop-2.7 -Dhadoop.version=2.7.3 -DskipTests clean package build/mvn -Pyarn -DskipTests clean package You can enable the yarn profile and optionally set the yarn.version property if it is different If unset, Spark will build against Hadoop 2.6.X by default. You can specify the exact version of Hadoop to compile against through the hadoop.version property.

dev/make-distribution.sh -help Specifying the Hadoop Version and Enabling YARN This will build Spark distribution along with Python pip and R packages. dev/make-distribution.sh -name custom-spark -pip -r -tgz -Psparkr -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pyarn -Pkubernetes With Maven profile settings and so on like the direct Maven build. dev/make-distribution.sh in the project root directory. Spark Downloads page, and that is laid out so as To create a Spark distribution like those distributed by the As an example, one can build a version of Spark as follows. build/mvn execution acts as a pass through to the mvn call allowing easy transition from previous build methods. It honors any mvn binary if present already, however, will pull down its own copy of Scala and Zinc regardless to ensure proper version requirements are met.
HOW TO INSTALL APACHE SPARK ON LINUX DOWNLOAD
This script will automatically download and setup all necessary build requirements ( Maven, Scala, and Zinc) locally within the build/ directory itself. Spark now comes packaged with a self-contained Maven installation to ease building and deployment of Spark from source located under the build/ directory.

If you don’t add these parameters to MAVEN_OPTS, you may see errors and warnings like the following: Compiling 203 Scala sources and 9 Java sources to /Users/me/Development/spark/core/target/scala-2.12/classes. (The ReservedCodeCacheSize setting is optional but recommended.) You’ll need to configure Maven to use more memory than usual by setting MAVEN_OPTS: export MAVEN_OPTS="-Xmx2g -XX:ReservedCodeCacheSize=512m" Note that support for Java 7 was removed as of Spark 2.2.0. The Maven-based build is the build of reference for Apache Spark.īuilding Spark using Maven requires Maven 3.5.4 and Java 8.
