Spark is available in the Universe and can be installed by using either the GUI or the DC/OS CLI.


Default Installation

To install the DC/OS Apache Spark service, run the following command on the DC/OS CLI. This installs the Spark DC/OS service, Spark CLI, dispatcher, and, optionally, the history server. See Custom Installation to install the history server.

dcos package install spark

Go to the Services > Deployments tab of the DC/OS GUI to monitor the deployment. When it has finished deploying , visit Spark at http://<dcos-url>/service/spark/.

You can also install Spark via the DC/OS GUI.

Spark CLI

You can install the Spark CLI with this command. This is useful if you already have a Spark cluster running, but need the Spark CLI.

Important: If you install Spark via the DC/OS GUI, you must install the Spark CLI as a separate step from the DC/OS CLI.

dcos package install spark --cli

Custom Installation

You can customize the default configuration properties by creating a JSON options file and passing it to dcos package install --options. For example, to install the history server, create a file called options.json:

  "history-server": {
    "enabled": true

Install Spark with the configuration specified in the options.json file:

dcos package install --options=options.json spark

Tip: Run this command to see all configuration options:

dcos package describe spark --config

Customize Spark Distribution

DC/OS Apache Spark does not support arbitrary Spark distributions, but Mesosphere does provide multiple pre-built distributions, primarily used to select Hadoop versions.

To use one of these distributions, select your Spark distribution from here, then select the corresponding Docker image from here, then use those values to set the following configuration variables:

  "service": {
    "spark-dist-uri": "<spark-dist-uri>",
    "docker-image": "<docker-image>"

Minimal Installation

For development purposes, you can install Spark on a local DC/OS cluster. For this, you can use dcos-vagrant.

  1. Install DC/OS Vagrant:

    Install a minimal DC/OS Vagrant according to the instructions here.

  2. Install Spark:

    dcos package install spark
  3. Run a simple Job:

    dcos spark run --submit-args="--class org.apache.spark.examples.SparkPi"

NOTE: A limited resource environment such as DC/OS Vagrant restricts some of the features available in DC/OS Apache Spark. For example, unless you have enough resources to start up a 5-agent cluster, you will not be able to install DC/OS HDFS, and you thus won’t be able to enable the history server.

Also, a limited resource environment can restrict how you size your executors, for example with spark.executor.memory.

Multiple Installations

Installing multiple instances of the DC/OS Apache Spark package provides basic multi-team support. Each dispatcher displays only the jobs submitted to it by a given team, and each team can be assigned different resources.

To install multiple instances of the DC/OS Apache Spark package, set each to a unique name (e.g.: spark-dev) in your JSON configuration file during installation. For example, create a JSON options file name multiple.json:

  "service": {
    "name": "spark-dev"

Install Spark with the options file specified:

dcos package install --options=multiple.json spark

Alternatively, you can specify a Spark instance directly from the CLI. For example:

dcos config set spark.app_id spark-dev