This page summarizes the steps to install Zeppelin version 0.7.3 on Windows 10 via Windows Subsystem for Linux (WSL).

Version 0.8.1

When running Zeppelin in Ubuntu, the server may pick up one host address that is not accessible, for example, and the the remote interpreter connection cannot be established successfully. The logic has changed in this version and I cannot find a fix yet without change the source code.

Thus, the page is still using version 0.7.3 to demo.


Follow either of the following pages to install WSL in a system or non-system drive on your Windows 10.

I also recommend you to install Hadoop 3.2.0 on your WSL following the second page.

After the above exercises, you WSL should already have JDK 1.8 installed.

Now let’s start to install Apache Zeppelin 0.7.3 in WSL.

Download binary package

Find out the latest Zeppelin binary package from the Download page:

For me the closest location is:

Run the following command to download the binary:


It may take a while to download.

Unzip the binary package

Unzip the binary package using the command below:

tar -xvzf zeppelin-0.7.3-bin-all.tgz -C ~/hadoop

Now Zeppelin is extracted to folder ~/hadoop/zeppelin-0.7.3-bin-all.

Start Zeppelin service

Start Zeppelin by running the following command:

cd ~/hadoop

cd zeppelin-0.7.3-bin-all/

bin/ start

You can see a successful log message like the following  in the zeppelin log file if the service is started:

({main}[main]:197) - Done, zeppelin server started

Open Zeppelin portal

In any browser, navigate to the following website: http://localhost:8080/.

8080 is the default port number. If you have changed it in Zeppelin configuration, remember to change the URL accordingly.

The UI looks like the following:


Stop Zeppelin service

Run the following command to stop the service:

bin/ stop

info Last modified by Raymond at 2 years ago * This page is subject to Site terms.

More from Kontext

Improve PySpark Performance using Pandas UDF with Apache Arrow

local_offer pyspark local_offer spark local_offer spark-2-x local_offer pandas

visibility 1781
thumb_up 4
access_time 7 months ago

Apache Arrow is an in-memory columnar data format that can be used in Spark to efficiently transfer data between JVM and Python processes. This currently is most beneficial to Python users that work with Pandas/NumPy data. In this article, ...

open_in_new Spark + PySpark

local_offer pyspark local_offer spark-2-x local_offer spark

visibility 2133
thumb_up 0
access_time 7 months ago

This article shows you how to read and write XML files in Spark. Sample XML file Create a sample XML file named test.xml with the following content: <?xml version="1.0"?> <data> <record id="1"> <rid>1</rid> <nam...

open_in_new Code snippets

local_offer pyspark local_offer spark-2-x local_offer spark local_offer python

visibility 2612
thumb_up 0
access_time 7 months ago

This article shows how to convert a Python dictionary list to a DataFrame in Spark using Python. Example dictionary list data = [{"Category": 'Category A', "ID": 1, "Value": 12.40}, {"Category": 'Category B', "ID": 2, "Value": 30.10}, {"Category": 'Category C', "...

open_in_new Spark + PySpark

local_offer pyspark local_offer spark-2-x local_offer spark

visibility 154
thumb_up 0
access_time 8 months ago

Sometime it is necessary to pass environment variables to Spark executors. To pass environment variable to executors, use setExecutorEnv function of SparkConf class. Code snippet In the following code snippet, an environment variable name ENV_NAME is set up with value ...

open_in_new Code snippets

info About author

comment Comments (0)

comment Add comment

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

No comments yet.

Dark theme mode

Dark theme mode is available on Kontext.

Learn more arrow_forward

Kontext Column

Created for everyone to publish data, programming and cloud related articles. Follow three steps to create your columns.

Learn more arrow_forward