Big Data Tools on Windows via Windows Subsystem for Linux (WSL)

access_time 2 years ago visibility1365 comment 0

This page summarizes the installation guides about big data tools on Windows through Windows Subsystem for Linux (WSL).

Install Hadoop 3.2.0 on Windows 10 using Windows Subsystem for Linux (WSL)

A framework that allows for distributed processing of the large data sets across clusters of computers.

View detail

Apache Spark 2.4.3 Installation on Windows 10 using Windows Subsystem for Linux

A unified analytics engine for large-scale both batch and streaming data processing.

View detail

Install Zeppelin 0.7.3 on Windows 10 using Windows Subsystem for Linux (WSL)

Web-based notebook that enables data-driven, interactive data analytics and collaborative documents with SQL, Scala, R, Python, Java, C#, F# and more.

View detail

Sqoop Installation on Windows 10 using Windows Subsystem for Linux

SQL to Hadoop: transferring data between Hadoop and structured datastores such as relational databases.

View detail

WSL installation

Install Windows Subsystem for Linux on a Non-System Drive

Follow this page if you want to install WSL on a non-system drive.

View detail
info Last modified by Raymond 2 years ago copyright This page is subject to Site terms.
Like this article?
Share on

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

Follow Kontext

Get our latest updates on LinkedIn or Twitter.

Want to publish your article on Kontext?

Learn more

More from Kontext

Apache Spark 3.0.1 Installation on Linux or WSL Guide
visibility 24
thumb_up 0
access_time 26 days ago

This article provides step by step guide to install the latest version of Apache Spark 3.0.1 on a UNIX alike system (Linux) or Windows Subsystem for Linux (WSL). These instructions can be applied to Ubuntu, Debian, Red Hat, OpenSUSE, etc.  If you are planning to configure Spark 3.0.1 on WSL ...

visibility 3292
thumb_up 0
access_time 3 years ago

This page continues with the following documentation about configuring a Hadoop multi-nodes cluster via adding a new edge node to configure administration or client tools. Configure Hadoop 3.1.0 in a Multi Node Cluster In this page, I’m going to show you how to add a edge node into the ...

visibility 3072
thumb_up 0
access_time 6 months ago

CSV is a commonly used data format. Spark provides rich APIs to load files from HDFS as data frame.  This page provides examples about how to load CSV from HDFS using Spark. If you want to read a local CSV file in Python, refer to this page  Python: Load / Read Multiline CSV File   ...