By using this site, you acknowledge that you have read and understand our Cookie policy, Privacy policy and Terms .

This page summarizes the installation guides about big data tools on Windows through Windows Subsystem for Linux (WSL).

Install Hadoop 3.2.0 on Windows 10 using Windows Subsystem for Linux (WSL)

A framework that allows for distributed processing of the large data sets across clusters of computers.

View detail

Apache Spark 2.4.3 Installation on Windows 10 using Windows Subsystem for Linux

A unified analytics engine for large-scale both batch and streaming data processing.

View detail

Install Zeppelin 0.7.3 on Windows 10 using Windows Subsystem for Linux (WSL)

Web-based notebook that enables data-driven, interactive data analytics and collaborative documents with SQL, Scala, R, Python, Java, C#, F# and more.

View detail

Sqoop Installation on Windows 10 using Windows Subsystem for Linux

SQL to Hadoop: transferring data between Hadoop and structured datastores such as relational databases.

View detail

WSL installation

Install Windows Subsystem for Linux on a Non-System Drive

Follow this page if you want to install WSL on a non-system drive.

View detail
info Last modified by Raymond at 11 months ago * This page is subject to Site terms.

More from Kontext

local_offer hdfs local_offer hadoop local_offer windows

visibility 57
thumb_up 0
access_time 30 days ago

Network Attached Storage are commonly used in many enterprises where files are stored remotely on those servers.  They typically provide access to files using network file sharing protocols such as  ...

open_in_new View open_in_new Hadoop

local_offer hive local_offer hdfs

visibility 60
thumb_up 0
access_time 2 months ago

In Hive, there are two types of tables can be created - internal and external table. Internal tables are also called managed tables. Different features are available to different types. This article lists some of the common differences.  Internal table By default, Hive creates ...

open_in_new View open_in_new Hadoop

Schema Merging (Evolution) with Parquet in Spark and Hive

local_offer parquet local_offer pyspark local_offer spark-2-x local_offer hive local_offer hdfs

visibility 296
thumb_up 0
access_time 3 months ago

Schema evolution is supported by many frameworks or data serialization systems such as Avro, Orc, Protocol Buffer and Parquet. With schema evolution, one set of data can be stored in multiple files with different but compatible schema. In Spark, Parquet data source can detect and merge sch...

open_in_new View open_in_new Spark + PySpark

Fix for Hadoop 3.2.1 namenode format issue on Windows 10

local_offer windows10 local_offer hadoop local_offer hdfs

visibility 247
thumb_up 0
access_time 3 months ago

Issue When installing Hadoop 3.2.1 on Windows 10,  you may encounter the following error when trying to format HDFS  namnode: ERROR namenode.NameNode: Failed to start namenode. The error happens when running the following comm...

open_in_new View open_in_new Hadoop

info About author

Kontext dark theme mode

Dark theme mode

Dark theme mode is available on Kontext.

Learn more arrow_forward
Kontext Column

Kontext Column

Created for everyone to publish data, programming and cloud related articles. Follow three steps to create your columns.

Learn more arrow_forward
info Follow us on Twitter to get the latest article updates. Follow us