This page summarizes the installation guides about big data tools on Windows through Windows Subsystem for Linux (WSL).

Install Hadoop 3.2.0 on Windows 10 using Windows Subsystem for Linux (WSL)

A framework that allows for distributed processing of the large data sets across clusters of computers.

View detail

Apache Spark 2.4.3 Installation on Windows 10 using Windows Subsystem for Linux

A unified analytics engine for large-scale both batch and streaming data processing.

View detail

Install Zeppelin 0.7.3 on Windows 10 using Windows Subsystem for Linux (WSL)

Web-based notebook that enables data-driven, interactive data analytics and collaborative documents with SQL, Scala, R, Python, Java, C#, F# and more.

View detail

Sqoop Installation on Windows 10 using Windows Subsystem for Linux

SQL to Hadoop: transferring data between Hadoop and structured datastores such as relational databases.

View detail

WSL installation

Install Windows Subsystem for Linux on a Non-System Drive

Follow this page if you want to install WSL on a non-system drive.

View detail
info Last modified by Raymond at 2 years ago * This page is subject to Site terms.

More from Kontext

local_offer hadoop local_offer hive local_offer Java

visibility 398
thumb_up 1
access_time 3 months ago

When I was configuring Hive 3.0.0 in Hadoop 3.2.1 environment, I encountered the following error: Exception in thread "main" java.lang.NoSuchMethodError: com.google.common.base.Preconditions.checkArgument(ZLjava/lang/String;Ljava/lang/Object;)V Ro...

open_in_new Hadoop

local_offer hdfs local_offer hadoop local_offer windows

visibility 354
thumb_up 0
access_time 4 months ago

Network Attached Storage are commonly used in many enterprises where files are stored remotely on those servers.  They typically provide access to files using network file sharing protocols such as  ...

open_in_new Hadoop

local_offer hive local_offer hdfs

visibility 115
thumb_up 0
access_time 5 months ago

In Hive, there are two types of tables can be created - internal and external table. Internal tables are also called managed tables. Different features are available to different types. This article lists some of the common differences.  Internal table By default, Hive creates ...

open_in_new Hadoop

Schema Merging (Evolution) with Parquet in Spark and Hive

local_offer parquet local_offer pyspark local_offer spark-2-x local_offer hive local_offer hdfs

visibility 2188
thumb_up 0
access_time 6 months ago

Schema evolution is supported by many frameworks or data serialization systems such as Avro, Orc, Protocol Buffer and Parquet. With schema evolution, one set of data can be stored in multiple files with different but compatible schema. In Spark, Parquet data source can detect and merge sch...

open_in_new Spark + PySpark

info About author

comment Comments (0)

comment Add comment

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

No comments yet.

Dark theme mode

Dark theme mode is available on Kontext.

Learn more arrow_forward

Kontext Column

Created for everyone to publish data, programming and cloud related articles. Follow three steps to create your columns.


Learn more arrow_forward