visibility 13,285 comment 4 access_time 3 years ago languageEnglish
more_vert
Raymond Raymond
articleArticles 549
codeCode 3
imageDiagrams 49
descriptionNotebooks 0
chat_bubble_outlineThreads 8
commentComments 268
loyaltyKontext Points 6058
account_circleProfile

Install Hadoop 3.3.0 on Linux

This article provides step-by-step guidance to install Hadoop 3.3.0 on Linux such as Debian, Ubuntu, Red Hat, openSUSE, etc.  Hadoop 3.3.0 was released on July 14 2020. It is the first release of Apache Hadoop 3.3 line. There are significant changes compared with Hadoop 3.2.0, such as ...
info Last modified by Raymond 3 years ago
thumb_up 8

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts

comment Comments
4 months ago link more_vert
Raymond Raymond
articleArticles 549
codeCode 3
imageDiagrams 49
descriptionNotebooks 0
chat_bubble_outlineThreads 8
commentComments 268
loyaltyKontext Points 6058
account_circleProfile
#1768 Re: Install Hadoop 3.3.0 on Linux

That link might not be available now. Please download directly from the release website: Apache Hadoop.


format_quote

person Attapol access_time 4 months ago
Re: Install Hadoop 3.3.0 on Linux

2 years ago link more_vert
Raymond Raymond
articleArticles 549
codeCode 3
imageDiagrams 49
descriptionNotebooks 0
chat_bubble_outlineThreads 8
commentComments 268
loyaltyKontext Points 6058
account_circleProfile
#1472 Re: Install Hadoop 3.3.0 on Linux

I'm glad the article is helping. If you don't specify the paths, it will use default paths. 

Good luck with your configurations for a 100TB cluster.

2 years ago link more_vert
Raymond Raymond
articleArticles 549
codeCode 3
imageDiagrams 49
descriptionNotebooks 0
chat_bubble_outlineThreads 8
commentComments 268
loyaltyKontext Points 6058
account_circleProfile
#1470 Re: Install Hadoop 3.3.0 on Linux

Hi,

It means that you need to ensure the account that runs Hadoop daemon services has full access to the namenode and datanode directories

<property>
     <name>dfs.namenode.name.dir</name>
     <value>/data/dfs/namespace_logs_330</value>
   </property>
   <property>
     <name>dfs.datanode.data.dir</name>
     <value>/data/dfs/data_330</value>
   </property>

For example, if you use your account to run the process, please your Linux account in WSL has write permissions to folder /data/dfs/namespace_logs_330 and /data/dfs/data_330 for the above HDFS configuration (in file etc/hadoop/core-site.xml). 

For connection refused, it is usually due to the SSH services were not started correctly. As mentioned in the article, please try the following commands in WSL bash command line and then restart your Hadoop services:

sudo apt-get install ssh
sudo service ssh restart 

Sometimes it might work by just simply restarting your Windows systems.

recommendMore from Kontext