arrow_back Install Hadoop 3.2.1 on Windows 10 Step by Step Guide

comment Comments
Raymond Raymond #307 access_time 4 years ago more_vert

Hi Tim,

Have you checked YARN web portal to see if you can see the Spark application is submitted successfully? You should be able to find more details there too (assuming you are run Spark with master set as yarn).

I’m working today and will try to replicate what you did in my machine when I am off work.



format_quote

Comment is deleted or blocked.

Raymond Raymond #298 access_time 4 years ago more_vert

Hi Saad,

Refer to the Reference section on this page: Default Ports Used by Hadoop Services (HDFS, MapReduce, YARN). It has the links to the official documentation about all the parameters you can configure in HDFS and YARN. It also shows the default values for each configurations.

For different versions of Hadoop, the default values might be different. 

format_quote

person Saad access_time 4 years ago

Hi,

http://localhost:9870/dfshealth.html#tab-overview

http://localhost:9864/datanode.html
these 2 links were not opening once i reached till end, then i started changing values in hdfs-site.xml 

to some other paths locations in E drive and then i think got lost.

Today when i run start-dfs.cmd then data and name node start without any error and i can see above 2 urls without any error. 

Thanks for quick reply.

Can you also guide me where can i find and change ports values like 8088,9870 etc.  

 Thanks again for this tutorial.

Regards,

Saad

S Saad United #297 access_time 4 years ago more_vert

Hi,

http://localhost:9870/dfshealth.html#tab-overview

http://localhost:9864/datanode.html
these 2 links were not opening once i reached till end, then i started changing values in hdfs-site.xml 

to some other paths locations in E drive and then i think got lost.

Today when i run start-dfs.cmd then data and name node start without any error and i can see above 2 urls without any error. 

Thanks for quick reply.

Can you also guide me where can i find and change ports values like 8088,9870 etc.  

 Thanks again for this tutorial.

Regards,

Saad

format_quote

person Raymond access_time 4 years ago

Hi Saad,

I don't see any error message in the log you pasted.

Can you please be more specific about the errors you encounterred.

For formatting namenode, it is correct to expect the namenode daemon to shutdown after the format is done. We will start all the HDFS and YAN daemons at the end. 

Raymond Raymond #296 access_time 4 years ago more_vert

Hi Saad,

I don't see any error message in the log you pasted.

Can you please be more specific about the errors you encounterred.

For formatting namenode, it is correct to expect the namenode daemon to shutdown after the format is done. We will start all the HDFS and YAN daemons at the end. 

format_quote

person Saad access_time 4 years ago

hello Raymond,


I learning about Hadoop and was following your detailed information on windows installation.




2020-04-27 22:32:04,347 INFO namenode.NameNode: createNameNode [-format]


Formatting using clusterid: CID-1d0c51aa-5dde-446b-99c1-3997255160fa


2020-04-27 22:32:05,369 INFO namenode.FSEditLog: Edit logging is async:true


2020-04-27 22:32:05,385 INFO namenode.FSNamesystem: KeyProvider: null


2020-04-27 22:32:05,387 INFO namenode.FSNamesystem: fsLock is fair: true


2020-04-27 22:32:05,388 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false


2020-04-27 22:32:05,428 INFO namenode.FSNamesystem: fsOwner             = saad (auth:SIMPLE)


2020-04-27 22:32:05,431 INFO namenode.FSNamesystem: supergroup          = supergroup


2020-04-27 22:32:05,431 INFO namenode.FSNamesystem: isPermissionEnabled = true


2020-04-27 22:32:05,432 INFO namenode.FSNamesystem: HA Enabled: false


2020-04-27 22:32:05,535 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling


2020-04-27 22:32:05,554 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000


2020-04-27 22:32:05,554 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true


2020-04-27 22:32:05,562 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000


2020-04-27 22:32:05,563 INFO blockmanagement.BlockManager: The block deletion will start around 2020 Apr 27 22:32:05


2020-04-27 22:32:05,566 INFO util.GSet: Computing capacity for map BlocksMap


2020-04-27 22:32:05,566 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,568 INFO util.GSet: 2.0% max memory 889 MB = 17.8 MB


2020-04-27 22:32:05,568 INFO util.GSet: capacity      = 2^21 = 2097152 entries


2020-04-27 22:32:05,579 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled


2020-04-27 22:32:05,580 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false


2020-04-27 22:32:05,588 INFO Configuration.deprecation: No unit for dfs.namenode.safemode.extension(30000) assuming MILLISECONDS


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.9990000128746033


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 30000


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: defaultReplication         = 1


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: maxReplication             = 512


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: minReplication             = 1


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: redundancyRecheckInterval  = 3000ms


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: encryptDataTransfer        = false


2020-04-27 22:32:05,593 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000


2020-04-27 22:32:05,646 INFO namenode.FSDirectory: GLOBAL serial map: bits=29 maxEntries=536870911


2020-04-27 22:32:05,646 INFO namenode.FSDirectory: USER serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,647 INFO namenode.FSDirectory: GROUP serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,647 INFO namenode.FSDirectory: XATTR serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,664 INFO util.GSet: Computing capacity for map INodeMap


2020-04-27 22:32:05,664 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,664 INFO util.GSet: 1.0% max memory 889 MB = 8.9 MB


2020-04-27 22:32:05,665 INFO util.GSet: capacity      = 2^20 = 1048576 entries


2020-04-27 22:32:05,666 INFO namenode.FSDirectory: ACLs enabled? false


2020-04-27 22:32:05,666 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true


2020-04-27 22:32:05,667 INFO namenode.FSDirectory: XAttrs enabled? true


2020-04-27 22:32:05,667 INFO namenode.NameNode: Caching file names occurring more than 10 times


2020-04-27 22:32:05,674 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536


2020-04-27 22:32:05,677 INFO snapshot.SnapshotManager: SkipList is disabled


2020-04-27 22:32:05,681 INFO util.GSet: Computing capacity for map cachedBlocks


2020-04-27 22:32:05,681 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,682 INFO util.GSet: 0.25% max memory 889 MB = 2.2 MB


2020-04-27 22:32:05,683 INFO util.GSet: capacity      = 2^18 = 262144 entries


2020-04-27 22:32:05,713 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10


2020-04-27 22:32:05,714 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10


2020-04-27 22:32:05,714 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25


2020-04-27 22:32:05,720 INFO namenode.FSNamesystem: Retry cache on namenode is enabled


2020-04-27 22:32:05,721 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis


2020-04-27 22:32:05,723 INFO util.GSet: Computing capacity for map NameNodeRetryCache


2020-04-27 22:32:05,723 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,724 INFO util.GSet: 0.029999999329447746% max memory 889 MB = 273.1 KB


2020-04-27 22:32:05,724 INFO util.GSet: capacity      = 2^15 = 32768 entries


2020-04-27 22:32:05,765 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1264791665-192.168.10.2-1588008725757


2020-04-27 22:32:05,810 INFO common.Storage: Storage directory E:\big-data\data\dfs\namespace_logs has been successfully formatted.


2020-04-27 22:32:05,841 INFO namenode.FSImageFormatProtobuf: Saving image file E:\big-data\data\dfs\namespace_logs\current\fsimage.ckpt_0000000000000000000 using no compression


2020-04-27 22:32:05,939 INFO namenode.FSImageFormatProtobuf: Image file E:\big-data\data\dfs\namespace_logs\current\fsimage.ckpt_0000000000000000000 of size 399 bytes saved in 0 seconds .


2020-04-27 22:32:05,957 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0


2020-04-27 22:32:05,963 INFO namenode.FSImage: FSImageSaver clean checkpoint: txid=0 when meet shutdown.


2020-04-27 22:32:05,963 INFO namenode.NameNode: SHUTDOWN_MSG:


/************************************************************


SHUTDOWN_MSG: Shutting down NameNode at DESKTOP-ROC4R5P/192.168.10.2


************************************************************/




i have downloaded jar and put in folder also.




https://github.com/FahaoTang/big-data/blob/master/hadoop-hdfs-3.2.1.jar 


Can you help me what thing i am setting wrong??? it will be great help and guidance.


Regards,

Saad

S Saad United #295 access_time 4 years ago more_vert

hello Raymond,


I learning about Hadoop and was following your detailed information on windows installation.




2020-04-27 22:32:04,347 INFO namenode.NameNode: createNameNode [-format]


Formatting using clusterid: CID-1d0c51aa-5dde-446b-99c1-3997255160fa


2020-04-27 22:32:05,369 INFO namenode.FSEditLog: Edit logging is async:true


2020-04-27 22:32:05,385 INFO namenode.FSNamesystem: KeyProvider: null


2020-04-27 22:32:05,387 INFO namenode.FSNamesystem: fsLock is fair: true


2020-04-27 22:32:05,388 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false


2020-04-27 22:32:05,428 INFO namenode.FSNamesystem: fsOwner             = saad (auth:SIMPLE)


2020-04-27 22:32:05,431 INFO namenode.FSNamesystem: supergroup          = supergroup


2020-04-27 22:32:05,431 INFO namenode.FSNamesystem: isPermissionEnabled = true


2020-04-27 22:32:05,432 INFO namenode.FSNamesystem: HA Enabled: false


2020-04-27 22:32:05,535 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling


2020-04-27 22:32:05,554 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000


2020-04-27 22:32:05,554 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true


2020-04-27 22:32:05,562 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000


2020-04-27 22:32:05,563 INFO blockmanagement.BlockManager: The block deletion will start around 2020 Apr 27 22:32:05


2020-04-27 22:32:05,566 INFO util.GSet: Computing capacity for map BlocksMap


2020-04-27 22:32:05,566 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,568 INFO util.GSet: 2.0% max memory 889 MB = 17.8 MB


2020-04-27 22:32:05,568 INFO util.GSet: capacity      = 2^21 = 2097152 entries


2020-04-27 22:32:05,579 INFO blockmanagement.BlockManager: Storage policy satisfier is disabled


2020-04-27 22:32:05,580 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false


2020-04-27 22:32:05,588 INFO Configuration.deprecation: No unit for dfs.namenode.safemode.extension(30000) assuming MILLISECONDS


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.9990000128746033


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0


2020-04-27 22:32:05,589 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 30000


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: defaultReplication         = 1


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: maxReplication             = 512


2020-04-27 22:32:05,591 INFO blockmanagement.BlockManager: minReplication             = 1


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: redundancyRecheckInterval  = 3000ms


2020-04-27 22:32:05,592 INFO blockmanagement.BlockManager: encryptDataTransfer        = false


2020-04-27 22:32:05,593 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000


2020-04-27 22:32:05,646 INFO namenode.FSDirectory: GLOBAL serial map: bits=29 maxEntries=536870911


2020-04-27 22:32:05,646 INFO namenode.FSDirectory: USER serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,647 INFO namenode.FSDirectory: GROUP serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,647 INFO namenode.FSDirectory: XATTR serial map: bits=24 maxEntries=16777215


2020-04-27 22:32:05,664 INFO util.GSet: Computing capacity for map INodeMap


2020-04-27 22:32:05,664 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,664 INFO util.GSet: 1.0% max memory 889 MB = 8.9 MB


2020-04-27 22:32:05,665 INFO util.GSet: capacity      = 2^20 = 1048576 entries


2020-04-27 22:32:05,666 INFO namenode.FSDirectory: ACLs enabled? false


2020-04-27 22:32:05,666 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true


2020-04-27 22:32:05,667 INFO namenode.FSDirectory: XAttrs enabled? true


2020-04-27 22:32:05,667 INFO namenode.NameNode: Caching file names occurring more than 10 times


2020-04-27 22:32:05,674 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536


2020-04-27 22:32:05,677 INFO snapshot.SnapshotManager: SkipList is disabled


2020-04-27 22:32:05,681 INFO util.GSet: Computing capacity for map cachedBlocks


2020-04-27 22:32:05,681 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,682 INFO util.GSet: 0.25% max memory 889 MB = 2.2 MB


2020-04-27 22:32:05,683 INFO util.GSet: capacity      = 2^18 = 262144 entries


2020-04-27 22:32:05,713 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10


2020-04-27 22:32:05,714 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10


2020-04-27 22:32:05,714 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25


2020-04-27 22:32:05,720 INFO namenode.FSNamesystem: Retry cache on namenode is enabled


2020-04-27 22:32:05,721 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis


2020-04-27 22:32:05,723 INFO util.GSet: Computing capacity for map NameNodeRetryCache


2020-04-27 22:32:05,723 INFO util.GSet: VM type       = 64-bit


2020-04-27 22:32:05,724 INFO util.GSet: 0.029999999329447746% max memory 889 MB = 273.1 KB


2020-04-27 22:32:05,724 INFO util.GSet: capacity      = 2^15 = 32768 entries


2020-04-27 22:32:05,765 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1264791665-192.168.10.2-1588008725757


2020-04-27 22:32:05,810 INFO common.Storage: Storage directory E:\big-data\data\dfs\namespace_logs has been successfully formatted.


2020-04-27 22:32:05,841 INFO namenode.FSImageFormatProtobuf: Saving image file E:\big-data\data\dfs\namespace_logs\current\fsimage.ckpt_0000000000000000000 using no compression


2020-04-27 22:32:05,939 INFO namenode.FSImageFormatProtobuf: Image file E:\big-data\data\dfs\namespace_logs\current\fsimage.ckpt_0000000000000000000 of size 399 bytes saved in 0 seconds .


2020-04-27 22:32:05,957 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0


2020-04-27 22:32:05,963 INFO namenode.FSImage: FSImageSaver clean checkpoint: txid=0 when meet shutdown.


2020-04-27 22:32:05,963 INFO namenode.NameNode: SHUTDOWN_MSG:


/************************************************************


SHUTDOWN_MSG: Shutting down NameNode at DESKTOP-ROC4R5P/192.168.10.2


************************************************************/




i have downloaded jar and put in folder also.




https://github.com/FahaoTang/big-data/blob/master/hadoop-hdfs-3.2.1.jar 


Can you help me what thing i am setting wrong??? it will be great help and guidance.


Regards,

Saad

Raymond Raymond #273 access_time 4 years ago more_vert

Did you follow step 3?

Step 3 - Install Hadoop native IO binary

If you've done that, you should be able to see the exe file in %HADOOP_HOME%/bin folder:

And also make sure HADOOP_HOME environment variable is configured correctly and also PATH environment variable has Hadoop bin folder.

You also need to restart PowerShell to source the latest environment variables if you are configure all these variables manually. 

Please let me know if that still exists. 

format_quote

person J Macklin access_time 4 years ago

When i type winutils.exe anr run

getting this error

winutils.exe : The term 'winutils.exe' is not recognized as the name of a cmdlet, function, script file, or operable

program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again.

At line:1 char:1

+ winutils.exe

+ ~~~~~~~~~~~~

    + CategoryInfo          : ObjectNotFound: (winutils.exe:String) [], CommandNotFoundException

    + FullyQualifiedErrorId : CommandNotFoundException

what to do


JM J Macklin Navamani #272 access_time 4 years ago more_vert

When i type winutils.exe anr run

getting this error

winutils.exe : The term 'winutils.exe' is not recognized as the name of a cmdlet, function, script file, or operable

program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again.

At line:1 char:1

+ winutils.exe

+ ~~~~~~~~~~~~

    + CategoryInfo          : ObjectNotFound: (winutils.exe:String) [], CommandNotFoundException

    + FullyQualifiedErrorId : CommandNotFoundException

what to do


Raymond Raymond #251 access_time 5 years ago more_vert

I'm glad it worked. I also updated the page to reflect the JDK requirement.

BTW, if you want to build a native Hadoop 3.2.1 on Windows, you can follow this guide:

Compile and Build Hadoop 3.2.1 on Windows 10 Guide

format_quote

person Matthias access_time 5 years ago

Hello Raymond,

you nailed it! Everything works fine now (except, I feel somewhat stupid, for I knew about JDK8, but for some reason assumed this to be but a minimum requirement. It didn't occur to me to revert to the older version. Let's hope, this will save someone at least something.)

Thanks a lot!

Cheers,
Matthias

M Matthias Wolf #250 access_time 5 years ago more_vert

Hello Raymond,

you nailed it! Everything works fine now (except, I feel somewhat stupid, for I knew about JDK8, but for some reason assumed this to be but a minimum requirement. It didn't occur to me to revert to the older version. Let's hope, this will save someone at least something.)

Thanks a lot!

Cheers,
Matthias

format_quote

person Raymond access_time 5 years ago

Hi Matthias,

As far as I know, JDK 13.0.1 is not supported. 

Please use JDK 1.8.x, it will work.

Even JDK 11 support is still in progress. 

https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+Java+Versions

Raymond Raymond #248 access_time 5 years ago more_vert

Hi Matthias,

As far as I know, JDK 13.0.1 is not supported. 

Please use JDK 1.8.x, it will work.

Even JDK 11 support is still in progress. 

https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+Java+Versions

format_quote

person Matthias access_time 5 years ago

Hi Raymond!

First of all: many thanks – a great how-to saving me hours!  Well, almost, that is.

I tried to install Hadoop 3.2.1 using your article. Everything went like a charm, I did encounter the problem mentioned when formatting the HDFS, the workaround worked fine.

However, when I reached Step 10 (launch YARN), both command windows run into a runtime exception. (See below)

The OS is a Windows 10 Edu N (english) and is up to date (as of today, Jan 25). The JDK is 13.0.1.

Any hint would be very much appreciated, as I'm far from being able to make sense of the messages!

Cheers from Vienna/Austria
Matthew



2020-01-25 15:54:39,323 INFO resourcemanager.ResourceManager: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting ResourceManager
STARTUP_MSG:   host = pitdb/10.0.2.15
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 3.2.1
STARTUP_MSG:   classpath = C:\hadoop-3.2.1\etc\hadoop;C:\hadoop-3.2.1\etc\hadoop;C:\hadoop-3.2.1\etc\hadoop;C:\hadoop-3.2.1\share\hadoop\common;C:\hadoop-3.2.1\share\hadoop\common\lib\accessors-smart-1.2.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\animal-sniffer-annotations-1.17.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\asm-5.0.4.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\audience-annotations-0.5.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\avro-1.7.7.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\checker-qual-2.5.2.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-beanutils-1.9.3.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-cli-1.2.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-codec-1.11.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-collections-3.2.2.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-compress-1.18.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-configuration2-2.1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-io-2.5.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-lang3-3.7.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-logging-1.1.3.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-math3-3.1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-net-3.6.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\commons-text-1.4.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\curator-client-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\curator-framework-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\curator-recipes-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\dnsjava-2.1.7.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\error_prone_annotations-2.2.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\failureaccess-1.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\gson-2.2.4.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\guava-27.0-jre.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\hadoop-annotations-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\hadoop-auth-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\htrace-core4-4.1.0-incubating.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\httpclient-4.5.6.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\httpcore-4.4.10.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\j2objc-annotations-1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-annotations-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-core-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-core-asl-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-databind-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-jaxrs-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-mapper-asl-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jackson-xc-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\javax.servlet-api-3.1.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jaxb-api-2.2.11.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jaxb-impl-2.2.3-1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jcip-annotations-1.0-1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jersey-core-1.19.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jersey-json-1.19.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jersey-server-1.19.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jersey-servlet-1.19.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jettison-1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-http-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-io-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-security-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-server-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-servlet-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-util-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-webapp-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jetty-xml-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jsch-0.1.54.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\json-smart-2.3.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jsp-api-2.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jsr305-3.0.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jsr311-api-1.1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\jul-to-slf4j-1.7.25.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-admin-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-client-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-common-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-core-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-crypto-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-identity-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-server-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-simplekdc-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerb-util-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerby-asn1-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerby-config-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerby-pkix-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerby-util-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\kerby-xdr-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\log4j-1.2.17.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\metrics-core-3.2.4.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\netty-3.10.5.Final.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\nimbus-jose-jwt-4.41.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\paranamer-2.3.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\protobuf-java-2.5.0.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\re2j-1.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\slf4j-api-1.7.25.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\slf4j-log4j12-1.7.25.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\snappy-java-1.0.5.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\stax2-api-3.1.4.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\token-provider-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\woodstox-core-5.0.3.jar;C:\hadoop-3.2.1\share\hadoop\common\lib\zookeeper-3.4.13.jar;C:\hadoop-3.2.1\share\hadoop\common\hadoop-common-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\common\hadoop-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\common\hadoop-kms-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\common\hadoop-nfs-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\accessors-smart-1.2.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\animal-sniffer-annotations-1.17.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\asm-5.0.4.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\audience-annotations-0.5.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\avro-1.7.7.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\checker-qual-2.5.2.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-beanutils-1.9.3.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-cli-1.2.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-codec-1.11.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-collections-3.2.2.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-compress-1.18.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-configuration2-2.1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-daemon-1.0.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-io-2.5.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-lang3-3.7.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-logging-1.1.3.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-math3-3.1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-net-3.6.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\commons-text-1.4.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\curator-client-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\curator-framework-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\curator-recipes-2.13.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\dnsjava-2.1.7.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\error_prone_annotations-2.2.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\failureaccess-1.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\gson-2.2.4.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\guava-27.0-jre.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\hadoop-annotations-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\hadoop-auth-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\htrace-core4-4.1.0-incubating.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\httpclient-4.5.6.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\httpcore-4.4.10.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\j2objc-annotations-1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-annotations-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-core-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-core-asl-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-databind-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-jaxrs-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-mapper-asl-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jackson-xc-1.9.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\javax.servlet-api-3.1.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jaxb-api-2.2.11.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jaxb-impl-2.2.3-1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jcip-annotations-1.0-1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jersey-core-1.19.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jersey-json-1.19.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jersey-server-1.19.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jersey-servlet-1.19.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jettison-1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-http-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-io-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-security-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-server-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-servlet-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-util-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-util-ajax-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-webapp-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jetty-xml-9.3.24.v20180605.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jsch-0.1.54.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\json-simple-1.1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\json-smart-2.3.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jsr305-3.0.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\jsr311-api-1.1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-admin-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-client-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-common-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-core-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-crypto-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-identity-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-server-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-simplekdc-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerb-util-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerby-asn1-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerby-config-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerby-pkix-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerby-util-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\kerby-xdr-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\leveldbjni-all-1.8.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\log4j-1.2.17.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\netty-3.10.5.Final.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\netty-all-4.0.52.Final.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\nimbus-jose-jwt-4.41.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\okhttp-2.7.5.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\okio-1.6.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\paranamer-2.3.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\protobuf-java-2.5.0.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\re2j-1.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\snappy-java-1.0.5.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\stax2-api-3.1.4.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\token-provider-1.0.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\woodstox-core-5.0.3.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\lib\zookeeper-3.4.13.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-client-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-client-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-httpfs-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-native-client-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-native-client-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-nfs-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-rbf-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\hdfs\hadoop-hdfs-rbf-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn;C:\hadoop-3.2.1\share\hadoop\yarn\lib\aopalliance-1.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\bcpkix-jdk15on-1.60.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\bcprov-jdk15on-1.60.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\ehcache-3.3.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\fst-2.50.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\geronimo-jcache_1.0_spec-1.0-alpha-1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\guice-4.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\guice-servlet-4.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\HikariCP-java7-2.4.12.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-jaxrs-base-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-jaxrs-json-provider-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-module-jaxb-annotations-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\java-util-1.9.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\javax.inject-1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jersey-client-1.19.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jersey-guice-1.19.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\json-io-2.5.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\metrics-core-3.2.4.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\mssql-jdbc-6.2.1.jre7.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\objenesis-1.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\snakeyaml-1.16.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\swagger-annotations-1.5.4.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-api-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-applications-distributedshell-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-applications-unmanaged-am-launcher-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-client-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-registry-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-applicationhistoryservice-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-nodemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-resourcemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-router-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-sharedcachemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-tests-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-timeline-pluginstorage-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-web-proxy-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-services-api-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-services-core-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-submarine-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\lib\hamcrest-core-1.3.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\lib\junit-4.11.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-app-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-core-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-hs-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-hs-plugins-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-jobclient-3.2.1-tests.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-jobclient-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-nativetask-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-shuffle-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-client-uploader-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\mapreduce\hadoop-mapreduce-examples-3.2.1.jar;C:\hadoop-2.7.7\etc\hadoop;C:\hadoop-2.7.7\lib;C:\hadoop-2.7.7\share\hadoop\common;C:\hadoop-2.7.7\share\hadoop\common\;C:\hadoop-2.7.7\share\hadoop\common\*;C:\hadoop-2.7.7\share\hadoop\common\lib\*;C:\hadoop-2.7.7\share\hadoop\hdfs;C:\hadoop-2.7.7\share\hadoop\hdfs\*;C:\hadoop-2.7.7\share\hadoop\hdfs\lib\*;C:\hadoop-2.7.7\share\hadoop\yarn\*;C:\hadoop-2.7.7\share\hadoop\yarn\lib\*;C:\hadoop-2.7.7\share\hadoop\mapreduce;C:\hadoop-2.7.7\share\hadoop\mapreduce\*;C:\hadoop-2.7.7\share\hadoop\mapreduce\lib\*;C:\Users\user\Desktop\hadoop\units.jar;C:\Users\user\Desktop\hadoop\wc.jar;;;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-api-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-applications-distributedshell-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-applications-unmanaged-am-launcher-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-client-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-registry-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-applicationhistoryservice-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-nodemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-resourcemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-router-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-sharedcachemanager-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-tests-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-timeline-pluginstorage-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-server-web-proxy-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-services-api-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-services-core-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\hadoop-yarn-submarine-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\aopalliance-1.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\bcpkix-jdk15on-1.60.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\bcprov-jdk15on-1.60.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\ehcache-3.3.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\fst-2.50.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\geronimo-jcache_1.0_spec-1.0-alpha-1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\guice-4.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\guice-servlet-4.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\HikariCP-java7-2.4.12.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-jaxrs-base-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-jaxrs-json-provider-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jackson-module-jaxb-annotations-2.9.8.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\java-util-1.9.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\javax.inject-1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jersey-client-1.19.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\jersey-guice-1.19.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\json-io-2.5.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\metrics-core-3.2.4.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\mssql-jdbc-6.2.1.jre7.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\objenesis-1.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\snakeyaml-1.16.jar;C:\hadoop-3.2.1\share\hadoop\yarn\lib\swagger-annotations-1.5.4.jar;C:\hadoop-3.2.1\etc\hadoop\rm-config\log4j.properties;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\hadoop-yarn-server-timelineservice-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\hadoop-yarn-server-timelineservice-hbase-client-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\hadoop-yarn-server-timelineservice-hbase-common-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\hadoop-yarn-server-timelineservice-hbase-coprocessor-3.2.1.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\commons-csv-1.0.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\commons-lang-2.6.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\hbase-annotations-1.2.6.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\hbase-client-1.2.6.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\hbase-common-1.2.6.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\hbase-protocol-1.2.6.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\htrace-core-3.1.0-incubating.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\jcodings-1.0.13.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\joni-2.1.2.jar;C:\hadoop-3.2.1\share\hadoop\yarn\timelineservice\lib\metrics-core-2.2.0.jar
STARTUP_MSG:   build = https://gitbox.apache.org/repos/asf/hadoop.git -r b3cbbb467e22ea829b3808f4b7b01d07e0bf3842; compiled by 'rohithsharmaks' on 2019-09-10T15:56Z
STARTUP_MSG:   java = 13.0.1
************************************************************/
2020-01-25 15:54:41,357 INFO conf.Configuration: found resource core-site.xml at file:/C:/hadoop-3.2.1/etc/hadoop/core-site.xml
2020-01-25 15:54:42,042 INFO conf.Configuration: resource-types.xml not found
2020-01-25 15:54:42,042 INFO resource.ResourceUtils: Unable to find 'resource-types.xml'.
2020-01-25 15:54:42,386 INFO conf.Configuration: found resource yarn-site.xml at file:/C:/hadoop-3.2.1/etc/hadoop/yarn-site.xml
2020-01-25 15:54:42,542 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.RMFatalEventType for class org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$RMFatalEventDispatcher
2020-01-25 15:54:42,698 INFO security.NMTokenSecretManagerInRM: NMTokenKeyRollingInterval: 86400000ms and NMTokenKeyActivationDelay: 900000ms
2020-01-25 15:54:42,714 INFO security.RMContainerTokenSecretManager: ContainerTokenKeyRollingInterval: 86400000ms and ContainerTokenKeyActivationDelay: 900000ms
2020-01-25 15:54:42,730 INFO security.AMRMTokenSecretManager: AMRMTokenKeyRollingInterval: 86400000ms and AMRMTokenKeyActivationDelay: 900000 ms
2020-01-25 15:54:42,792 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.recovery.RMStateStoreEventType for class org.apache.hadoop.yarn.server.resourcemanager.recovery.RMStateStore$ForwardingEventHandler
2020-01-25 15:54:42,792 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.NodesListManagerEventType for class org.apache.hadoop.yarn.server.resourcemanager.NodesListManager
2020-01-25 15:54:42,839 INFO resourcemanager.ResourceManager: Using Scheduler: org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler
2020-01-25 15:54:42,979 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.scheduler.event.SchedulerEventType for class org.apache.hadoop.yarn.event.EventDispatcher
2020-01-25 15:54:43,026 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.rmapp.RMAppEventType for class org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$ApplicationEventDispatcher
2020-01-25 15:54:43,104 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.rmapp.attempt.RMAppAttemptEventType for class org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$ApplicationAttemptEventDispatcher
2020-01-25 15:54:43,104 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNodeEventType for class org.apache.hadoop.yarn.server.resourcemanager.ResourceManager$NodeEventDispatcher
2020-01-25 15:54:43,666 INFO impl.MetricsConfig: Loaded properties from hadoop-metrics2.properties
2020-01-25 15:54:44,025 INFO impl.MetricsSystemImpl: Scheduled Metric snapshot period at 10 second(s).
2020-01-25 15:54:44,042 INFO impl.MetricsSystemImpl: ResourceManager metrics system started
2020-01-25 15:54:44,119 INFO security.YarnAuthorizationProvider: org.apache.hadoop.yarn.security.ConfiguredYarnAuthorizer is instantiated.
2020-01-25 15:54:44,151 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.RMAppManagerEventType for class org.apache.hadoop.yarn.server.resourcemanager.RMAppManager
2020-01-25 15:54:44,166 INFO event.AsyncDispatcher: Registering class org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncherEventType for class org.apache.hadoop.yarn.server.resourcemanager.amlauncher.ApplicationMasterLauncher
2020-01-25 15:54:44,182 INFO resourcemanager.RMNMInfo: Registered RMNMInfo MBean
2020-01-25 15:54:44,198 INFO monitor.RMAppLifetimeMonitor: Application lifelime monitor interval set to 3000 ms.
2020-01-25 15:54:44,198 INFO placement.MultiNodeSortingManager: Initializing NodeSortingService=MultiNodeSortingManager
2020-01-25 15:54:44,261 INFO util.HostsFileReader: Refreshing hosts (include/exclude) list
2020-01-25 15:54:44,308 INFO conf.Configuration: found resource capacity-scheduler.xml at file:/C:/hadoop-3.2.1/etc/hadoop/capacity-scheduler.xml
2020-01-25 15:54:44,448 INFO scheduler.AbstractYarnScheduler: Minimum allocation = <memory:1024, vCores:1>
2020-01-25 15:54:44,464 INFO scheduler.AbstractYarnScheduler: Maximum allocation = <memory:8192, vCores:4>
2020-01-25 15:54:44,605 INFO capacity.CapacitySchedulerConfiguration: max alloc mb per queue for root is undefined
2020-01-25 15:54:44,682 INFO capacity.CapacitySchedulerConfiguration: max alloc vcore per queue for root is undefined
2020-01-25 15:54:44,776 INFO capacity.ParentQueue: root, capacity=1.0, absoluteCapacity=1.0, maxCapacity=1.0, absoluteMaxCapacity=1.0, state=RUNNING, acls=ADMINISTER_QUEUE:*SUBMIT_APP:*, labels=*,
, reservationsContinueLooking=true, orderingPolicy=utilization, priority=0
2020-01-25 15:54:44,808 INFO capacity.ParentQueue: Initialized parent-queue root name=root, fullname=root
2020-01-25 15:54:44,902 INFO capacity.CapacitySchedulerConfiguration: max alloc mb per queue for root.default is undefined
2020-01-25 15:54:45,011 INFO capacity.CapacitySchedulerConfiguration: max alloc vcore per queue for root.default is undefined
2020-01-25 15:54:45,089 INFO capacity.LeafQueue: Initializing default
capacity = 1.0 [= (float) configuredCapacity / 100 ]
absoluteCapacity = 1.0 [= parentAbsoluteCapacity * capacity ]
maxCapacity = 1.0 [= configuredMaxCapacity ]
absoluteMaxCapacity = 1.0 [= 1.0 maximumCapacity undefined, (parentAbsoluteMaxCapacity * maximumCapacity) / 100 otherwise ]
effectiveMinResource=<memory:0, vCores:0>
 , effectiveMaxResource=<memory:0, vCores:0>
userLimit = 100 [= configuredUserLimit ]
userLimitFactor = 1.0 [= configuredUserLimitFactor ]
maxApplications = 10000 [= configuredMaximumSystemApplicationsPerQueue or (int)(configuredMaximumSystemApplications * absoluteCapacity)]
maxApplicationsPerUser = 10000 [= (int)(maxApplications * (userLimit / 100.0f) * userLimitFactor) ]
usedCapacity = 0.0 [= usedResourcesMemory / (clusterResourceMemory * absoluteCapacity)]
absoluteUsedCapacity = 0.0 [= usedResourcesMemory / clusterResourceMemory]
maxAMResourcePerQueuePercent = 0.1 [= configuredMaximumAMResourcePercent ]
minimumAllocationFactor = 0.875 [= (float)(maximumAllocationMemory - minimumAllocationMemory) / maximumAllocationMemory ]
maximumAllocation = <memory:8192, vCores:4> [= configuredMaxAllocation ]
numContainers = 0 [= currentNumContainers ]
state = RUNNING [= configuredState ]
acls = ADMINISTER_QUEUE:*SUBMIT_APP:* [= configuredAcls ]
nodeLocalityDelay = 40
rackLocalityAdditionalDelay = -1
labels=*,
reservationsContinueLooking = true
preemptionDisabled = true
defaultAppPriorityPerQueue = 0
priority = 0
maxLifetime = -1 seconds
defaultLifetime = -1 seconds
2020-01-25 15:54:45,152 INFO capacity.CapacitySchedulerQueueManager: Initialized queue: default: capacity=1.0, absoluteCapacity=1.0, usedResources=<memory:0, vCores:0>, usedCapacity=0.0, absoluteUsedCapacity=0.0, numApps=0, numContainers=0, effectiveMinResource=<memory:0, vCores:0> , effectiveMaxResource=<memory:0, vCores:0>
2020-01-25 15:54:45,167 INFO capacity.CapacitySchedulerQueueManager: Initialized queue: root: numChildQueue= 1, capacity=1.0, absoluteCapacity=1.0, usedResources=<memory:0, vCores:0>usedCapacity=0.0, numApps=0, numContainers=0
2020-01-25 15:54:45,167 INFO capacity.CapacitySchedulerQueueManager: Initialized root queue root: numChildQueue= 1, capacity=1.0, absoluteCapacity=1.0, usedResources=<memory:0, vCores:0>usedCapacity=0.0, numApps=0, numContainers=0
2020-01-25 15:54:45,167 INFO placement.UserGroupMappingPlacementRule: Initialized queue mappings, override: false
2020-01-25 15:54:45,167 INFO placement.MultiNodeSortingManager: MultiNode scheduling is 'false', and configured policies are
2020-01-25 15:54:45,197 INFO capacity.CapacityScheduler: Initialized CapacityScheduler with calculator=class org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator, minimumAllocation=<<memory:1024, vCores:1>>, maximumAllocation=<<memory:8192, vCores:4>>, asynchronousScheduling=false, asyncScheduleInterval=5ms,multiNodePlacementEnabled=false
2020-01-25 15:54:45,197 INFO conf.Configuration: dynamic-resources.xml not found
2020-01-25 15:54:45,213 INFO resourcemanager.AMSProcessingChain: Initializing AMS Processing chain. Root Processor=[org.apache.hadoop.yarn.server.resourcemanager.DefaultAMSProcessor].
2020-01-25 15:54:45,213 INFO resourcemanager.ApplicationMasterService: disabled placement handler will be used, all scheduling requests will be rejected.
2020-01-25 15:54:45,213 INFO resourcemanager.AMSProcessingChain: Adding [org.apache.hadoop.yarn.server.resourcemanager.scheduler.constraint.processor.DisabledPlacementProcessor] tp top of AMS Processing chain.
2020-01-25 15:54:45,229 INFO resourcemanager.ResourceManager: TimelineServicePublisher is not configured
2020-01-25 15:54:45,291 INFO util.log: Logging initialized @8928ms
2020-01-25 15:54:45,400 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets.
2020-01-25 15:54:45,417 INFO http.HttpRequestLog: Http request log for http.requests.resourcemanager is not defined
2020-01-25 15:54:45,432 INFO http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)
2020-01-25 15:54:45,432 INFO http.HttpServer2: Added filter RMAuthenticationFilter (class=org.apache.hadoop.yarn.server.security.http.RMAuthenticationFilter) to context cluster
2020-01-25 15:54:45,432 INFO http.HttpServer2: Added filter RMAuthenticationFilter (class=org.apache.hadoop.yarn.server.security.http.RMAuthenticationFilter) to context static
2020-01-25 15:54:45,432 INFO http.HttpServer2: Added filter RMAuthenticationFilter (class=org.apache.hadoop.yarn.server.security.http.RMAuthenticationFilter) to context logs
2020-01-25 15:54:45,448 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context cluster
2020-01-25 15:54:45,448 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2020-01-25 15:54:45,448 INFO http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2020-01-25 15:54:45,448 INFO http.HttpServer2: adding path spec: /cluster/*
2020-01-25 15:54:45,448 INFO http.HttpServer2: adding path spec: /ws/*
2020-01-25 15:54:45,448 INFO http.HttpServer2: adding path spec: /app/*
WARNING: An illegal reflective access operation has occurred
WARNING: Illegal reflective access by com.google.inject.internal.cglib.core.$ReflectUtils$2 (file:/C:/hadoop-3.2.1/share/hadoop/yarn/lib/guice-4.0.jar) to method java.lang.ClassLoader.defineClass(java.lang.String,byte[],int,int,java.security.ProtectionDomain)
WARNING: Please consider reporting this to the maintainers of com.google.inject.internal.cglib.core.$ReflectUtils$2
WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations
WARNING: All illegal access operations will be denied in a future release
2020-01-25 15:54:47,277 INFO webapp.WebApps: Registered webapp guice modules
2020-01-25 15:54:47,402 INFO http.HttpServer2: Jetty bound to port 8088
2020-01-25 15:54:47,574 INFO server.Server: jetty-9.3.24.v20180605, build timestamp: 2018-06-05T19:11:56+02:00, git hash: 84205aa28f11a4f31f2a3b86d1bba2cc8ab69827
2020-01-25 15:54:47,885 INFO server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets.
2020-01-25 15:54:47,964 INFO delegation.AbstractDelegationTokenSecretManager: Updating the current master key for generating delegation tokens
2020-01-25 15:54:48,057 INFO delegation.AbstractDelegationTokenSecretManager: Starting expired delegation token remover thread, tokenRemoverScanInterval=60 min(s)
2020-01-25 15:54:48,136 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@53ab0286{/logs,file:///C:/hadoop-3.2.1/logs/,AVAILABLE}
2020-01-25 15:54:48,230 INFO delegation.AbstractDelegationTokenSecretManager: Updating the current master key for generating delegation tokens
2020-01-25 15:54:48,479 INFO handler.ContextHandler: Started o.e.j.s.ServletContextHandler@2ce86164{/static,jar:file:/C:/hadoop-3.2.1/share/hadoop/yarn/hadoop-yarn-common-3.2.1.jar!/webapps/static,AVAILABLE}
Jõn. 25, 2020 3:54:49 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver as a provider class
Jõn. 25, 2020 3:54:49 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebServices as a root resource class
Jõn. 25, 2020 3:54:49 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory register
INFO: Registering org.apache.hadoop.yarn.webapp.GenericExceptionHandler as a provider class
Jõn. 25, 2020 3:54:49 PM com.sun.jersey.server.impl.application.WebApplicationImpl _initiate
INFO: Initiating Jersey application, version 'Jersey: 1.19 02/11/2015 03:25 AM'
Jõn. 25, 2020 3:54:50 PM com.sun.jersey.guice.spi.container.GuiceComponentProviderFactory getComponentProvider
INFO: Binding org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver to GuiceManagedComponentProvider with the scope "Singleton"
2020-01-25 15:54:50,729 WARN webapp.WebAppContext: Failed startup of context o.e.j.w.WebAppContext@4422dd48{/,file:///C:/Users/user/AppData/Local/Temp/jetty-0.0.0.0-8088-cluster-_-any-4591039387602299968.dir/webapp/,UNAVAILABLE}{/cluster}
com.google.inject.ProvisionException: Unable to provision, see the following errors:

1) Error injecting constructor, java.lang.NoClassDefFoundError: javax/activation/DataSource
  at org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver.<init>(JAXBContextResolver.java:41)
  at org.apache.hadoop.yarn.server.resourcemanager.webapp.RMWebApp.setup(RMWebApp.java:54)
  while locating org.apache.hadoop.yarn.server.resourcemanager.webapp.JAXBContextResolver

1 error
        at com.google.inject.internal.InjectorImpl$2.get(InjectorImpl.java:1025)
        at com.google.inject.internal.InjectorImpl.getInstance(InjectorImpl.java:1051)

Please log in or register to comment.

account_circle Log in person_add Register

Log in with external accounts