EducationSoftwareStrategy.com
StrategyCommunity

Knowledge Base

Product

Community

Knowledge Base

TopicsBrowse ArticlesDeveloper Zone

Product

Download SoftwareProduct DocumentationSecurity Hub

Education

Tutorial VideosSolution GalleryEducation courses

Community

GuidelinesGrandmastersEvents
x_social-icon_white.svglinkedin_social-icon_white.svg
Strategy logoCommunity

© Strategy Inc. All Rights Reserved.

LegalTerms of UsePrivacy Policy
  1. Home
  2. Topics

How to deploy Hadoop Gateway Automatically


David Harsh

Software Engineer in Test II • Strategy


This video provides the steps necessary to deploy the Hadoop Gateway automatically in web

From Data import interface, navigate to hadoop. From this portal, you can navigate to the Gateway Manager. This is where you will add your cluster configuration.  
You will need to provide the following parameters in Web: 
Name: Avoid special characters
Hadoop Properties:
Hadoop NameNode: FQDN or IP
HDFS Port: browse files, def. 8020
WebHDFS: preview file, def. 50070
Gateway Properties: 
Host: machine to install Gateway
Port: I-Server to HG, def. 30004 
Spark Properties:
YARN: Jar: path of spark assembly
Standalone: Master: Spark service; ex. spark://SparkMasterNode:7077 
Local: Threads: simulated threads
Advanced: 
Memory of executor:
  ~RAM/executors per node
spark.executor.instance:
  # executors on cluster
spark.executor.cores:
  # cores per executor
 
Once we do this, we can deploy the gateway. Once it is deployed (installed); we can start the gateway from the gateway manager.  We can go back to the file browsing interface to ensure the WebHDFS, HDFS ports have been defined properly. And then we can navigate to the yarn application to ensure the application has properly started. Even though you might not be Big Data cluster admin, we have seen that most people in the network will typically have access to this interface to monitor the application. 


Comment

0 comments

Details

Knowledge Article

Published:

May 3, 2017

Last Updated:

May 3, 2017