Hadoop Uses The Concept Of Parallelism Or Serialism To Upload The Split Data To The Cluster?

  1. Create an account on AWS
  2. Launch four Ec2 instance on AWS
  3. Configure one instance as NameNode, one as Client and remaining two as DataNodes.
  4. Install jdk and hadoop package in all instances
  5. Configure “hdfs-site.xml” and “core-site.xml” file in both datanodes and namenode. (Reminder, no need to configure “hdfs-site.xml” file in hadoop-client , only configure “core-site.xml” file)
  6. Format the NameNode
  7. Start the Hadoop daemon services in both DataNodes and NameNode and check by using “jps” command
  8. Check Datanodes available to the Hadoop-Cluster by using command “#hadoop dfsadmin -report”
  9. Hadoop-Client uploads a file to the Hadoop-Cluster by using command:-
DataNode-1
DataNode-2

--

--

--

Aeromodeller|Passionate|Technoholic|Learner|Technical writer

Love podcasts or audiobooks? Learn on the go with our new app.

Recommended from Medium

How to Use Post Processing | Unity 3D

An Introduction to Linked Lists

Modelling dispersive materials with gprMax software during Google Summer of Code 2021 adventure

Different Database Related Jobs- An overview

Creating a Maturity Level 3 Rest API with HATEOAS

How many tests should I use?

How to Think About Reliability

How I built a serverless joke machine for Alexa

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Lalita Sharma

Lalita Sharma

Aeromodeller|Passionate|Technoholic|Learner|Technical writer

More from Medium

AWS Parameter Store — How to get all parameters at a given path

Amazon S3 Storage Classes

How to use AWS EC2, EBS, and S3 services using AWS CLI………………

AWS S3 with IAM user credentials in Postman