1 / 24

Hadoop Tutorial for Big Data Enthusiasts

Hadoop Tutorial u2013 One of the most searched terms on the internet today. Do you know the reason? It is because Hadoop is the major part or framework of Big Data.<br>If you donu2019t know anything about Big Data then you are in major trouble.

joshii223
Download Presentation

Hadoop Tutorial for Big Data Enthusiasts

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Hadoop Tutorial for Big Data Enthusiasts The Optimal way of Learning Hadoop

  2. Hadoop Tutorial – One of the most searched terms on the internet today. Do you know the reason? It is because Hadoop is the major part or framework of Big Data. If you don’t know anything about Big Data then you are in major trouble. But don’t worry I have something for you which is completely FREE – 520+ Big Data Tutorials. This free tutorial series will make you a master of Big Data in just a few weeks. Also, I have explained a little about Big Data in this blog. “Hadoop is a technology to store massive datasets on a cluster of cheap machines in a distributed manner”. It was originated by Doug Cutting and Mike Cafarella. Doug Cutting’s kid named Hadoop to one of his toy that was a yellow elephant. Doug then used the name for his open source project because it was easy to spell, pronounce, and not used elsewhere. Interesting, right?

  3. Hadoop Tutorial Now, let’s begin our interesting Hadoop tutorial with the basic introduction to Big Data. What is Big Data? Big Data refers to the datasets too large and complex for traditional systems to store and process. The major problems faced by Big Data majorly falls under three Vs. They are volume, velocity, and variety. Do you know – Every minute we send 204 million emails, generate 1.8 million Facebook likes, send 278 thousand Tweets, and up-load 200,000 photos to Facebook. Volume: The data is getting generated in order of Tera to petabytes. The largest contributor of data is social media. For instance, Facebook generates 500 TB of data every day. Twitter generates 8TB of data daily. Velocity: Every enterprise has its own requirement of the time frame within which they have process data. Many use cases like credit card

  4. fraud detection have only a few seconds to process the data in real-time and detect fraud. Hence there is a need of framework which is capable of high-speed data computations. Variety: Also the data from various sources have varied formats like text, XML, images, audio, video, etc. Hence the Big Data technology should have the capability of performing analytics on a variety of data. Hope you have checked the Free Big Data DataFlair Tutorial Series. Here is one more interesting article for you –Top Big Data Quotes by the Experts

  5. Why is Hadoop Invented? Let us discuss the shortcomings of the traditional approach which led to the invention of Hadoop – 1. Storage for Large Datasets The conventional RDBMS is incapable of storing huge amounts of Data. The cost of data storage in available RDBMS is very high. As it incurs the cost of hardware and software both. 2. Handling data in different formats The RDBMS is capable of storing and manipulating data in a structured format. But in the real world we have to deal with data in a structured, unstructured and semi-structured format. 3. Data getting generated with high speed: The data is oozing out in the order of tera to peta bytes daily. Hence we need a system to process data in real-time within a few seconds. The traditional RDBMS fails to provide real-time processing at great speeds.

  6. What is Hadoop? Hadoop is the solution to the above Big Data problems. It is the technology to store massive datasets on a cluster of cheap machines in a distributed manner. Not only this it provides Big Data analytics through a distributed computing framework. It is an open-source software developed as a project by Apache Software Foundation. Doug Cutting created Hadoop. In the year 2008 Yahoo gave Hadoop to Apache Software Foundation. Since then two versions of Hadoop has come. Version 1.0 in the year 2011 and version 2.0.6 in the year 2013. Hadoop comes in various flavors like Cloudera, IBM BigInsight, MapR and Hortonworks. Prerequisites to Learn Hadoop ● Familiarity with some basic Linux Command – Hadoop is set up over Linux Operating System preferable Ubuntu. So one must know certain basic Linux commands. These commands are for uploading the file in HDFS, downloading the file from HDFS and so on. ● Basic Java concepts – Folks who want to learn Hadoop can get started in Hadoop while simultaneously graspingbasic concepts of Java. We can write maps and reduce functions in Hadoop using other languages too. And these are Python, Perl, C, Ruby, etc. This is possible via streaming

  7. API. It supports reading from standard input and writing to standard output. Hadoop also has high-level abstractions tools like Pig and Hive which do not require familiarity with Java. Hadoop consists of three core components – ● Hadoop Distributed File System (HDFS) – It is the storage layer of Hadoop. ● Map-Reduce – It is the data processing layer of Hadoop. ● YARN – It is the resource management layer of Hadoop. Core Components of Hadoop Let us understand these Hadoop components in detail. 1. HDFS Short for Hadoop Distributed File System provides for distributed storage for Hadoop. HDFS has a master-slave topology.

  8. Master is a high-end machine where as slaves are inexpensive computers. The Big Data files get divided into the number of blocks. Hadoop stores these blocks in a distributed fashion on the cluster of slave nodes. On the master, we have metadata stored. HDFS has two daemons running for it. They are : NameNode : NameNode performs following functions – ● NameNode Daemon runs on the master machine. ● It is responsible for maintaining, monitoring and managing DataNodes.

  9. ● It records the metadata of the files like the location of blocks, file size, permission, hierarchy etc. ● Namenode captures all the changes to the metadata like deletion, creation and renaming of the file in edit logs. ● It regularly receives heartbeat and block reports from the DataNodes. DataNode: The various functions of DataNode are as follows – ● DataNode runs on the slave machine. ● It stores the actual business data. ● It serves the read-write request from the user. ● DataNode does the ground work of creating, replicating and deleting the blocks on the command of NameNode. ● After every 3 seconds, by default, it sends a heartbeat to NameNode reporting the health of HDFS. Explore the top features of HDFS that a Hadoop developer must know

  10. 2. MapReduce It is the data processing layer of Hadoop. It processes data in two phases. They are:- Map Phase- This phase applies business logic to the data. The input data gets converted into key-value pairs. Reduce Phase- The Reduce phase takes as input the output of Map Phase. It applies aggregation based on the key of the key-value pairs.

  11. You must check this MapReduce tutorial to start your learning. Map-Reduce works in the following way: ● The client specifies the file for input to the Map function. It splits it into tuples ● Map function defines key and value from the input file. The output of the map function is this key-value pair. ● MapReduce framework sorts the key-value pair from map function. ● The framework merges the tuples having the same key together. ● The reducers get these merged key-value pairs as input.

  12. ● Reducer applies aggregate functions on key-value pairs. ● The output from the reducer gets written to HDFS.

  13. 3. YARN Short for Yet Another Resource Locator has the following components:- Resource Manager ● Resource Manager runs on the master node. ● It knows the location of slaves (Rack Awareness). ● It is aware of how many resources each slave has. ● Resource Scheduler is one of the important services run by the Resource Manager. ● Resource Scheduler decides how the resources get assigned to various tasks. ● Application Manager is one more service run by Resource Manager. ● Application Manager negotiates the first container for an application.

  14. ● Resource Manager keeps track of the heart beats from the Node Manager. Node Manager ● It runs on slave machines. ● It manages containers. Containers are nothing but a fraction of Node Manager’s resource capacity ● Node manager monitors resource utilization of each container. ● It sends a heartbeat to the Resource Manager.

  15. Job Submitter The application startup process is as follows:- ● The client submits the job to the Resource Manager. ● Resource Manager contacts Resource Scheduler and allocates containers. ● Now Resource Manager contacts the relevant Node Manager to launch the container. ● Container runs Application Master. The basic idea of YARN was to split the task of resource management and job scheduling. It has one global Resource Manager and per-application Application Master. An application can be either one job or DAG of jobs.

  16. The Resource Manager’s job is to assign resources to various competing applications. Node Manager runs on the slave nodes. It is responsible for containers, monitoring resource utilization and informing about the same to the Resource Manager. The job of Application master is to negotiate resources from the Resource Manager. It also works with NodeManager to execute and monitor the tasks. Wait before scrolling further! This is the time to read about the top 15 Hadoop Ecosystem components.

  17. Why Hadoop? Let us now understand why Big Data Hadoop is very popular, why Apache Hadoop captures more than 90% of the big data market. Apache Hadoop is not only a storage system but is a platform for data storage as well as processing. It is scalable (as we can add more nodes on the fly), Fault-tolerant (Even if nodes go down, data processed by another node). Following characteristics of Hadoop make it a unique platform: ● Flexibility to store and mine any type of data whether it is structured, semi-structured or unstructured. It is not bounded by a single schema. ● Excels at processing data of complex nature. Its scale-out architecture divides workloads across many nodes. Another added advantage is that its flexible file-system eliminates ETL bottlenecks. ● Scales economically, as discussed it can deploy on commodity hardware. Apart from this its open-source nature guards against vendor lock.

  18. Hadoop Architecture After understanding what Apache Hadoop is, let us now understand the Hadoop Architecture in detail. How Hadoop Works

  19. Hadoop works in master-slave fashion. There is a master node and there are n numbers of slave nodes where n can be 1000s. Master manages, maintains and monitors the slaves while slaves are the actual worker nodes. In Hadoop architecture, the Master should deploy on good configuration hardware, not just commodity hardware. As it is the centerpiece of Hadoop cluster. Master stores the metadata (data about data) while slaves are the nodes which store the data. Distributedly data stores in the cluster. The client connects with the master node to perform any task. Now in this Hadoop tutorial for beginners, we will discuss different features of Hadoop in detail.

  20. Hadoop Features Here are the top Hadoop features that make it popular – 1. Reliability In the Hadoop cluster, if any node goes down, it will not disable the whole cluster. Instead, another node will take the place of the failed node. The Hadoop cluster will continue functioning as nothing has happened. Hadoop has a built-in fault tolerance feature. 2. Scalable Hadoop gets integrated with cloud-based service. If you are installing Hadoop on the cloud you need not worry about scalability. You can easily procure more hardware and expand your Hadoop cluster within minutes. 3. Economical

  21. Hadoop gets deployed on commodity hardware which is cheap machines. This makes Hadoop very economical. Also as Hadoop is an open system software there is no cost of license too. 4. Distributed Processing In Hadoop, any job submitted by the client gets divided into the number of sub-tasks. These sub-tasks are independent of each other. Hence they execute in parallel giving high throughput. 5. Distributed Storage Hadoop splits each file into the number of blocks. These blocks get stored distributedly on the cluster of machines. 6. Fault Tolerance Hadoop replicates every block of file many times depending on the replication factor. Replication factor is 3 by default. In Hadoop suppose any node goes down then the data on that node gets recovered. This is because this copy of the data would be available on other nodes due to replication. Hadoop is fault tolerant. Are you looking for more Features? Here are the additional Hadoop Features that make it special.

  22. Hadoop Flavors This section of the Hadoop Tutorial talks about the various flavors of Hadoop. ● Apache – Vanilla flavor, as the actual code is residing in Apache repositories. ● Hortonworks – Popular distribution in the industry. ● Cloudera – It is the most popular in the industry. ● MapR – It has rewritten HDFS and its HDFS is faster as compared to others. ● IBM – Proprietary distribution is known as Big Insights. All the databases have provided native connectivity with Hadoop for fast data transfer. Because, to transfer data from Oracle to Hadoop, you need a connector. All flavors are almost the same and if you know one, you can easily work on other flavors as well.

  23. Hadoop Future Scope There is going to be a lot of investment in the Big Data industry in the coming years. According to a report by FORBES, 90% of global organizations will be investing in Big Data technology. Hence the demand for Hadoop resources will also grow. Learning Apache Hadoop will give you accelerated growth in your career. It also tends to increase your pay package. There is a lot of gap between the supply and demand of Big Data professional. The skill in Big Data technologies continues to be in high demand. This is because companies grow as they try to get the most out of their data. Therefore, their salary package is quite high as compared to professionals in other technology. The managing director of Dice, Alice Hills has said that Hadoop jobs have seen 64% increase from the previous year. It is evident that Hadoop is ruling the Big Data market and its future is bright. The demand for Big Data Analytics professional is ever increasing. As it is a known fact that data is nothing without power to analyze it. You must check Expert’s Prediction for the Future of Hadoop

  24. Summary – Hadoop Tutorial On concluding this Hadoop tutorial, we can say that Apache Hadoop is the most popular and powerful big data tool. Big Data stores huge amount of data in the distributed manner and processes the data in parallel on a cluster of nodes. It provides the world’s most reliable storage layer- HDFS. Batch processing engine MapReduce and Resource management layer- YARN. On summarizing this Hadoop Tutorial, I want to give you a quick revision of all the topics we have discussed ● The concept of Big Data ● Reason for Hadoop’s Invention ● Prerequisites to learn Hadoop ● Introduction to Hadoop ● Core components of Hadoop ● Why Hadoop ● Hadoop Architecture ● Features of Hadoop ● Hadoop Flavours ● Future Scope of Hadoop ● Boost your career with Free Big Data Course!!

More Related