1 / 11

Basics of Hadoop

We are providing best training on hadoop. We train professionals across all experience 0 -15 years and we have separate modules like Developer module, Project manager module etc. We customize the syllabus covered according to the role requirements in the industry.

sindhubell
Download Presentation

Basics of Hadoop

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Hadoop is an open-source framework that was created to make it easier to work with big data.  It provides a method to access data that is distributed among multiple clustered computers, process the data, and manage resources across the computing and network resources that are involved.

  2. Hadoop” commonly refers to the core technology that consists of the four main components, but is also frequently used in reference to the entire ecosystem of supporting technologies and applications.

  3. Hadoop also is often used interchangeably with “big data”, but it shouldn’t be.  Hadoop is a framework for working with big data. It is part of the big data ecosystem, which consists of much more than Hadoop itself.

  4. Hadoop is not a single technology or product.  Instead, Hadoop is made up of four core modules that are supported by a large ecosystem of supporting technologies and products. The modules are: Hadoop Distributed File System. Hadoop YARN. Hadoop MapReduce. Hadoop Common.

  5. It Provides access to application data. Hadoop can also work with other file systems, including FTP, Amazon S3 and Windows Azure Storage Blobs (WASB), among others.

  6. Hadoop YARN– Provides the framework to schedule jobs and manage resources across the cluster that holds the data

  7. Hadoop MapReduce – A YARN-based parallel processing system for large data sets. The “Map” in MapReduce refers to the Map Tasks function. Map Tasks is the process of formatting data into key- value pairs and assigning them to nodes for the “Reduce” function, which is executed byReduce Tasks, where data is reduced to tuples.

  8. Common, which is also known as Hadoop Core, is a set of utilities that support the other Hadoop components.  Common is intended to give the Hadoop framework ways to manage typical (common) hardware failures.

  9. Click here For more detail www.peridotsystems.in

More Related