1 / 8

Hadoop Big Data

Hadoop is an open-source, a Java-based programming framework that continues the processing of large data sets in a distributed computing environment. It based on the Google File System or GFS.

karyatech45
Télécharger la présentation

Hadoop Big Data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. HADOOP BIG DATA SOLUTIONS

  2. CONTENT : • What is Hadoop ? • Hadoop Framework • Hadoop Distributed File system • Hadoop Map Reduce

  3. WHAT IS HADOOP ? • Hadoop is an open-source, a Java-based programming framework that continues the processing of large data sets in a distributed computing environment. It based on the Google File System or GFS.

  4. HADOOP FRAMEWORK

  5. Hadoop Distribution File System • It provides a storage layer for Hadoop. It is suitable for distributed storage and processing. • HDFS Provides a command line interface to interact with Hadoop. • It provides streaming access to file system data. So, it includes file permission and authentication.

  6. Hbase : It helps to store data in HDFS. It is a NoSQL database or non-relational database. HBase mainly used when you need random, real-time, read/write access to your big data. • Sqoop : A sqoop is a tool designed to transfer data between Hadoop and NoSQL. It is managed to import data from relational databases such as Oracle and MySQL to HDFS and export data from HDFS to relational database. • Flume : Flume distributed service for ingesting streaming data. So, Distributed data that collect event data & transfer it to HDFS. It is ideally suitable for event data from multiple systems.

  7. HADOOP MAP REDUCE • It is another framework that processes the data. The original Hadoop processing engine which primarily based on JAVA. Based on the Map and Reduce programming model. • Many tools such as Hive, Pig are build on Map Reduce Model. • It is broad & mature fault tolerance framework. It is the most commonly used framework.

  8. THANK YOU For More Information : Visit : https://www.karyatech.com/big-data.html

More Related