Hadoop Training Courses
Hadoop is an Apache Software Foundation (ASF) project that contains a suite of open-source software for scalable, reliable, distributed computing. Our Hadoop training includes courses in Hadoop application development and administration with specific emphasis on using the MapReduce framework. We also offer a non-technical Hadoop Overview course designed for managers and key decision makers who desire a better understanding of what Hadoop does and where it should be used.
You can take a Hadoop course online, at your own site, or at our training facility in the San Francisco Bay Area, California.
|Training Course||Jan, 2015||Feb, 2015||Mar, 2015||Apr, 2015|
|Hadoop Overview for Managers|
|Hadoop Developer Training with MapReduce|
|- Classroom - Online|
Hadoop Overview For Managers training course is designed for technical personnel and management who are evaluating and considering using Hadoop to solve their data scalability problems. We will cover Hadoop basics and discuss best practices using Hadoop in enterprises dealing with large data sets. Hadoop benefits and case studies on how other organizations are using Hadoop will be presented.
This 3-day hands-on Hadoop training course is designed for experienced developers and provides a fast track to building reliable and scalable application systems using Hadoop open-source software. Hadoop frameworks and tools are specifically geared towards processing of large datasets. Practical case studies will be demonstrated in class to show how Hadoop is used in real world today to solve different problems. MapReduce training is an essential component of this course.
This 3-day hands-on Hadoop for System Administrators class is designed for technical operations personnel whose job is to install and maintain production Hadoop clusters in real world. We will cover Hadoop architecture and its components, installation process, monitoring and troubleshooting of the complex Hadoop issues. The class includes practical hands-on exercises and encourages open discussions of how people are using Hadoop in enterprises dealing with large data sets.
Hive is a system for querying and managing structured data built on top of Hadoop. It uses Map-Reduce for execution, HDFS for storage, structured data with rich data types (structs, lists and maps). Allows to directly query data from different formats (text/binary) and file formats (Flat/Sequence) using SQL as a familiar programming tool for standard analytics. Hive provides extensibility using embedded scripts for non standard applications and supports rich MetaData to allow data discovery and optimization.