Basement Supercomputing

Developed in conjunction with Joomla extensions.


See Below for Dates and Locations

Workshop Enrollment is Limited to 15 Participants

Download Brochure with Workshop Outline

This one-day workshop covers the essential introductory aspects of the Apache Hadoop and Spark ecosystems. After completing the workshop attendees will gain an understanding Hadoop's technical value proposition and acquire hands-on experience with some basic Hadoop tools including Spark (with Zeppelin GUI). Approximately 30% of the workshop time is devoted to assisted hands-on exercises.

Unique Features:

Intended Audience:

Though no prior database experience is needed, those that work in the traditional database and data warehousing sectors should find the workshop useful. Devops and application programmers should find the material helps them understand the Hadoop processing models and ecosystem. Finally, those involved in data science or statistics will learn about how Hadoop and Spark can be used as an analytics tool.

What You Will Learn:

Attendees will learn why Hadoop is different from more traditional approaches to data analysis. The Hadoop core components will be presented and related to the various Hadoop capabilities. In addition, attendees will gain hands-on experience with the Hadoop Distributed File Systems (HDFS), the Hadoop resource manager (YARN), and several high level Hadoop tools including Spark. After completing the workshop attendees will be able to use and navigate a production Hadoop cluster and develop their own projects by building on the workshop examples.

About the Presenter:

Douglas Eadline, PhD, is a consultant and writer in the Big Data (Hadoop) and High Performance Computing (HPC) industries. Doug has written hundreds of articles, white papers, and instructional documents covering many aspects of HPC and Hadoop computing. Prior to starting and editing the popular ClusterMonkey.net website in 2005, he served as editor-in-chief for ClusterWorld Magazine, and was senior HPC editor for Linux Magazine. He has authored Hadoop Fundamentals LiveLessons, Second Edition (2015), and Apache Hadoop YARN LiveLessons (2014), and is coauthor of Apache Hadoop YARN (2014) and Hadoop 2 Quick-Start Guide (2016), all from Addison-Wesley.

Prerequisites:

  • Familiarity with the Linux command line and text editing is helpful
  • A wifi capable laptop with an up-to-date web browser and an ssh client (For Windows users, we highly recommend MobaXterm)
  • To minimize connection issues, attendees can rent a pre-configured Chromebook that meets these requirements

Note: The workshop can be presented in-house for your organization. Educational discounts available for qualifying academic institutions. Contact Basement Supercomputing directly for more information.

Dates and Locations

Basement Supercomputing may cancel (with refund) or suggest alternate days due to low enrollment.

(Can’t make the date? Join our mailing list for future Hadoop and Spark one-day workshops)

Chrombook Rental ($50) $50.00 each
14 items in stock
Results 1 - 1 of 1

Your Cart

The cart is empty