An Introduction to Apache Hadoop and HDFS
- Big Data, Hadoop and the Hortonworks Data Platform
- Installing the Hortonworks Data Platform
- Using HDFS Storage
- Managing Apache Ambari Users and Groups
- Managing Hadoop Services
LABS
- Setting Up the Lab Environment
- Installing HDP
- Managing Apache Ambari Users and Groups
- Managing Hadoop Services
Working with HDFS
- Using HDFS Storage
- Managing HDFS Storage
- Adding, Deleting, and Replacing Worker Nodes
- Configuring Rack Awareness
LABS
- Using Hadoop Storage
- Using WebHDFS
- Using HDFS Access Control Lists
- Managing Hadoop Storage
- Managing HDFS Quotas
- Adding, Decommissioning, and Recommissioning Worker Nodes
- Configuring Rack Awareness
Working with Apache YARN
- YARN Resource Management
- YARN Applications
- YARN Capacity Scheduler
LABS
- Managing YARN Using Ambari
- Managing YARN Using CLI
- Running Sample YARN Applications
- Setting Up for Capacity Scheduler
- Managing YARN Containers and Queues
- Managing YARN ACLs and User Limits
- YARN Node Labels
High Availability, Backups and Configuring Centralized Cache
- HDFS and YARN High Availability
- Monitoring a Cluster
- Protecting a Cluster with Backups
- Configuring Heterogenous HDFS Storage
- Managing the HDFS NFS Gateway
- Configuring HDFS Centralized Cache
LABS
- Configuring NameNode High Availability
- Configuring ResourceManager High Availability
- Managing Apache Ambari Alerts
- Managing HDFS Snapshots
- Using DistCP
- Configuring HDFS Storage Policies
- Configuring an NFS Gateway
- Configuring HDFS Centralized Cache
Performing a Rolling Upgrade
- Apache Hive Tuning
- Managing Workflows Using Apache Oozie
- Integrating Ambari with LDAP
- Automating Cluster Provisioning Using Ambari Blueprints
- Performing an HDP Rolling Upgrade
LABS
- Configuring Apache Hive High Availability
- Managing Workflows Using Apache Oozie
- Integrating Apache Ambari with AD/LDAP
- Automating Cluster Provisioning using Apache Ambari
- Performing an HDP Upgrade