Thank you for sending your enquiry! One of our team members will contact you shortly.
Thank you for sending your booking! One of our team members will contact you shortly.
Course Outline
Section 1: Data Management in HDFS
- Various Data Formats (JSON, Avro, Parquet)
- Compression Schemes
- Data Masking
- Labs: Analyzing different data formats; enabling compression
Section 2: Advanced Pig
- User-defined Functions
- Introduction to Pig Libraries (ElephantBird, Data-Fu)
- Loading Complex Structured Data using Pig
- Pig Tuning
- Labs: Advanced Pig scripting, parsing complex data types
Section 3: Advanced Hive
- User-defined Functions
- Compressed Tables
- Hive Performance Tuning
- Labs: Creating compressed tables, evaluating table formats and configuration
Section 4: Advanced HBase
- Advanced Schema Modelling
- Compression
- Bulk Data Ingest
- Wide-table vs. Tall-table comparison
- HBase and Pig
- HBase and Hive
- HBase Performance Tuning
- Labs: Tuning HBase; accessing HBase data from Pig & Hive; Using Phoenix for data modeling
Requirements
- Proficiency in the Java programming language (as most programming exercises are conducted in Java)
- Comfort with the Linux environment (including navigation via the Linux command line and file editing using vi or nano)
- A solid working knowledge of Hadoop.
Lab environment
Zero Install: Students do not need to install Hadoop software on their personal machines; a fully functional Hadoop cluster will be provided.
Students will need the following
21 Hours
Testimonials (1)
Hands on exercises. Class should have been 5 days, but the 3 days helped to clear up a lot of questions that I had from working with NiFi already