Hadoop development is one of the top skills most desired in software development. One of the reasons is because Hadoop is early in the product life cycle. It’s like getting involved with Relational Databases back in the early 80’s. Huge opportunity!!
Do I need to know Java?
If you are looking to learn how to write MapReduce jobs but don’t know Java that is okay. One way to write MapReduce jobs is to use Pig. Pig an application that works on top of Java and abstracts away the complexity of writing MapReduce jobs.
Pig developers are in high demand, checkout the number of open positions for Pig developers. If that doesn’t peek your interest run a search on “Pig Developer Salary” or checkout this post I wrote a couple weeks ago.
Now are you ready to learn Pig?
Here is a course that will show you how to write MapReduce jobs with only basic SQL skills. In fact, this tutorial will show you the basics of the Pig Latin language in under 2 hours.
Learn How to Get Started using Pig Latin with this Pig Tutorial!
What will you learn?
- We start with learning the history of Pig, where it fits in the Hadoop stack, and doing a comparison of Pig Latin, HiveQL, and SQL.
- Next we set our Hadoop development environment using both the Hortonworks sandbox and Cloudera QuickStart. Both of these environments run in a virtual machine on your desktop/laptop.
- Once we have our Hadoop development setup, we will start to learn the basic Pig Latin syntax and begin using the Pig editor to run our own Pig scripts.
- After learning the basics, we will begin to tackle a real-world problem and use Pig Latin to solve it.
- Need to extend the Pig? No problem, we will walk through User Defined Functions (UDF) and take a look at community provided UDFs in the Piggy Bank.
- In the end, we will wrap it up by learning how to use the Pig Grunt Shell to test, run, and debug our Pig scripts from the command line.