Thanks for your kind words Jim.
I agree with the pain points on virtual machines.
We’ve created a series of videos and assessments on writing and running MapReduce code on CloudxLab. These videos are part of our Big Data with Hadoop and Spark course.
Please access the videos and slides here
In above videos, we have shown
- How to write MapReduce code using Java and Eclipse
- Build MapReduce project using Apache Ant
- Run MapReduce code using Hadoop Streaming
In the end, we have assessments, where you have to write code for problems using MapReduce and run it on CloudxLab. Some of the problems include
- Count the frequency of characters in a file stored in HDFS
- Find anagrams in a text file stored in HDFS
- Find users having same DNA
- Find users having mirror DNA
Please watch the videos and write the code for the above problems. Hope these videos and assessments will help you in writing MapReduce logic properly.
Happy learning!