question archive Project Goal The goal of this project is to design an automated data pipeline using hdfs commands and MapReduce

Project Goal The goal of this project is to design an automated data pipeline using hdfs commands and MapReduce

Subject:Computer SciencePrice: Bought3

Project Goal

The goal of this project is to design an automated data pipeline using hdfs commands and MapReduce. 

 

Overview

Your data pipeline will consist of a Unix script that completes the following steps:

  1. Takes some file(s) from the local file system and copies them to a staging folder in hdfs.
  2. Runs a MapReduce job on the files contained in the staging folder.
  3. Takes the output files, renames them and moves them to a target folder.
  4. Deletes the temporary (MR-generated) output folder.
  5. Moves the content of the staging folder to an archive folder in hdfs.

By local file system, we mean the file system on the client machine. For example, if you are accessing Hadoop from a virtual machine, the virtual machine would be the local system.

pur-new-sol

Purchase A New Answer

Custom new solution created by our subject matter experts

GET A QUOTE