Hitachi Vantara Pentaho Community Wiki
Child pages
  • Loading Data into the MapR filesystem
Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

Version 1 Next »

Loading Data into CLDB
This guide shows you how to use a PDI job to move a file into CLDB.

Prerequisites

In order follow along with this how-to guide you will need the following:

  • MapR
  • Pentaho Data Integration

Sample Files

The sample data files needed for this guide is:

File Name

Content

weblogs_rebuild.txt

Unparsed, raw weblog data

Step-By-Step Instructions

Setup

Start MapR if it is not already running.

Create a Job to Put the Files into MapR

In this task you will load a file into MapR's CLDB.

  1. Start PDI on your desktop. Once it is running choose 'File' -> 'New' -> 'Job' from the menu system or click on the 'New file' icon on the toolbar and choose the 'Job' option.
  2. Add a Start Job Entry: You need to tell PDI where to start the job, so expand the 'General' section of the Design palette and drag a 'Start' job entry onto the job canvas.

  1. Add a Copy Files Job Entry: You will copy files from your local disk to the MapR CLDB, so expand the 'Hadoop' section of the Design palette and drag a 'Copy Files' job entry onto the job canvas. Your canvas should look like this:

  1. Connect the Start and Copy Files Job Entries: Hover the mouse over the 'Start' node and a tooltip will appear. Click on the output connector (the green arrow pointing to the right) and drag a connector arrow to the 'Copy Files' node. Your canvas should look like this:

  1. Edit the Copy Files Job Entry: Double-click on the 'Copy Files' node to edit its properties. Enter this information:
    1. File/Folder source(s): The folder containing the sample files you want to add to the CLDB.
    2. File/Folder destination(s): maprfs://<CLDB>:<PORT>/weblogs/raw

When running PDI on the same machine as the MapR cluster use: maprfs:///weblogs/raw the CLDB and port are not required.
<CLDB> is the server name of the machine running the MapR CLDB.
<PORT> is the port the MapR CLDB is running on.

    1. Wildcard (RegExp): Enter ^.*\.txt
    2. Click the Add button to add the above entries to the list of files you wish to copy.
    3. Check the "Create destination folder" option to ensure that the weblogs folder is created in CLDB the first time this job is executed.

When you are done your 'Copy Files' window should look like this (your file paths may be different):

Click 'OK' to close the window.

  1. Save the Job: Choose 'File' -> 'Save as...' from the menu system. Save the transformation as 'load_cldb.kjb' into a folder of your choice.
  2. Run the Job: Choose 'Action' -> 'Run' from the menu system or click on the green run button on the job toolbar. An 'Execute a job' window will open. Click on the 'Launch' button. An 'Execution Results' panel will open at the bottom of the PDI window and it will show you the progress of the job as it runs. After a few seconds the job should finish successfully:

If any errors occurred the job step that failed will be highlighted in red and you can use the 'Logging' tab to view error messages.

Check MapR

  1. If you have mounted your MapR CLDB onto your local machine you may verify the file loaded by navigating to the MapR directory.


ls /mapr/my.cluster.com/weblogs/raw
This should return:
weblog_raw.txt

  1. If you have not mounted your MapR CLDB onto your local machine you may alternatively check MapR by:


Hadoop fs -ls /weblogs/raw
This should return:
-rwxrwxrwx 3 demo demo 77908174 2011-12-28 07:16 /weblogs/raw/weblog_raw.txt

Summary

In this guide you learned how to copy local files into MapR's CLDB using PDI's graphical design tool. You can use this tool to put files into the CLDB from many different sources.

  • No labels