Hitachi Vantara Pentaho Community Wiki
Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 45 Next »

Resources

Unknown macro: {div}

">Pentaho Big Data Plugin
Unknown macro: {div}

The Pentaho Big Data Plugin Project provides support for an ever-expanding Big Data community within the Pentaho ecosystem. It is a plugin for the Pentaho Kettle engine which can be used within Pentaho Data Integration (Kettle), Pentaho Reporting, and the Pentaho BI Platform.

Pentaho Big Data Plugin Features

This project contains the implementations for connecting to or preforming the following:

  • Pentaho MapReduce: visually design MapReduce jobs as Kettle transformations
  • HDFS File Operations: Read/write directly from any Kettle step. All made possible by the ubiquitous use of Apache VFS throughout Kettle
  • Data Sources
    • JDBC connectivity
      • Apache Hive
    • Native RPC connectivity for reading/writing
      • Apache HBase
      • Cassandra
      • MongoDB

Key Links

Community and where to find help

The Big Data Forum exists for both users and developers. The community also manages the ##pentaho IRC channel on irc.freenode.net.

Quick Start: Building the project

The Pentaho Big Data Plugin is built with Apache Ant and uses Apache Ivy for dependency management. All you'll need to get started is Ant 1.8.0 or newer to build the project. The build scripts will download Ivy if you do not already have it installed.

git clone git://github.com/pentaho/big-data-plugin.git
cd big-data-plugin
ant

Developing with Eclipse

We recommend Apache IvyDE to manage your Ivy dependencies within Eclipse.

  1. Import pentaho-big-data-plugin into Eclipse
  2. Resolve the project using IvyDE

If IvyDE is not an option then you can manually add the jars from lib/ and libswt/ to your class path. This project, like all other Pentaho projects, uses the open source Subfloor Ant build framework. Running the following targets will configure the Eclipse project to reference the required libraries:

ant resolve create-dot-classpath

Then import or refresh the project in Eclipse.

Contributing Changes

We use the Fork + Pull Model to manage community contributions. Please fork the repository and submit a pull request with your changes.

Here's a sample git workflow to get you started:

  1. Create a Github account
  2. Fork the project from https://github.com/pentaho/big-data-plugin
  3. Clone your repository:
    git clone git@github.com:USERNAME/big-data-plugin.git
  4. * Hack away *
  5. Stage and commit changes. Please make sure your commit messages include the JIRA case for your changes. It should be in the format: [JIRA-CASE] Short description of fixes.:
    git add . && git commit
  6. Push changes back up to Github:
    git push
  7. Submit a pull request from your project page. Please include a brief summary of what you changed and why.

Git Resources

Here's a short list of resources to help you learn and master Git:

Documentation

Kettle Plugin Development

Getting started with the Pentaho Data Integration Java API

Step Documentation

Job Entry Documentation

Community Plugins

Here's a list of known community plugins that fall into the "big data" category:

Voldemort Lookup
HPCC Systems ECL Plugins

  • No labels