- Pentaho MapReduce: visually design MapReduce jobs as Kettle transformations
- HDFS File Operations
- SVN Repository: svn://source.pentaho.org/svnkettleroot/pentaho-big-data-plugin (GitHub mirror: <TODO>)
- Documentation: <TODO: add dev doc page and aggregate links to wiki pages such as Cassandra Input, Cassandra Output, MongoDB Input, MongoDB Output)
- Link to Kettle plugin development
Quick Start: Building the project
The Pentaho Big Data Plugin is built with Apache Ant and uses Apache Ivy for dependency management. All you'll need to get started is Ant 1.8.0 or newer to build the project. The build scripts will download Ivy if you do not already have it installed.
svn co svn://source.pentaho.org/svnkettleroot/pentaho-big-data-plugin/trunk pentaho-big-data-plugin cd pentaho-big-data-plugin ant
Developing with Eclipse
We recommend Apache IvyDE to manage your Ivy dependencies within Eclipse.