compared with
Current by Jens Bleuel
on Mar 19, 2014 12:23.

Key
This line was removed.
This word was removed. This word was added.
This line was added.

Changes (83)

View Page History
| [EAI:Add value fields changing sequence] | Transform | FieldsChangeSequence | Add sequence depending of fields value change. Each time value of at least one field change, PDI will reset sequence. | opdts.fieldschangesequence.FieldsChangeSequenceMeta |
| [EAI:Add XML] | Transform | AddXML | Encode several fields into an XML fragment | opdts.addxml.AddXMLMeta |
| [EAI:Aggregate Rows]\\ | Deprecated \\ | | | |
| [EAI:Analytic Query] | Statistics | AnalyticQuery | Execute analytic queries over a sorted dataset (LEAD/LAG/FIRST/LAST) | opdts.analyticquery.AnalyticQueryMeta |
| [EAI:Append streams] | Flow | Append | Append 2 streams in an ordered way | opdts.append.AppendMeta |
| [EAI:Arff Output] | Data Mining | Arff Output | Writes data in ARFF format to a file | opdts.append.arff.ArffOutputMeta |
| [EAI:Automatic Documentation Output] | Output | AutoDoc | This step automatically generates documentation based on input in the form of a list of transformations and jobs | opdts.autodoc.AutoDocMeta |
| [Avro input|EAI:Avro Input] | Input | AvroInput | Decode binary or Json Avro data from a file or a field | opdts.avroinput.AvroInputMeta |
| [EAI:Call DB Procedure] | Lookup | DBProc | Get back information by calling a database procedure. | opdts.dbproc.DBProcMeta |
| [EAI:Change file encoding] | Utility | ChangeFileEncoding | Change file encoding and create a new file | opdts.changefileencoding.ChangeFileEncodingMeta |
| [Cassandra input|EAI:Cassandra Input] | Input Big Data | CassandraInput | Read from a Cassandra column family | opdts.cassandrainput.CassandraInputMeta |
| [Cassandra output|EAI:Cassandra Output] | Output Big Data | CassandraOutput | Write to a Cassandra column family | opdts.cassandraoutput.CassandraOutputMeta |
| [EAI:Check if a column exists] | Lookup | ColumnExists | Check if a column exists in a table on a specified connection. | opdts.columnexists.ColumnExistsMeta |
| [EAI:Check if file is locked] | Lookup | FileLocked | Check if a file is locked by another process | opdts.filelocked.FileLockedMeta |
| [EAI:Closure Generator] | Transform | ClosureGenerator | This step allows you to generates a closure table using parent-child relationships. | opdts.closure.ClosureGeneratorMeta |
| [Combination lookup/update|EAI:Combination lookup-update] | Data Warehouse | CombinationLookup | Update a junk dimension in a data warehouse. Alternatively, look up information in this dimension. The primary key of a junk dimension are all the fields. | opdts.combinationlookup.CombinationLookupMeta |
| [EAI:Concat Fields]\\ | Transform | ConcatFields \\ | The Concat Fields step is used to concatenate multiple fields into one target field. The fields can be separated by a separator and the enclosure logic is completely compatible with the Text File Output step. | opdts.concatfields.ConcatFieldsMeta |
| [EAI:Copy rows to result] | Job | RowsToResult | Use this step to write rows to the executing job. The information will then be passed to the next entry in this job. | opdts.rowstoresult.RowsToResultMeta |
| [EAI:CouchDB Input] | Big Data | CouchDbInput | Retrieves all documents from a given view in a given design document from a given database | opdts.couchdbinput.CouchDbInputMeta |
| [EAI:Credit card validator] | Validation | CreditCardValidator | The Credit card validator step will help you tell: (1) if a credit card number is valid (uses LUHN10 (MOD-10) algorithm) (2) which credit card vendor handles that number (VISA, MasterCard, Diners Club, EnRoute, American Express (AMEX),...) | opdts.creditcardvalidator.CreditCardValidatorMeta |
| [CSV file input|EAI:CSV Input] | Input | CsvInput | Simple CSV file input | opdts.csvinput.CsvInputMeta |
| [EAI:Dynamic SQL row] | Lookup | DynamicSQLRow | Execute dynamic SQL statement build in a previous field | opdts.dynamicsqlrow.DynamicSQLRowMeta |
| [EAI:Edi to XML]\\ | Utility | TypeExitEdi2XmlStep | Converts an Edifact message to XML to simplify data extraction (Available in PDI 4.4, already present in [CI trunk|http://ci.pentaho.com/view/Data%20Integration/job/Kettle/] builds) | opdts.edi2xml.Edi2XmlMeta |
| [ElasticSearch Bulk Insert |EAI:ElasticSearch Bulk Insert] | Bulk loading | ElasticSearchBulk | Performs bulk inserts into ElasticSearch | opdts.elasticsearchbulk.ElasticSearchBulkMeta |
| [Email messages input |EAI:Email Messages Input] | Input | MailInput | Read POP3/IMAP server and retrieve messages | opdts.mailinput.MailInputMeta |
| [ESRI Shapefile Reader |EAI:ESRI Shapefile Reader] | Input | ShapeFileReader | Reads shape file data from an ESRI shape file and linked DBF file | org.pentaho.di.shapefilereader.ShapeFileReaderMeta |
| [EAI:ETL Metadata Injection] | Flow | MetaInject | This step allows you to inject metadata into an existing transformation prior to execution. This allows for the creation of dynamic and highly flexible data integration solutions. | opdts.metainject.MetaInjectMeta |
| [Example plugin|EAI:Example plugin (Transformation Step)] | Transform | DummyPlugin | This is an example for a plugin test step | be.ibridge.kettle.dummy.DummyPluginMeta |
| [EAI:Execute a process] | Utility | ExecProcess | Execute a process and return the result | opdts.execprocess.ExecProcessMeta |
| [EAI:Execute row SQL script] | Scripting | ExecSQLRow | Execute SQL script extracted from a field created in a previous step. | opdts.execsqlrow.ExecSQLRowMeta |
| [EAI:Execute SQL script] | Scripting | ExecSQL | Execute an SQL script, optionally parameterized using input rows | opdts.sql.ExecSQLMeta |
| [EAI:File exists] | Lookup | FileExists | Check if a file exists | opdts.fileexists.FileExistsMeta |
| [EAI:Filter rows] [Filter Rows] | Flow | FilterRows | Filter rows using simple equations | opdts.filterrows.FilterRowsMeta |
| [Fixed file input|EAI:Fixed File Input] | Input | FixedInput | Fixed file input | opdts.fixedinput.FixedInputMeta |
| [EAI:Formula] | Scripting | Formula | Calculate a formula using Pentaho's libformula | opdts.formula.FormulaMeta |
| [EAI:Fuzzy match] | Lookup | FuzzyMatch | Finding approximate matches to a string using matching algorithms. Read a field from a main stream and output approximative value from lookup stream. | opdts.fuzzymatch.FuzzyMatchMeta |
| [EAI:Generate random credit card numbers] | Input | RandomCCNumberGenerator | Generate random valide (luhn check) credit card numbers | opdts.randomccnumber.RandomCCNumberGeneratorMeta |
| [Generate random value|EAI:Generate Random Value] | Input | RandomValue | Generate random value | opdts.randomvalue.RandomValueMeta |
| [EAI:Get Files Rows Count] | Input | GetFilesRowsCount | Get Files Rows Count | opdts.getfilesrowscount.GetFilesRowsCountMeta |
| [Get ID from slave server|EAI:Get ID from Slave Server] | Transform | GetSlaveSequence | Retrieves unique IDs in blocks from a slave server. The referenced sequence needs to be configured on the slave server in the XML configuration file. | opdts.getslavesequence.GetSlaveSequenceMeta |
| [EAI:Get previous row fields]\\ | Deprecated \\ | | | |
| [EAI:Get repository names] | Input | GetRepositoryNames | Lists detailed information about transformations and/or jobs in a repository | opdts.getrepositorynames.GetRepositoryNamesMeta |
| [EAI:Get rows from result] | Job | RowsFromResult | This allows you to read rows from a previous entry in a job. | opdts.rowsfromresult.RowsFromResultMeta |
| [Get SubFolder names |EAI:Get SubFolder names] | Input | GetSubFolders | Read a parent folder and return all subfolders | opdts.getsubfolders.GetSubFoldersMeta |
| [EAI:Get System Info] | Input | SystemInfo | Get information from the system like system date, arguments, etc. | opdts.systemdata.SystemDataMeta |
| [EAI:Get table names] | Input | GetTableNames | Get table names from database connection and send them to the next step | opdts.gettablenames.GetTableNamesMeta |
| [Get Variables|EAI:Get Variable] | Job | GetVariable | Determine the values of certain (environment or Kettle) variables and put them in field values. | opdts.getvariable.GetVariableMeta |
| [EAI:Google Analytics] | Input | TypeExitGoogleAnalyticsInputStep | Fetches data from google analytics account | opdts.googleanalytics.GaInputStepMeta |
| [EAI:Google Docs Input]\\ | Input \\ | | | |
| [Greenplum Bulk Loader|PMOPEN:GreenPlum bulk loader] | Bulk loading | GPBulkLoader | Greenplum Bulk Loader | opdts.gpbulkloader.GPBulkLoaderMeta |
| [EAI:Greenplum Load]\\ | Bulk loading \\ | GPLoad \\ | Greenplum Load \\ | |
| [Group by|EAI:Group By] | Statistics | GroupBy | Builds aggregates in a group by fashion. This works only on a sorted input. If the input is not sorted, only double consecutive rows are handled correctly. | opdts.groupby.GroupByMeta |
| [EAI:GZIP CSV Input] | Input | ParallelGzipCsvInput | Parallel GZIP CSV file input reader | opdts.parallelgzipcsv.ParGzipCsvInputMeta |
| [EAI:Hadoop File Input] | Big Data | HadoopFileInputPlugin | Read data from a variety of different text-file types stored on a Hadoop cluster | opdts.hadoopfileinput.HadoopFileInputMeta |
| [EAI:Hadoop File Output] | Big Data | HadoopFileOutputPlugin | Write data to a variety of different text-file types stored on a Hadoop cluster | opdts.hadoopfileoutput.HadoopFileOutputMeta |
| [HBase input|EAI:HBase Input] | Input Big Data | HbaseInput | Read from an HBase column family | opdts.hbaseinput.HBaseInputMeta |
| [HBase output|EAI:HBase Output] | Output Big Data | HbaseOutput | Write to an HBase column family | opdts.hbaseoutput.HBaseOutputMeta |
| [EAI:HBase Row Decoder] | Big Data | HBaseRowDecoder | Decodes an incoming key and HBase result object according to a mapping | opdts.hbaserowdecoder.HBaseRowDecoderMeta |
| [EAI:HL7 Input]\\ | Input \\ | HL7Input | Read data from HL7 data streams. | opdt.hl7.plugins.hl7input |
| [HTTP client|EAI:HTTP Client] | Lookup | HTTP | Call a web service over HTTP by supplying a base URL by allowing parameters to be set dynamically | opdts.http.HTTPMeta |
| [EAI:HTTP Post] | Lookup | HTTPPOST | Call a web service request over HTTP by supplying a base URL by allowing parameters to be set dynamically | opdts.httppost.HTTPPOSTMeta |
| [EAI:IBM Websphere MQ Consumer] | Input \\ | | Receive messages from any IBM Websphere MQ Server | |
| [EAI:IBM Websphere MQ Producer] | Output \\ | | Send messages to any IBM Websphere MQ Server | |
| [Identify last row in a stream |EAI:Identify last row in a stream] | Flow | DetectLastRow | Last row will be marked | opdts.detectlastrow.DetectLastRowMeta |
| [EAI:If field value is null] | Utility | IfNull | Sets a field value to a constant if it is null. | opdts.ifnull.IfNullMeta |
| Infobright Loader [EAI:Infobright Loader] | Bulk loading | InfobrightOutput | Load data to an Infobright database table | opdts.infobrightoutput.InfobrightLoaderMeta |
| [EAI:Ingres VectorWise Bulk Loader] | Bulk loading | VectorWiseBulkLoader | This step interfaces with the Ingres VectorWise Bulk Loader "COPY TABLE" command. | opdts.ivwloader.IngresVectorwiseLoaderMeta |
| [EAI:Injector] | Inline | Injector | Injector step to allow to inject rows into the transformation through the java API | opdts.injector.InjectorMeta |
| [Insert / Update|EAI:Insert - Update] | Output | InsertUpdate | Update or insert rows in a database based upon keys. | opdts.insertupdate.InsertUpdateMeta |
| [EAI:Java Filter] | Flow | JavaFilter | Filter rows using java code | opdts.javafilter.JavaFilterMeta |
| [EAI:JMS Consumer]\\ | Input \\ | | Receive messages from a JMS server | |
| [EAI:JMS Producer]\\ | Output \\ | | Send messages to a JMS server | |
| [EAI:Job Executor]\\ | Flow \\ | JobExecutor \\ | This step executes a Pentaho Data Integration Job, passes parameters and rows. \\ | opdts.jobexecutor.JobExecutorMeta \\ |
| [Join Rows (cartesian product) |EAI:Join Rows (Cartesian product)]\\ | Joins | JoinRows | The output of this step is the cartesian product of the input streams. The number of rows is the multiplication of the number of rows in the input streams. | opdts.joinrows.JoinRowsMeta |
| [EAI:Json Input] | Input | JsonInput | Extract relevant portions out of JSON structures (file or incoming field) and output rows | opdts.jsoninput.JsonInputMeta |
| Json output [JSON output] | Output | JsonOutput | Create Json bloc and output it in a field ou a file. | opdts.jsonoutput.JsonOutputMeta |
| [EAI:Knowledge Flow] | Data Mining | KF | Executes a Knowledge Flow data mining process | org.pentaho.di.kf.KFMeta |
| [EAI:LDAP Input] | Input | LDAPInput | Read data from LDAP host | opdts.ldapinput.LDAPInputMeta |
| [EAI:LDAP Output] | Output | LDAPOutput | Perform Insert, upsert, update, add or delete operations on records based on their DN (Distinguished Name). | opdts.ldapoutput.LDAPOutputMeta |
| [EAI:LDIF Input] | Input | LDIFInput | Read data from LDIF files | opdts.ldifinput.LDIFInputMeta |
| [Load file content in memory |EAI:Load file content in memory] | Input | LoadFileInput | Load file content in memory | opdts.loadfileinput.LoadFileInputMeta |
| [LucidDB Bulk Loader|EAI:LucidDB bulk loader]\\ | Deprecated \\ | | | |
| [LucidDB Streaming Loader |EAI:LucidDB Streaming Loader] | Bulk loading | LucidDBStreamingLoader | Load data into LucidDB by using Remote Rows UDX. | opdts.luciddbstreamingloader.LucidDBStreamingLoaderMeta |
| [Mail|EAI:Mail_transformation] [Mail|Mail (step)] | Utility | Mail | Send eMail. | opdts.mail.MailMeta |
| [EAI:Mail Validator] | Validation | MailValidator | Check if an email address is valid. | opdts.mailvalidator.MailValidatorMeta |
| [Mapping (sub-transformation)|EAI:Mapping] | Mapping | Mapping | Run a mapping (sub-transformation), use MappingInput and MappingOutput to specify the fields interface | opdts.mapping.MappingMeta |
| [Mapping input specification|EAI:Mapping Input] | Mapping | MappingInput | Specify the input interface of a mapping | opdts.mappinginput.MappingInputMeta |
| [Mapping output specification|EAI:Mapping Output] | Mapping | MappingOutput | Specify the output interface of a mapping | opdts.mappingoutput.MappingOutputMeta |
| [MapReduce Input|EAI:MapReduce Input] | Big Data | HadoopEnterPlugin | Key Value pairs enter here from Hadoop MapReduce | opdts.hadoopenter.HadoopEnterMeta |
| [MapReduce Output|EAI:MapReduce Output] | Big Data | HadoopExitPlugin | Key Value pairs exit here and are pushed into Hadoop MapReduce | opdts.hadoopexit.HadoopExitMeta |
| [EAI:MaxMind GeoIP Lookup] | Lookup | MaxMindGeoIPLookup | Lookup an IPv4 address in a MaxMind database and add fields such as geography, ISP, or organization. | com.maxmind.geoip.MaxMindGeoIPLookupMeta |
| [EAI:Memory Group by] | Statistics | MemoryGroupBy | Builds aggregates in a group by fashion. This step doesn't require sorted input. | opdts.memgroupby.MemoryGroupByMeta |
| [EAI:Merge Join] | Joins | MergeJoin | Joins two streams on a given key and outputs a joined set. The input streams must be sorted on the join key | opdts.mergejoin.MergeJoinMeta |
| [Merge Rows (diff)|EAI:Merge rows] | Joins | MergeRows | Merge two streams of rows, sorted on a certain key. The two streams are compared and the equals, changed, deleted and new rows are flagged. | opdts.mergerows.MergeRowsMeta |
| [EAI:Modified Java Script Value] | Scripting | ScriptValueMod | This steps allows the execution of JavaScript programs (and much more) \\ | opdts.scriptvalues_mod.ScriptValuesMetaMod |
| [Mondrian Input |EAI:Mondrian Input]\\ | Input | MondrianInput | Execute and retrieve data using an MDX query against a Pentaho Analyses OLAP server (Mondrian) | opdts.mondrianinput.MondrianInputMeta |
| [EAI:MonetDB Agile Mart]\\ | Agile \\ | | | |
| [MonetDB Bulk Loader|EAI:MonetDB bulk loader] | Bulk loading | MonetDBBulkLoader | Load data into MonetDB by using their bulk load command in streaming mode. | opdts.monetdbbulkloader.MonetDBBulkLoaderMeta |
| [EAI:MongoDB Input] | Input Big Data | MongoDbInput | Reads all entries from a MongoDB collection in the specified database. | opdts.mongodbinput.MongoDbInputMeta |
| [EAI:MongoDB Output] | Output Big Data | MongoDbOutput | Write to a MongoDB collection. | opdts.mongodboutput.MongoDbOutputMeta |
| {color:#c0c0c0}Multiway Merge Join{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}MultiwayMergeJoin{color} | {color:#c0c0c0}Multiway Merge Join{color} | {color:#c0c0c0}opdts.multimerge.MultiMergeJoinMeta{color} |
| {color:#000000}[EAI:Multiway Merge Join]{color} | {color:#000000}Joins{color} | {color:#000000}MultiwayMergeJoin{color} | {color:#000000}Multiway Merge Join{color} | {color:#000000}opdts.multimerge.MultiMergeJoinMeta{color} |
| [EAI:MySQL Bulk Loader] | Bulk loading | MySQLBulkLoader | MySQL bulk loader step, loading data over a named pipe (not available on MS Windows) | opdts.mysqlbulkloader.MySQLBulkLoaderMeta |
| [Null if...|EAI:Null If] | Utility | NullIf | Sets a field value to null if it is equal to a constant value | opdts.nullif.NullIfMeta |
| [EAI:Number range] | Transform | NumberRange | Create ranges based on numeric field | opdts.numberrange.NumberRangeMeta |
| [EAI:OLAP Input] | Input | OlapInput | Execute and retrieve data using an MDX query against any XML/A OLAP datasource using olap4j | opdts.olapinput.OlapInputMeta |
| [EAI:OpenERP Object Delete]\\ | Delete \\ | OpenERPObjectDelete | Deletes data from the OpenERP server using the XMLRPC interface with the 'unlink' function. \\ | opdts.openerp.objectdelete.OpenERPObjectDeleteMeta |
| [EAI:OpenERP Object Output]\\ | Output \\ | OpenERPObjectOutputImport | Updates data on the OpenERP server using the XMLRPC interface and the 'import' function \\ | opdts.openerp.objectoutput.OpenERPObjectOutputMeta |
| [EAI:Oracle Bulk Loader] | Bulk loading | OraBulkLoader | Use Oracle Bulk Loader to load data | opdts.orabulkloader.OraBulkLoaderMeta |
| [Output steps metrics|EAI:Output Steps Metrics] | Statistics | StepsMetrics | Return metrics for one or several steps | opdts.stepsmetrics.StepsMetricsMeta |
| [EAI:Palo Cell Input] | Input \\ | PaloCellInput | Retrieves all cell data from a Palo cube \\ | opdts.palo.cellinput |
| [EAI:Palo Cell Output]\\ | Output \\ | PaloCellOutput | Updates cell data in a Palo cube \\ | opdts.palo.celloutput |
| [EAI:Pentaho Reporting Output] | Output | PentahoReportingOutput | Executes an existing report (PRPT) | opdts.pentahoreporting.PentahoReportingOutputMeta |
| [EAI:PostgreSQL Bulk Loader] | Bulk loading | PGBulkLoader | PostgreSQL Bulk Loader | opdts.pgbulkloader.PGBulkLoaderMeta |
| [EAI:Prioritize streams] | Flow | PrioritizeStreams | Prioritize streams in an order way. | opdts.prioritizestreams.PrioritizeStreamsMeta |
| [EAI:Process files] | Utility | ProcessFiles | Process one file per row (copy or move or delete). This step only accept filename in input. | opdts.processfiles.ProcessFilesMeta |
| [EAI:Properties Output] | Output | PropertyOutput | Write data to properties file | opdts.propertyoutput.PropertyOutputMeta |
| [EAI:Property Input] | Input | PropertyInput | Read data (key, value) from properties files. | opdts.propertyinput.PropertyInputMeta |
| [EAI:R script executor]\\ | Statistics | RScriptExecutor \\ | Executes an R script within a PDI transformation | |
| [EAI:Regex Evaluation] | Scripting | RegexEval | Regular expression Evaluation. This step uses a regular expression to evaluate a field. It can also extract new fields out of an existing field with capturing groups. | opdts.regexeval.RegexEvalMeta |
| [Replace in string|EAI:Replace in String] | Transform | ReplaceString | Replace all occurences a word in a string with another word. | opdts.replacestring.ReplaceStringMeta |
| [Reservoir Sampling|DATAMINING:Using the Reservoir Sampling Plugin] [EAI:Reservoir Sampling] | Statistics | ReservoirSampling | Transform Samples a fixed number of rows from the incoming stream | opdts.reservoirsampling.ReservoirSamplingMeta |
| [REST Client|EAI:Rest Client] | Lookup | Rest | Consume RESTfull services. REpresentational State Transfer (REST) is a key design idiom that embraces a stateless client-server architecture in which the web services are viewed as resources and can be identified by their URLs | opdts.rest.RestMeta |
| [Row denormaliser|EAI:Row De-normalizer] denormaliser|Row denormaliser] | Transform | Denormaliser | Denormalises rows by looking up key-value pairs and by assigning them to new fields in the output rows. This method aggregates and needs the input rows to be sorted on the grouping fields | opdts.denormaliser.DenormaliserMeta |
| [Row flattener|EAI:Flattener] | Transform | Flattener | Flattens consecutive rows based on the order in which they appear in the input stream | opdts.flattener.FlattenerMeta |
| [Row Normaliser|EAI:Row Normalizer] Normaliser|Row Normaliser] | Transform | Normaliser | De-normalised information can be normalised using this step type. | opdts.normaliser.NormaliserMeta |
| [EAI:RSS Input] | Input | RssInput | Read RSS feeds | opdts.rssinput.RssInputMeta |
| [EAI:RSS Output] | Output | RssOutput | Read RSS stream. | opdts.rssoutput.RssOutputMeta |
| {color:#c0c0c0}[EAI:Rule Executor]{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}RuleExecutor{color} | {color:#c0c0c0}Execute a rule against each row{color} | {color:#c0c0c0}opdts.rules.RulesExecutorMeta{color} |
| {color:#c0c0c0}[EAI:Rule Accumulator]{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}RuleAccumulator{color} | {color:#c0c0c0}Execute a rule against a set of rows{color} | {color:#c0c0c0}opdts.rules.RulesAccumulatorMeta{color} |
| [EAI:Rule Executor] | Scripting | RuleExecutor | Execute a rule against each row (using Drools) | opdts.rules.RulesExecutorMeta |
| [EAI:Rule Accumulator] | Scripting | RuleAccumulator | Execute a rule against a set of rows (using Drools) | opdts.rules.RulesAccumulatorMeta |
| [EAI:Run SSH commands] | Utility | SSH | Run SSH commands and returns result. | opdts.ssh.SSHMeta |
| [S3 CSV Input |EAI:S3 CSV Input] | Input | S3CSVINPUT | S3 CSV Input | opdts.s3csvinput.S3CsvInputMeta |
| S3 File Output | Output | S3FileOutputPlugin | Create files in an S3 location | com.pentaho.amazon.s3.S3FileOutputMeta |
| [EAI:S3 File Output] | Output | S3FileOutputPlugin | Exports data to a text file on an Amazon Simple Storage Service (S3) | com.pentaho.amazon.s3.S3FileOutputMeta |
| [EAI:Salesforce Delete] | Output | SalesforceDelete | Delete records in Salesforce module. | opdts.salesforcedelete.SalesforceDeleteMeta |
| [Salesforce Input|EAI:SalesForce Input] | Input | SalesforceInput | Reads information from SalesForce \\ | opdts.salesforceinput.SalesforceInputMeta |
| [Salesforce Insert |EAI:Salesforce Insert] | Output | SalesforceInsert | Insert records in Salesforce module. | opdts.salesforceinsert.SalesforceInsertMeta |
| [EAI:Salesforce Update] | Output | SalesforceUpdate | Update records in Salesforce module. | opdts.salesforceupdate.SalesforceUpdateMeta |
| [Salesforce Upsert |EAI:Salesforce Upsert] | Output | SalesforceUpsert | Insert or update records in Salesforce module. | opdts.salesforceupsert.SalesforceUpsertMeta |
| [EAI:Sample rows] | Statistics | SampleRows | Filter rows based on the line number. | opdts.samplerows.SampleRowsMeta |
| [EAI:SAP Input] | Input | SapInput | Read data from SAP ERP, optionally with parameters | opdts.sapinput.SapInputMeta |
| [EAI:SAS Input] | Input | SASInput | This step reads files in sas7bdat (SAS) native format \\ | opdts.sasinput.SasInputMeta \\ |
| [EAI:Script]\\ | Experimental \\ | | | |
| {color:#c0c0c0}Secret key generator{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}SecretKeyGenerator{color} | {color:#c0c0c0}Generate secrete key for algorithms such as DES, AEC, TripleDES{color}. | {color:#c0c0c0}opdts.symmetriccrypto.secretkeygenerator.SecretKeyGeneratorMeta{color} |
| [Select values |EAI:Select Values]\\ | Transform | SelectValues | Select or remove fields in a row. Optionally, set the field meta-data: type, length and precision. | opdts.selectvalues.SelectValuesMeta |
| [Send message to Syslog |EAI:Send message to Syslog] | Utility | SyslogMessage | Send message to Syslog server | opdts.syslog.SyslogMessageMeta |
| [EAI:Serialize to file] | Output | CubeOutput | Write rows of data to a data cube | opdts.cubeoutput.CubeOutputMeta |
| [Set field value |EAI:Set field value] | Transform | SetValueField | Set Replace value of a field with another value field | opdts.setvaluefield.SetValueFieldMeta |
| [Set field value to a constant |EAI:Set field value to a constant] | Transform | SetValueConstant | Set Replace value of a field to a constant | opdts.setvalueconstant.SetValueConstantMeta |
| [EAI:Set files in result] | Job | FilesToResult | This step allows you to set filenames in the result of this transformation. Subsequent job entries can then use this information. | opdts.filestoresult.FilesToResultMeta |
| [EAI:Set Variables] | Job | SetVariable | Set environment variables based on a single input row. | opdts.setvariable.SetVariableMeta |
| [EAI:SFTP Put]\\ | Experimental \\ | | | |
| [EAI:Simple Mapping]\\ | Mapping \\ | SimpleMapping | Turn a repetitive, re-usable part of a transformation (a sequence of steps) into a mapping (sub-transformation). | opdts.simplemapping.SimpleMapping |
| [EAI:Single Threader] | Flow | SingleThreader | Executes a transformation snippet in a single thread. You need a standard mapping or a transformation with an Injector step where data from the parent transformation will arive in blocks. | opdts.singlethreader.SingleThreaderMeta |
| [EAI:Socket reader] | Inline | SocketReader | Socket reader. A socket client that connects to a server (Socket Writer step). | opdts.socketreader.SocketReaderMeta |
| [Sorted Merge |EAI:Sorted Merge]\\ | Joins | SortedMerge | Sorted Merge | opdts.sortedmerge.SortedMergeMeta |
| [EAI:Split field to rows] | Transform | SplitFieldToRows3 | Splits a single string field by delimiter and creates a new row for each split term | opdts.splitfieldtorows.SplitFieldToRowsMeta |
| [Split Fields|EAI:Field Splitter] Fields|Split Fields] | Transform | FieldSplitter | When you want to split a single field into more then one, use this step type. | opdts.fieldsplitter.FieldSplitterMeta |
| [Splunk Input|EAI:Splunk Input] | Transform | SplunkInput | Reads data from Splunk. | opdts.splunk.SplunkInputMeta |
| [Splunk Output|EAI:Splunk Output] | Transform | SplunkOutput | Writes data to Splunk. | opdts.splunk.SplunkOutputMeta |
| [EAI:SQL File Output] | Output | SQLFileOutput | Output SQL INSERT statements to file | opdts.sqlfileoutput.SQLFileOutputMeta |
| [Stream lookup|EAI:Stream Lookup] | Lookup | StreamLookup | Look up values coming from another stream in the transformation. | opdts.streamlookup.StreamLookupMeta |
| [SSTable Output|EAI:SSTable Output] | Big Data | SSTableOutput | writes to a filesystem directory as a Cassandra SSTable | opdts.cassandrasstableoutput.SSTableOutputMeta |
| [EAI:Streaming XML Input]\\ | Deprecated \\ | | | |
| [EAI:String operations] | Transform | StringOperations | Apply certain operations like trimming, padding and others to string value. | opdts.stringoperations.StringOperationsMeta |
| [EAI:Strings cut] | Transform | StringCut | Strings cut (substring). | opdts.stringcut.StringCutMeta |
| [Switch / Case|EAI:Switch-Case] | Flow | SwitchCase | Switch a row to a certain target step based on the case value in a field. | opdts.switchcase.SwitchCaseMeta |
| {color:#c0c0c0}Symmetric Cryptography{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}SymmetricCryptoTrans{color} | {color:#c0c0c0}Encrypt or decrypt a string using symmetric encryption. Available algorithms are DES, AEC, TripleDES.{color} | {color:#c0c0c0}opdts.symmetriccrypto.symmetriccryptotrans.SymmetricCryptoTransMeta{color} |
| [EAI:Synchronize after merge] | Output | SynchronizeAfterMerge | This step perform insert/update/delete in one go based on the value of a field. | opdts.synchronizeaftermerge.SynchronizeAfterMergeMeta |
| [EAI:Table Agile Mart]\\ | Agile \\ | | | |
| [EAI:Table Compare]\\ | Utility \\ | TableCompare | This step compares the data from two tables (provided they have the same lay-out). It'll find differences between the data in the two tables and log it. | opdts.tablecompare.TableCompareMeta |
| [Table exists|EAI:Table Exists] | Lookup | TableExists | Check if a table exists on a specified connection | opdts.tableexists.TableExistsMeta |
| [Table input|EAI:Table Input] | Input | TableInput | Read information from a database table. | opdts.tableinput.TableInputMeta |
| [Table output |EAI:Table Output]\\ | Output | TableOutput | Write information to a database table | opdts.tableoutput.TableOutputMeta |
| [EAI:Teradata Fastload Bulk Loader] | Bulk loading | TeraFast | The Teradata Fastload Bulk loader | opdts.terafast.TeraFastMeta |
| [EAI:Teradata TPT Insert Upsert Bulk Loader] | Bulk loading | TeraDataBulkLoader | Bulk loading via TPT using the tbuild command. | |
| [Text file input|EAI:Text File Input] | Input | TextFileInput | Read data from a text file in several formats. This data can then be passed on to the next step(s)... | opdts.textfileinput.TextFileInputMeta |
| [Text file output|EAI:Text File Output] | Output | TextFileOutput | Write rows to a text file. | opdts.textfileoutput.TextFileOutputMeta |
| [EAI:Transformation Executor]\\ | Flow \\ | | | |
| [EAI:Transformation Executor]\\ | Flow \\ | | | |
| [Unique rows|EAI:Unique Rows] | Transform | Unique | Remove double rows and leave only unique occurrences. This works only on a sorted input. If the input is not sorted, only double consecutive rows are handled correctly. | opdts.uniquerows.UniqueRowsMeta |
| [Unique rows (HashSet) |EAI:Unique Rows (HashSet)] | Transform | UniqueRowsByHashSet | Remove double rows and leave only unique occurrences by using a HashSet. | opdts.uniquerowsbyhashset.UniqueRowsByHashSetMeta |
| [Univariate Statistics|DATAMINING:Using the Univariate Statistics Plugin] | Statistics | UnivariateStats | This step computes some simple stats based on a single input field | opdts.univariatestats.UnivariateStatsMeta |
| [EAI:Update] | Output | Update | Update data in a database table based upon keys | opdts.update.UpdateMeta |
| [EAI:User Defined Java Expression] | Scripting | Janino | Calculate the result of a Java Expression using Janino | opdts.janino.JaninoMeta |
| [EAI:Value Mapper] | Transform | ValueMapper | Maps values of a certain field from one value to another | opdts.valuemapper.ValueMapperMeta |
| [EAI:Vertica Bulk Loader]\\ | Bulk loading | VerticaBulkLoader | Bulk loads data into a Vertica table using their high performance COPY feature \\ | opdts.verticabulkload.VerticaBulkLoaderMeta |
| [EAI:Web services lookup] | Lookup | WebServiceLookup | Look up information using web services (WSDL) | opdts.webservices.WebServiceMeta |
| [EAI:Knowledge Flow] | Data Mining | KF | Executes a Knowledge Flow data mining process | org.pentaho.di.kf.KFMeta |
| [Write to log |EAI:Write to log (step)]\\ | Utility | WriteToLog | Write data to log | opdts.writetolog.WriteToLogMeta |
| [XBase input|EAI:XBase Input] | Input | XBaseInput | Reads records from an XBase type of database file (DBF) | opdts.xbaseinput.XBaseInputMeta |
| [XML Input Stream (StAX)|EAI:XML Input Stream (StAX)] | Input | XMLInputStream | This step is capable of processing very large and complex XML files very fast. | opdts.xmlinputstream.XMLInputStreamMeta |
| [EAI:XML Input]\\ | Deprecated \\ | | | |
| [EAI:XML Join] | Joins | XMLJoin | Joins a stream of XML-Tags into a target XML string | opdts.xmljoin.XMLJoinMeta |
| [EAI:XML Output] | Output | XMLOutput | Write data to an XML file | opdts.xmloutput.XMLOutputMeta |
| [EAI:XSD Validator] | Validation | XSDValidator | Validate XML source (files or streams) against XML Schema Definition. | opdts.xsdvalidator.XsdValidatorMeta |
| [EAI:XSL Transformation] | Transform | XSLT | Transform XML stream using XSL (eXtensible Stylesheet Language). | opdts.xslt.XsltMeta |
| [Yaml Input |EAI:Yaml Input] | Input | YamlInput | Read YAML source (file or stream) parse them and convert them to rows and writes these to one or more output. | opdts.yamlinput.YamlInputMeta |
| [Zip File|EAI:Zip file (step)]\\ | Utility \\ | ZipFile | Creates a standard ZIP archive from the data stream fields \\ | opdts.zipfile.ZipFileMeta |