Access Keys:
Skip to content (Access Key - 0)

compared with
Current by ravi nanganay
on Dec 15, 2014 14:05.

Key
This line was removed.
This word was removed. This word was added.
This line was added.

Changes (83)

View Page History
| [EAI:Add value fields changing sequence] | Transform | FieldsChangeSequence | Add sequence depending of fields value change. Each time value of at least one field change, PDI will reset sequence. | opdts.fieldschangesequence.FieldsChangeSequenceMeta |
| [EAI:Add XML] | Transform | AddXML | Encode several fields into an XML fragment | opdts.addxml.AddXMLMeta |
| [EAI:Aggregate Rows]\\ | Deprecated \\ | | | |
| [EAI:Analytic Query] | Statistics | AnalyticQuery | Execute analytic queries over a sorted dataset (LEAD/LAG/FIRST/LAST) | opdts.analyticquery.AnalyticQueryMeta |
| [EAI:Append streams] | Flow | Append | Append 2 streams in an ordered way | opdts.append.AppendMeta |
| [EAI:Arff Output] | Data Mining | Arff Output | Writes data in ARFF format to a file | opdts.append.arff.ArffOutputMeta |
| [EAI:Automatic Documentation Output] | Output | AutoDoc | This step automatically generates documentation based on input in the form of a list of transformations and jobs | opdts.autodoc.AutoDocMeta |
| [Avro input|EAI:Avro Input] | Input | AvroInput | Decode binary or Json Avro data from a file or a field | opdts.avroinput.AvroInputMeta |
| [EAI:Call DB Procedure] | Lookup | DBProc | Get back information by calling a database procedure. | opdts.dbproc.DBProcMeta |
| [EAI:Change file encoding] | Utility | ChangeFileEncoding | Change file encoding and create a new file | opdts.changefileencoding.ChangeFileEncodingMeta |
| [Cassandra input|EAI:Cassandra Input] | Input Big Data | CassandraInput | Read from a Cassandra column family | opdts.cassandrainput.CassandraInputMeta |
| [Cassandra output|EAI:Cassandra Output] | Output Big Data | CassandraOutput | Write to a Cassandra column family | opdts.cassandraoutput.CassandraOutputMeta |
| [EAI:Check if a column exists] | Lookup | ColumnExists | Check if a column exists in a table on a specified connection. | opdts.columnexists.ColumnExistsMeta |
| [EAI:Check if file is locked] | Lookup | FileLocked | Check if a file is locked by another process | opdts.filelocked.FileLockedMeta |
| [EAI:Closure Generator] | Transform | ClosureGenerator | This step allows you to generates a closure table using parent-child relationships. | opdts.closure.ClosureGeneratorMeta |
| [Combination lookup/update|EAI:Combination lookup-update] | Data Warehouse | CombinationLookup | Update a junk dimension in a data warehouse. Alternatively, look up information in this dimension. The primary key of a junk dimension are all the fields. | opdts.combinationlookup.CombinationLookupMeta |
| [EAI:Concat Fields]\\ | Transform | ConcatFields \\ | The Concat Fields step is used to concatenate multiple fields into one target field. The fields can be separated by a separator and the enclosure logic is completely compatible with the Text File Output step. | opdts.concatfields.ConcatFieldsMeta |
| [EAI:Copy rows to result] | Job | RowsToResult | Use this step to write rows to the executing job. The information will then be passed to the next entry in this job. | opdts.rowstoresult.RowsToResultMeta |
| [EAI:CouchDB Input] | Big Data | CouchDbInput | Retrieves all documents from a given view in a given design document from a given database | opdts.couchdbinput.CouchDbInputMeta |
| [EAI:Credit card validator] | Validation | CreditCardValidator | The Credit card validator step will help you tell: (1) if a credit card number is valid (uses LUHN10 (MOD-10) algorithm) (2) which credit card vendor handles that number (VISA, MasterCard, Diners Club, EnRoute, American Express (AMEX),...) | opdts.creditcardvalidator.CreditCardValidatorMeta |
| [CSV file input|EAI:CSV Input] | Input | CsvInput | Simple CSV file input | opdts.csvinput.CsvInputMeta |
| [EAI:Dynamic SQL row] | Lookup | DynamicSQLRow | Execute dynamic SQL statement build in a previous field | opdts.dynamicsqlrow.DynamicSQLRowMeta |
| [EAI:Edi to XML]\\ | Utility | TypeExitEdi2XmlStep | Converts an Edifact message to XML to simplify data extraction (Available in PDI 4.4, already present in [CI trunk|http://ci.pentaho.com/view/Data%20Integration/job/Kettle/] builds) | opdts.edi2xml.Edi2XmlMeta |
| [ElasticSearch Bulk Insert |EAI:ElasticSearch Bulk Insert] | Bulk loading | ElasticSearchBulk | Performs bulk inserts into ElasticSearch | opdts.elasticsearchbulk.ElasticSearchBulkMeta |
| [Email messages input |EAI:Email Messages Input] | Input | MailInput | Read POP3/IMAP server and retrieve messages | opdts.mailinput.MailInputMeta |
| [ESRI Shapefile Reader |EAI:ESRI Shapefile Reader] | Input | ShapeFileReader | Reads shape file data from an ESRI shape file and linked DBF file | org.pentaho.di.shapefilereader.ShapeFileReaderMeta |
| [EAI:ETL Metadata Injection] | Flow | MetaInject | This step allows you to inject metadata into an existing transformation prior to execution. This allows for the creation of dynamic and highly flexible data integration solutions. | opdts.metainject.MetaInjectMeta |
| [Example plugin|EAI:Example plugin (Transformation Step)] | Transform | DummyPlugin | This is an example for a plugin test step | be.ibridge.kettle.dummy.DummyPluginMeta |
| [EAI:Execute a process] | Utility | ExecProcess | Execute a process and return the result | opdts.execprocess.ExecProcessMeta |
| [EAI:Execute row SQL script] | Scripting | ExecSQLRow | Execute SQL script extracted from a field created in a previous step. | opdts.execsqlrow.ExecSQLRowMeta |
| [EAI:Execute SQL script] | Scripting | ExecSQL | Execute an SQL script, optionally parameterized using input rows | opdts.sql.ExecSQLMeta |
| [EAI:File exists] | Lookup | FileExists | Check if a file exists | opdts.fileexists.FileExistsMeta |
| [EAI:Filter rows] [Filter Rows] | Flow | FilterRows | Filter rows using simple equations | opdts.filterrows.FilterRowsMeta |
| [Fixed file input|EAI:Fixed File Input] | Input | FixedInput | Fixed file input | opdts.fixedinput.FixedInputMeta |
| [EAI:Formula] | Scripting | Formula | Calculate a formula using Pentaho's libformula | opdts.formula.FormulaMeta |
| [EAI:Fuzzy match] | Lookup | FuzzyMatch | Finding approximate matches to a string using matching algorithms. Read a field from a main stream and output approximative value from lookup stream. | opdts.fuzzymatch.FuzzyMatchMeta |
| [EAI:Generate random credit card numbers] | Input | RandomCCNumberGenerator | Generate random valide (luhn check) credit card numbers | opdts.randomccnumber.RandomCCNumberGeneratorMeta |
| [Generate random value|EAI:Generate Random Value] | Input | RandomValue | Generate random value | opdts.randomvalue.RandomValueMeta |
| [EAI:Get Files Rows Count] | Input | GetFilesRowsCount | Get Files Rows Count | opdts.getfilesrowscount.GetFilesRowsCountMeta |
| [Get ID from slave server|EAI:Get ID from Slave Server] | Transform | GetSlaveSequence | Retrieves unique IDs in blocks from a slave server. The referenced sequence needs to be configured on the slave server in the XML configuration file. | opdts.getslavesequence.GetSlaveSequenceMeta |
| [EAI:Get previous row fields]\\ | Deprecated \\ | | | |
| [EAI:Get repository names] | Input | GetRepositoryNames | Lists detailed information about transformations and/or jobs in a repository | opdts.getrepositorynames.GetRepositoryNamesMeta |
| [EAI:Get rows from result] | Job | RowsFromResult | This allows you to read rows from a previous entry in a job. | opdts.rowsfromresult.RowsFromResultMeta |
| [Get SubFolder names |EAI:Get SubFolder names] | Input | GetSubFolders | Read a parent folder and return all subfolders | opdts.getsubfolders.GetSubFoldersMeta |
| [EAI:Get System Info] | Input | SystemInfo | Get information from the system like system date, arguments, etc. | opdts.systemdata.SystemDataMeta |
| [EAI:Get table names] | Input | GetTableNames | Get table names from database connection and send them to the next step | opdts.gettablenames.GetTableNamesMeta |
| [Get Variables|EAI:Get Variable] | Job | GetVariable | Determine the values of certain (environment or Kettle) variables and put them in field values. | opdts.getvariable.GetVariableMeta |
| [EAI:Google Analytics] | Input | TypeExitGoogleAnalyticsInputStep | Fetches data from google analytics account | opdts.googleanalytics.GaInputStepMeta |
| [EAI:Google Docs Input]\\ | Input \\ | | | |
| [Greenplum Bulk Loader|PMOPEN:GreenPlum bulk loader] | Bulk loading | GPBulkLoader | Greenplum Bulk Loader | opdts.gpbulkloader.GPBulkLoaderMeta |
| [EAI:Greenplum Load]\\ | Bulk loading \\ | GPLoad \\ | Greenplum Load \\ | |
| [Group by|EAI:Group By] | Statistics | GroupBy | Builds aggregates in a group by fashion. This works only on a sorted input. If the input is not sorted, only double consecutive rows are handled correctly. | opdts.groupby.GroupByMeta |
| [EAI:GZIP CSV Input] | Input | ParallelGzipCsvInput | Parallel GZIP CSV file input reader | opdts.parallelgzipcsv.ParGzipCsvInputMeta |
| [EAI:Hadoop File Input] | Big Data | HadoopFileInputPlugin | Read data from a variety of different text-file types stored on a Hadoop cluster | opdts.hadoopfileinput.HadoopFileInputMeta |
| [EAI:Hadoop File Output] | Big Data | HadoopFileOutputPlugin | Write data to a variety of different text-file types stored on a Hadoop cluster | opdts.hadoopfileoutput.HadoopFileOutputMeta |
| [HBase input|EAI:HBase Input] | Input Big Data | HbaseInput | Read from an HBase column family | opdts.hbaseinput.HBaseInputMeta |
| [HBase output|EAI:HBase Output] | Output Big Data | HbaseOutput | Write to an HBase column family | opdts.hbaseoutput.HBaseOutputMeta |
| [EAI:HBase Row Decoder] | Big Data | HBaseRowDecoder | Decodes an incoming key and HBase result object according to a mapping | opdts.hbaserowdecoder.HBaseRowDecoderMeta |
| [EAI:HL7 Input]\\ | Input \\ | HL7Input | Read data from HL7 data streams. | opdt.hl7.plugins.hl7input |
| [HTTP client|EAI:HTTP Client] | Lookup | HTTP | Call a web service over HTTP by supplying a base URL by allowing parameters to be set dynamically | opdts.http.HTTPMeta |
| [EAI:HTTP Post] | Lookup | HTTPPOST | Call a web service request over HTTP by supplying a base URL by allowing parameters to be set dynamically | opdts.httppost.HTTPPOSTMeta |
| [EAI:IBM Websphere MQ Consumer] | Input \\ | MQInput | Receive messages from any IBM Websphere MQ Server | |
| [EAI:IBM Websphere MQ Producer] | Output \\ | MQOutput | Send messages to any IBM Websphere MQ Server | |
| [Identify last row in a stream |EAI:Identify last row in a stream] | Flow | DetectLastRow | Last row will be marked | opdts.detectlastrow.DetectLastRowMeta |
| [EAI:If field value is null] | Utility | IfNull | Sets a field value to a constant if it is null. | opdts.ifnull.IfNullMeta |
| Infobright Loader [EAI:Infobright Loader] | Bulk loading | InfobrightOutput | Load data to an Infobright database table | opdts.infobrightoutput.InfobrightLoaderMeta |
| [EAI:Ingres VectorWise Bulk Loader] | Bulk loading | VectorWiseBulkLoader | This step interfaces with the Ingres VectorWise Bulk Loader "COPY TABLE" command. | opdts.ivwloader.IngresVectorwiseLoaderMeta |
| [EAI:Injector] | Inline | Injector | Injector step to allow to inject rows into the transformation through the java API | opdts.injector.InjectorMeta |
| [Insert / Update|EAI:Insert - Update] | Output | InsertUpdate | Update or insert rows in a database based upon keys. | opdts.insertupdate.InsertUpdateMeta |
| [EAI:Java Filter] | Flow | JavaFilter | Filter rows using java code | opdts.javafilter.JavaFilterMeta |
| [EAI:JMS Consumer]\\ | Input \\ | JmsInput | Receive messages from a JMS server | |
| [EAI:JMS Producer]\\ | Output \\ | JmsOutput \\ | Send messages to a JMS server | |
| [EAI:Job Executor]\\ | Flow \\ | JobExecutor \\ | This step executes a Pentaho Data Integration Job, passes parameters and rows. \\ | opdts.jobexecutor.JobExecutorMeta \\ |
| [Join Rows (cartesian product) |EAI:Join Rows (Cartesian product)]\\ | Joins | JoinRows | The output of this step is the cartesian product of the input streams. The number of rows is the multiplication of the number of rows in the input streams. | opdts.joinrows.JoinRowsMeta |
| [EAI:Json Input] | Input | JsonInput | Extract relevant portions out of JSON structures (file or incoming field) and output rows | opdts.jsoninput.JsonInputMeta |
| Json output [JSON output] | Output | JsonOutput | Create Json bloc and output it in a field ou a file. | opdts.jsonoutput.JsonOutputMeta |
| [EAI:Knowledge Flow] | Data Mining | KF | Executes a Knowledge Flow data mining process | org.pentaho.di.kf.KFMeta |
| [EAI:LDAP Input] | Input | LDAPInput | Read data from LDAP host | opdts.ldapinput.LDAPInputMeta |
| [EAI:LDAP Output] | Output | LDAPOutput | Perform Insert, upsert, update, add or delete operations on records based on their DN (Distinguished Name). | opdts.ldapoutput.LDAPOutputMeta |
| [EAI:LDIF Input] | Input | LDIFInput | Read data from LDIF files | opdts.ldifinput.LDIFInputMeta |
| [Load file content in memory |EAI:Load file content in memory] | Input | LoadFileInput | Load file content in memory | opdts.loadfileinput.LoadFileInputMeta |
| [LucidDB Bulk Loader|EAI:LucidDB bulk loader]\\ | Deprecated \\ | | | |
| [LucidDB Streaming Loader |EAI:LucidDB Streaming Loader] | Bulk loading | LucidDBStreamingLoader | Load data into LucidDB by using Remote Rows UDX. | opdts.luciddbstreamingloader.LucidDBStreamingLoaderMeta |
| [Mail|EAI:Mail_transformation] [Mail|Mail (step)] | Utility | Mail | Send eMail. | opdts.mail.MailMeta |
| [EAI:Mail Validator] | Validation | MailValidator | Check if an email address is valid. | opdts.mailvalidator.MailValidatorMeta |
| [Mapping (sub-transformation)|EAI:Mapping] | Mapping | Mapping | Run a mapping (sub-transformation), use MappingInput and MappingOutput to specify the fields interface | opdts.mapping.MappingMeta |
| [Mapping input specification|EAI:Mapping Input] | Mapping | MappingInput | Specify the input interface of a mapping | opdts.mappinginput.MappingInputMeta |
| [Mapping output specification|EAI:Mapping Output] | Mapping | MappingOutput | Specify the output interface of a mapping | opdts.mappingoutput.MappingOutputMeta |
| [MapReduce Input|EAI:MapReduce Input] | Big Data | HadoopEnterPlugin | Key Value pairs enter here from Hadoop MapReduce | opdts.hadoopenter.HadoopEnterMeta |
| [MapReduce Output|EAI:MapReduce Output] | Big Data | HadoopExitPlugin | Key Value pairs exit here and are pushed into Hadoop MapReduce | opdts.hadoopexit.HadoopExitMeta |
| [EAI:MaxMind GeoIP Lookup] | Lookup | MaxMindGeoIPLookup | Lookup an IPv4 address in a MaxMind database and add fields such as geography, ISP, or organization. | com.maxmind.geoip.MaxMindGeoIPLookupMeta |
| [EAI:Memory Group by] | Statistics | MemoryGroupBy | Builds aggregates in a group by fashion. This step doesn't require sorted input. | opdts.memgroupby.MemoryGroupByMeta |
| [EAI:Merge Join] | Joins | MergeJoin | Joins two streams on a given key and outputs a joined set. The input streams must be sorted on the join key | opdts.mergejoin.MergeJoinMeta |
| [Merge Rows (diff)|EAI:Merge rows] | Joins | MergeRows | Merge two streams of rows, sorted on a certain key. The two streams are compared and the equals, changed, deleted and new rows are flagged. | opdts.mergerows.MergeRowsMeta |
| [EAI:Modified Java Script Value] | Scripting | ScriptValueMod | This steps allows the execution of JavaScript programs (and much more) \\ | opdts.scriptvalues_mod.ScriptValuesMetaMod |
| [Mondrian Input |EAI:Mondrian Input]\\ | Input | MondrianInput | Execute and retrieve data using an MDX query against a Pentaho Analyses OLAP server (Mondrian) | opdts.mondrianinput.MondrianInputMeta |
| [EAI:MonetDB Agile Mart]\\ | Agile \\ | | | |
| [MonetDB Bulk Loader|EAI:MonetDB bulk loader] | Bulk loading | MonetDBBulkLoader | Load data into MonetDB by using their bulk load command in streaming mode. | opdts.monetdbbulkloader.MonetDBBulkLoaderMeta |
| [EAI:MongoDB Input] | Input Big Data | MongoDbInput | Reads all entries from a MongoDB collection in the specified database. | opdts.mongodbinput.MongoDbInputMeta |
| [EAI:MongoDB Output] | Output Big Data | MongoDbOutput | Write to a MongoDB collection. | opdts.mongodboutput.MongoDbOutputMeta |
| {color:#c0c0c0}Multiway Merge Join{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}MultiwayMergeJoin{color} | {color:#c0c0c0}Multiway Merge Join{color} | {color:#c0c0c0}opdts.multimerge.MultiMergeJoinMeta{color} |
| {color:#000000}[EAI:Multiway Merge Join]{color} | {color:#000000}Joins{color} | {color:#000000}MultiwayMergeJoin{color} | {color:#000000}Multiway Merge Join{color} | {color:#000000}opdts.multimerge.MultiMergeJoinMeta{color} |
| [EAI:MySQL Bulk Loader] | Bulk loading | MySQLBulkLoader | MySQL bulk loader step, loading data over a named pipe (not available on MS Windows) | opdts.mysqlbulkloader.MySQLBulkLoaderMeta |
| [Null if...|EAI:Null If] | Utility | NullIf | Sets a field value to null if it is equal to a constant value | opdts.nullif.NullIfMeta |
| [EAI:Number range] | Transform | NumberRange | Create ranges based on numeric field | opdts.numberrange.NumberRangeMeta |
| [EAI:OLAP Input] | Input | OlapInput | Execute and retrieve data using an MDX query against any XML/A OLAP datasource using olap4j | opdts.olapinput.OlapInputMeta |
| [EAI:OpenERP Object Delete]\\ | Delete \\ | OpenERPObjectDelete | Deletes data from the OpenERP server using the XMLRPC interface with the 'unlink' function. \\ | opdts.openerp.objectdelete.OpenERPObjectDeleteMeta |
| [EAI:OpenERP Object Output]\\ | Output \\ | OpenERPObjectOutputImport | Updates data on the OpenERP server using the XMLRPC interface and the 'import' function \\ | opdts.openerp.objectoutput.OpenERPObjectOutputMeta |
| [EAI:Oracle Bulk Loader] | Bulk loading | OraBulkLoader | Use Oracle Bulk Loader to load data | opdts.orabulkloader.OraBulkLoaderMeta |
| [Output steps metrics|EAI:Output Steps Metrics] | Statistics | StepsMetrics | Return metrics for one or several steps | opdts.stepsmetrics.StepsMetricsMeta |
| [EAI:Palo Cell Input] | Input \\ | PaloCellInput | Retrieves all cell data from a Palo cube \\ | opdts.palo.cellinput |
| [EAI:Palo Cell Output]\\ | Output \\ | PaloCellOutput | Updates cell data in a Palo cube \\ | opdts.palo.celloutput |
| [EAI:Pentaho Reporting Output] | Output | PentahoReportingOutput | Executes an existing report (PRPT) | opdts.pentahoreporting.PentahoReportingOutputMeta |
| [EAI:PostgreSQL Bulk Loader] | Bulk loading | PGBulkLoader | PostgreSQL Bulk Loader | opdts.pgbulkloader.PGBulkLoaderMeta |
| [EAI:Prioritize streams] | Flow | PrioritizeStreams | Prioritize streams in an order way. | opdts.prioritizestreams.PrioritizeStreamsMeta |
| [EAI:Process files] | Utility | ProcessFiles | Process one file per row (copy or move or delete). This step only accept filename in input. | opdts.processfiles.ProcessFilesMeta |
| [EAI:Properties Output] | Output | PropertyOutput | Write data to properties file | opdts.propertyoutput.PropertyOutputMeta |
| [EAI:Property Input] | Input | PropertyInput | Read data (key, value) from properties files. | opdts.propertyinput.PropertyInputMeta |
| [EAI:R script executor]\\ | Statistics | RScriptExecutor \\ | Executes an R script within a PDI transformation | |
| [EAI:Regex Evaluation] | Scripting | RegexEval | Regular expression Evaluation. This step uses a regular expression to evaluate a field. It can also extract new fields out of an existing field with capturing groups. | opdts.regexeval.RegexEvalMeta |
| [Replace in string|EAI:Replace in String] | Transform | ReplaceString | Replace all occurences a word in a string with another word. | opdts.replacestring.ReplaceStringMeta |
| [Reservoir Sampling|DATAMINING:Using the Reservoir Sampling Plugin] [EAI:Reservoir Sampling] | Statistics | ReservoirSampling | Transform Samples a fixed number of rows from the incoming stream | opdts.reservoirsampling.ReservoirSamplingMeta |
| [REST Client|EAI:Rest Client] | Lookup | Rest | Consume RESTfull services. REpresentational State Transfer (REST) is a key design idiom that embraces a stateless client-server architecture in which the web services are viewed as resources and can be identified by their URLs | opdts.rest.RestMeta |
| [Row denormaliser|EAI:Row De-normalizer] denormaliser|Row denormaliser] | Transform | Denormaliser | Denormalises rows by looking up key-value pairs and by assigning them to new fields in the output rows. This method aggregates and needs the input rows to be sorted on the grouping fields | opdts.denormaliser.DenormaliserMeta |
| [Row flattener|EAI:Flattener] | Transform | Flattener | Flattens consecutive rows based on the order in which they appear in the input stream | opdts.flattener.FlattenerMeta |
| [Row Normaliser|EAI:Row Normalizer] Normaliser|Row Normaliser] | Transform | Normaliser | De-normalised information can be normalised using this step type. | opdts.normaliser.NormaliserMeta |
| [EAI:RSS Input] | Input | RssInput | Read RSS feeds | opdts.rssinput.RssInputMeta |
| [EAI:RSS Output] | Output | RssOutput | Read RSS stream. | opdts.rssoutput.RssOutputMeta |
| {color:#c0c0c0}[EAI:Rule Executor]{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}RuleExecutor{color} | {color:#c0c0c0}Execute a rule against each row{color} | {color:#c0c0c0}opdts.rules.RulesExecutorMeta{color} |
| {color:#c0c0c0}[EAI:Rule Accumulator]{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}RuleAccumulator{color} | {color:#c0c0c0}Execute a rule against a set of rows{color} | {color:#c0c0c0}opdts.rules.RulesAccumulatorMeta{color} |
| [EAI:Rule Executor] | Scripting | RuleExecutor | Execute a rule against each row (using Drools) | opdts.rules.RulesExecutorMeta |
| [EAI:Rule Accumulator] | Scripting | RuleAccumulator | Execute a rule against a set of rows (using Drools) | opdts.rules.RulesAccumulatorMeta |
| [EAI:Run SSH commands] | Utility | SSH | Run SSH commands and returns result. | opdts.ssh.SSHMeta |
| [S3 CSV Input |EAI:S3 CSV Input] | Input | S3CSVINPUT | S3 CSV Input | opdts.s3csvinput.S3CsvInputMeta |
| S3 File Output | Output | S3FileOutputPlugin | Create files in an S3 location | com.pentaho.amazon.s3.S3FileOutputMeta |
| [EAI:S3 File Output] | Output | S3FileOutputPlugin | Exports data to a text file on an Amazon Simple Storage Service (S3) | com.pentaho.amazon.s3.S3FileOutputMeta |
| [EAI:Salesforce Delete] | Output | SalesforceDelete | Delete records in Salesforce module. | opdts.salesforcedelete.SalesforceDeleteMeta |
| [Salesforce Input|EAI:SalesForce Input] | Input | SalesforceInput | Reads information from SalesForce \\ | opdts.salesforceinput.SalesforceInputMeta |
| [Salesforce Insert |EAI:Salesforce Insert] | Output | SalesforceInsert | Insert records in Salesforce module. | opdts.salesforceinsert.SalesforceInsertMeta |
| [EAI:Salesforce Update] | Output | SalesforceUpdate | Update records in Salesforce module. | opdts.salesforceupdate.SalesforceUpdateMeta |
| [Salesforce Upsert |EAI:Salesforce Upsert] | Output | SalesforceUpsert | Insert or update records in Salesforce module. | opdts.salesforceupsert.SalesforceUpsertMeta |
| [EAI:Sample rows] | Statistics | SampleRows | Filter rows based on the line number. | opdts.samplerows.SampleRowsMeta |
| [EAI:SAP Input] | Input | SapInput | Read data from SAP ERP, optionally with parameters | opdts.sapinput.SapInputMeta |
| [EAI:SAS Input] | Input | SASInput | This step reads files in sas7bdat (SAS) native format \\ | opdts.sasinput.SasInputMeta \\ |
| [EAI:Script]\\ | Experimental \\ | | | |
| {color:#c0c0c0}Secret key generator{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}SecretKeyGenerator{color} | {color:#c0c0c0}Generate secrete key for algorithms such as DES, AEC, TripleDES{color}. | {color:#c0c0c0}opdts.symmetriccrypto.secretkeygenerator.SecretKeyGeneratorMeta{color} |
| [Select values |EAI:Select Values]\\ | Transform | SelectValues | Select or remove fields in a row. Optionally, set the field meta-data: type, length and precision. | opdts.selectvalues.SelectValuesMeta |
| [Send message to Syslog |EAI:Send message to Syslog] | Utility | SyslogMessage | Send message to Syslog server | opdts.syslog.SyslogMessageMeta |
| [EAI:Serialize to file] | Output | CubeOutput | Write rows of data to a data cube | opdts.cubeoutput.CubeOutputMeta |
| [Set field value |EAI:Set field value] | Transform | SetValueField | Set Replace value of a field with another value field | opdts.setvaluefield.SetValueFieldMeta |
| [Set field value to a constant |EAI:Set field value to a constant] | Transform | SetValueConstant | Set Replace value of a field to a constant | opdts.setvalueconstant.SetValueConstantMeta |
| [EAI:Set files in result] | Job | FilesToResult | This step allows you to set filenames in the result of this transformation. Subsequent job entries can then use this information. | opdts.filestoresult.FilesToResultMeta |
| [EAI:Set Variables] | Job | SetVariable | Set environment variables based on a single input row. | opdts.setvariable.SetVariableMeta |
| [EAI:SFTP Put]\\ | Experimental \\ | | | |
| [EAI:Simple Mapping]\\ | Mapping \\ | SimpleMapping | Turn a repetitive, re-usable part of a transformation (a sequence of steps) into a mapping (sub-transformation). | opdts.simplemapping.SimpleMapping |
| [EAI:Single Threader] | Flow | SingleThreader | Executes a transformation snippet in a single thread. You need a standard mapping or a transformation with an Injector step where data from the parent transformation will arive in blocks. | opdts.singlethreader.SingleThreaderMeta |
| [EAI:Socket reader] | Inline | SocketReader | Socket reader. A socket client that connects to a server (Socket Writer step). | opdts.socketreader.SocketReaderMeta |
| [Sorted Merge |EAI:Sorted Merge]\\ | Joins | SortedMerge | Sorted Merge | opdts.sortedmerge.SortedMergeMeta |
| [EAI:Split field to rows] | Transform | SplitFieldToRows3 | Splits a single string field by delimiter and creates a new row for each split term | opdts.splitfieldtorows.SplitFieldToRowsMeta |
| [Split Fields|EAI:Field Splitter] Fields|Split Fields] | Transform | FieldSplitter | When you want to split a single field into more then one, use this step type. | opdts.fieldsplitter.FieldSplitterMeta |
| [Splunk Input|EAI:Splunk Input] | Transform | SplunkInput | Reads data from Splunk. | opdts.splunk.SplunkInputMeta |
| [Splunk Output|EAI:Splunk Output] | Transform | SplunkOutput | Writes data to Splunk. | opdts.splunk.SplunkOutputMeta |
| [EAI:SQL File Output] | Output | SQLFileOutput | Output SQL INSERT statements to file | opdts.sqlfileoutput.SQLFileOutputMeta |
| [Stream lookup|EAI:Stream Lookup] | Lookup | StreamLookup | Look up values coming from another stream in the transformation. | opdts.streamlookup.StreamLookupMeta |
| [SSTable Output|EAI:SSTable Output] | Big Data | SSTableOutput | writes to a filesystem directory as a Cassandra SSTable | opdts.cassandrasstableoutput.SSTableOutputMeta |
| [EAI:Streaming XML Input]\\ | Deprecated \\ | | | |
| [EAI:String operations] | Transform | StringOperations | Apply certain operations like trimming, padding and others to string value. | opdts.stringoperations.StringOperationsMeta |
| [EAI:Strings cut] | Transform | StringCut | Strings cut (substring). | opdts.stringcut.StringCutMeta |
| [Switch / Case|EAI:Switch-Case] | Flow | SwitchCase | Switch a row to a certain target step based on the case value in a field. | opdts.switchcase.SwitchCaseMeta |
| {color:#c0c0c0}Symmetric Cryptography{color} | {color:#c0c0c0}Experimental{color} | {color:#c0c0c0}SymmetricCryptoTrans{color} | {color:#c0c0c0}Encrypt or decrypt a string using symmetric encryption. Available algorithms are DES, AEC, TripleDES.{color} | {color:#c0c0c0}opdts.symmetriccrypto.symmetriccryptotrans.SymmetricCryptoTransMeta{color} |
| [EAI:Synchronize after merge] | Output | SynchronizeAfterMerge | This step perform insert/update/delete in one go based on the value of a field. | opdts.synchronizeaftermerge.SynchronizeAfterMergeMeta |
| [EAI:Table Agile Mart]\\ | Agile \\ | | | |
| [EAI:Table Compare]\\ | Utility \\ | TableCompare | This step compares the data from two tables (provided they have the same lay-out). It'll find differences between the data in the two tables and log it. | opdts.tablecompare.TableCompareMeta |
| [Table exists|EAI:Table Exists] | Lookup | TableExists | Check if a table exists on a specified connection | opdts.tableexists.TableExistsMeta |
| [Table input|EAI:Table Input] | Input | TableInput | Read information from a database table. | opdts.tableinput.TableInputMeta |
| [Table output |EAI:Table Output]\\ | Output | TableOutput | Write information to a database table | opdts.tableoutput.TableOutputMeta |
| [EAI:Teradata Fastload Bulk Loader] | Bulk loading | TeraFast | The Teradata Fastload Bulk loader | opdts.terafast.TeraFastMeta |
| [Teradata TPT Insert Upsert Bulk Loader] | Bulk loading | TeraDataBulkLoader | Bulk loading via TPT using the tbuild command. | |
| [Text file input|EAI:Text File Input] | Input | TextFileInput | Read data from a text file in several formats. This data can then be passed on to the next step(s)... | opdts.textfileinput.TextFileInputMeta |
| [Text file output|EAI:Text File Output] | Output | TextFileOutput | Write rows to a text file. | opdts.textfileoutput.TextFileOutputMeta |
| [EAI:Transformation Executor]\\ | Flow \\ | | | |
| [EAI:Transformation Executor]\\ | Flow \\ | | | |
| [Unique rows|EAI:Unique Rows] | Transform | Unique | Remove double rows and leave only unique occurrences. This works only on a sorted input. If the input is not sorted, only double consecutive rows are handled correctly. | opdts.uniquerows.UniqueRowsMeta |
| [Unique rows (HashSet) |EAI:Unique Rows (HashSet)] | Transform | UniqueRowsByHashSet | Remove double rows and leave only unique occurrences by using a HashSet. | opdts.uniquerowsbyhashset.UniqueRowsByHashSetMeta |
| [Univariate Statistics|DATAMINING:Using the Univariate Statistics Plugin] | Statistics | UnivariateStats | This step computes some simple stats based on a single input field | opdts.univariatestats.UnivariateStatsMeta |
| [EAI:Update] | Output | Update | Update data in a database table based upon keys | opdts.update.UpdateMeta |
| [EAI:User Defined Java Expression] | Scripting | Janino | Calculate the result of a Java Expression using Janino | opdts.janino.JaninoMeta |
| [EAI:Value Mapper] | Transform | ValueMapper | Maps values of a certain field from one value to another | opdts.valuemapper.ValueMapperMeta |
| [EAI:Vertica Bulk Loader]\\ | Bulk loading | VerticaBulkLoader | Bulk loads data into a Vertica table using their high performance COPY feature \\ | opdts.verticabulkload.VerticaBulkLoaderMeta |
| [EAI:Web services lookup] | Lookup | WebServiceLookup | Look up information using web services (WSDL) | opdts.webservices.WebServiceMeta |
| [EAI:Knowledge Flow] | Data Mining | KF | Executes a Knowledge Flow data mining process | org.pentaho.di.kf.KFMeta |
| [Write to log |EAI:Write to log (step)]\\ | Utility | WriteToLog | Write data to log | opdts.writetolog.WriteToLogMeta |
| [XBase input|EAI:XBase Input] | Input | XBaseInput | Reads records from an XBase type of database file (DBF) | opdts.xbaseinput.XBaseInputMeta |
| [XML Input Stream (StAX)|EAI:XML Input Stream (StAX)] | Input | XMLInputStream | This step is capable of processing very large and complex XML files very fast. | opdts.xmlinputstream.XMLInputStreamMeta |
| [EAI:XML Input]\\ | Deprecated \\ | | | |
| [EAI:XML Join] | Joins | XMLJoin | Joins a stream of XML-Tags into a target XML string | opdts.xmljoin.XMLJoinMeta |
| [EAI:XML Output] | Output | XMLOutput | Write data to an XML file | opdts.xmloutput.XMLOutputMeta |
| [EAI:XSD Validator] | Validation | XSDValidator | Validate XML source (files or streams) against XML Schema Definition. | opdts.xsdvalidator.XsdValidatorMeta |
| [EAI:XSL Transformation] | Transform | XSLT | Transform XML stream using XSL (eXtensible Stylesheet Language). | opdts.xslt.XsltMeta |
| [Yaml Input |EAI:Yaml Input] | Input | YamlInput | Read YAML source (file or stream) parse them and convert them to rows and writes these to one or more output. | opdts.yamlinput.YamlInputMeta |
| [Zip File|EAI:Zip file (step)]\\ | Utility \\ | ZipFile | Creates a standard ZIP archive from the data stream fields \\ | opdts.zipfile.ZipFileMeta |

This documentation is maintained by the Pentaho community, and members are encouraged to create new pages in the appropriate spaces, or edit existing pages that need to be corrected or updated.

Please do not leave comments on Wiki pages asking for help. They will be deleted. Use the forums instead.

Adaptavist Theme Builder (4.2.0) Powered by Atlassian Confluence 3.3.3, the Enterprise Wiki