Today I would like to talk about the “User Defined Java Class” a.k.a. UDJC step introduced in Kettle 4.0. This step is incredibly versatile. It allows you to put arbitrary processing code into the ETL without the penalty of a performance hit. This article shows how to use the step in different scenarios, explaining each of . . . → Read More: The User Defined Java Class Step
Sometimes an ETL process needs to generate files in very specific non row based formats. This can be standard files like EDIFACT record files or maybe files using some ancient format you need to feed to a legacy system. In this post I would like to show some techniques to create those files using Pentaho Kettle . . . → Read More: Writing custom output formats in Pentaho Kettle
This article shows how to develop a simple plugin which provides a custom transformation step for Kettle 4.0. The transformation step should accept any row stream and append a string field at the end, filling it with a fixed value. The user should be able to define the name of the added field. For starters, that should be enough. Keeping the step functionality at a minimum allows me to explain how the plugin interfaces with Kettle with as little distraction as possible. . . . → Read More: Developing a custom Kettle Plugin: A Simple Transformation Step
Sometimes it would be nice to access a Java library directly from Kettle. You might find it useful for validation, lookup or custom cryptography support, just to give a few examples. Sometimes even basic access to data is not as straightforward as getting a file dump or a using a database connection. Some data sources might be encapsulated in an application, and the only way to get your hands on it, is using a custom Java client. This article explains how you can directly utilize your Java classes in Kettle a.k.a. PDI.
NOTE: If you’re using Kettle 4.0 or later, you also have the option to use the new User Defined Java Class step.
Continue reading Using Java in Pentaho Kettle