Today, one of our community members posted a deviously simply XML format on the forum that needed to be parsed. The format looks like this: <RESPONSE> <EXPR>USD</EXPR> <EXCH>GBP</EXCH> <AMOUNT>1</AMOUNT> <NPRICES>1</NPRICES> <CONVERSION> <DATE>Fri, 01 Jun 2001 22:50:00 GMT</DATE> <ASK>1.4181</ASK> <BID>1.4177</BID> </CONVERSION> <EXPR>USD</EXPR> <EXCH>JPY</EXCH> <AMOUNT>1</AMOUNT> <NPRICES>1</NPRICES> <CONVERSION> <DATE>Fri, 01 Jun 2001 22:50:02
Tag Archives: Pentaho Data Integration
Dear Kettle users, Most of you usually use a data integration engine to process data in a batch-oriented way. Pentaho Data Integration (Kettle) is typically deployed to run monthly, nightly, hourly workloads. Sometimes folks run micro-batches of work every minute or so. However, it’s lesser known that our beloved transformation
I took the time out to build a high level overview of all the new big ticket items that are going to be in the upcoming version 4.2 of Kettle (Pentaho Data Integration).
Dear Kettle friends, on occasion we need to support environments where not only a lot of data needs to be processed but also in frequent batches. For example, a new data file with hundreds of thousands of rows arrives in a folder every few seconds. In this setting we want
Dear Kettlers, A couple of years ago I wrote a post about key/value tables and how they can ruin the day of any honest person that wants to create BI solutions. The obvious advice I gave back then was to not use those tables in the first place if you’re