I am new to pentaho BI and been with pentaho for quite few days.
But I couldn't fetch the current time from pentaho. May I know if any of you could help me.
And i would also like to know how to fetch the filename of the file that is passed as input
Related
I am using Pentaho (Spoon PDI), I have created input and output transformations. When running the sql in OracleDB I get back all the rows and data Pentaho I can generate a preview of the data set(generally 1000 rows).
Is there a tool where I can verify if all the returned values are the same between the two ? I have tried returning rows in each data set as a start but it seems like an innefficent way to go about this. Any input would be appreciated.
On my Pentaho CE 7.1 I often need to generate large number of reports (*.prpt) with different attributes.
For example, I have a report that shows data for a day, and I need to generate those reports for each day since September 2017.
Is there any way how create a script, that would execute those *.prpt files one by one for each day since September 2017 until now?
I have been checking API on official Pentaho documentation, but it does not seem to be such option there. Perhaps some kind of hack, like sending parameters in URL or so?
Create your *.prpt with the Report Designer and use a parameter to select one day in your data.
Then open PDI, with first step to generate a date starting from 2017-09-10, and give this date to a Pentaho Reporting Example step. Then do what you need with the report output (mail, save them in the Pentaho-solutions,...).
You have a use case very similar in the sample directory which is shipped with the Pentaho Data Integrator. It is named Pentaho Reporting Output Example.ktr.
I have a transformation which has table input (which fetch data from db) and and csv output (which will save table input data into csv file.)
And a Job which runs this transformation on weekly basis.
What I want now that whenever my report will get generated, a new dynamic password would create.
Please help me on this. Iam using pdi.
We have a SQL SSIS package setup which is being run manually when needed. This will grab two CSV from the file system and process them to import data to sql table. This should be run once a month when the CSV file is updated.
Now I would like to automate this process. I want to add a step to the existing job that should check for the modified file dates (of two particular files) and compare it to the Today's date and if true, proceed to the next step otherwise not. I am not sure how to check for the modified date of the files (two files) on the file system and compare to today's date. Any thoughts on how to accomplish this? Thanks!!
You'll have to use a script task to pull the modified date in for comparison, using the following medthod:
File.GetLastWriteTime
Im new to the pentaho suite and its automation functionality. i have files that come in on a daily basis and two columns need to be put in place. I have figured out how to add the columns but now i am stuck on the automation side of things. The filename is constant but it has a datestamp at the end. EG: LEAVER_REPORT_NEW_20110623.csv. The file will always be in the same directory. How do i go about using Pentaho data integration to solve this issue? ive tried get files but that doesnt seem to work.
create a variable in a previous transform which contains 20110623 (easy with a get system info step to get the date, and then a select values step to format to string, then a set variables step)
then change the filename of the textfile input to use:
LEAVER_REPORT_NEW_${variablename}.csv