Run ktr remotly using on pentaho BI server - pentaho

I'm using pentaho BI server 7 EE.
I have uploaded a simple transformation (.ktr) and run it successfully throught PUC.
But when I try to call it throught REST service (http://localhost:8080/pentaho/kettle/runTrans?trans=%3Apublic%3Ademo%3AsimplePdiJob.ktr) it fails with the following response:
<webresult>
<result>ERROR</result>
<message>!RunTransServlet.Error.UnexpectedError!</message>
<id/>
</webresult>
I didn't find any exception in the log.
What can be the cause of this problem?
Thanks

Hi I do not know if you still have this problem but I managed to solve it these days doing an xaction and inserting in the in the XML code that I inserted a PHP script that made the call of my .ktr file and uploaded to the server a file Csv that was with which I would perform my ETL. I know that is not the best solution but for my case there is no way even in version 7 of bi server to upload csv files or some other extension.

Related

simple-jdin/jdbc.properties is ignored in pentaho-server 8.2

If we need to perform queries in pentaho data integration (IDE), we need to add manually the datasource in simple-jdin/jdbc.properties
MyDatabase/type=javax.sql.DataSource
MyDatabase/driver=org.hsqldb.jdbcDriver
MyDatabase/url=jdbc:hsqldb:hsql://localhost/sampledata
MyDatabase/user=pentaho_admin
MyDatabase/password=password
This works as expected in the ide known as pentaho data integration, spoon or kettle.
But, the same in pentaho server 8.2 does not works.
Steps to reproduce the error
deploy or upload the transformation(.ktr) in the pentaho-server 8.2
add manually the datasource in the server /../pentaho-server/pentaho-solutions/system/simple-jndi/jdbc.properties
executed the transformation using pentaho server web console : run in background or schedule options
error: datasource not found
Alternative
Create the datasource manually using web console of pentaho server instead of manually modification of file : /../pentaho-server/pentaho-solutions/system/simple-jndi/jdbc.properties
Question
simple-jndi/jdbc.properties works in server or is just for development purposes (pdi)?
are the settings available in jdbc.properties file on server can you verify the same by referring to the pentaho data integrator > simple-jndi folder.
also are you able to connect to the database from the server may be using a database client to confirm.
NOTE:whenever you work with databases make sure you relevant library files to connect to the respective database.
From my personal experience, I was not able to make the server pick JNDI connection definitions from the simple-jndi/jdbc.properties file when the Pentaho Server was running with Tomcat.
I was only able to use JNDI on the Pentaho Server by defining JNDI Datasources in the Tomcat configuration files.

Pentaho report which contains more than 7 connections are not working in pentaho DI server

I am new to pentaho. Recently i created a job which generates some report. I am using JNDI connection in the report(.prpt). If I am using more than 7 connections inside a report, it will fail to generate report but it's working smoothly with reports which has lesser connections.
Note: If i execute using spoon, it will work fine even report has more JNDI connections but my requirement is to execute job using pentaho kettle API like this ->
https://address/pentaho-di/kettle/executeJob/?job=/home/pentaho/Test/main.kjb&level=Rowlevel
Seeking for help from pentaho experts.
It was actually due to issue in data factory.
Steps to do:
Stop your Pentaho DI Server if it is already running.
Navigate to
server\data-integration-server\tomcat\webapps\pentaho\META-INF and
edit the context.xml file.
In your jdbc/mart JNDI connection change the factory class from,
factory="org.apache.commons.dbcp.BasicDataSourceFactory" to,
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
Save and close the file.
Clear tomcat\work and tomcat\temp directories.
Restart your PDI server.
Execute the transformation

Import data from CouchDB to Sql Server

Is there a way to export data from couchDB to Sql Server ?
We tried with SSiS using http connection manager, JSON component source etc. But the problem is that couchDB has some security and we can't retrieve data info. We tried installing their certificates no success.
Any other suggestions?
CounchDB has a number of .net client libraries you can use. I would get that working in a console app and then just write that to a csv on the disk and use SSIS to import it.
If you don't want the additional step you could write a .net script source component that used the .net client to retrieve data but get it working in a console app first so you know it is good before bringing SSIS into the equation.
Getting started with CouchDb and c#:
https://wiki.apache.org/couchdb/Getting_started_with_C%23
ed

Adding Informix JDBC to Jasper Server

i have installed Jasper Server on my local. I need to generate a report which will extract data from Informix DB. Therefore a JDBC for it is needed (duh). I found the JDBC file for informix called ifxjdbc.jar and i copy it to "/apache-tomcat/lib". Restarted the server. Then when i want to create a new data source, the dropdown list doesn't show my JDBC driver that i just added.
I even tried copying the JDBC file to "/apache-tomcat/webapps/jasperserver-pro/lib" then restart the server but still no luck
You should install JDBC driver which is available on IBM pages. It contains more jar files than just only ifxjdbc.jar. In my environment I copied: ifxjdbc.jar, ifxjdbcx.jar, ifxlang.jar, ifxlsupp.jar, ifxsqlj.jar and ifxtools.jar.
I think that you should run simple Java/Jython program that tries to connect to your database. If it will work then you are sure you have complete environment.

My job is running but no data is comming

I have got a ssis package which runs when I manually run from Integration Services. But when I try to run it from a job. Then it runs but no data is seen in the data. There seems to be some permission issue. Can somebody tell me what permissions are required for running a package from a SQL Server Job?
State the error message.
If you are using a flat file connection manager, and that's where the error is occurring, click 'start' then 'computer' then check to make sure you are mapped to that drive. If not, click the tab upper right corner to map to the drive then when you access the file through SSIS you shouldn't have an error.
If the package runs successfully as a job using the SQL Server Agent then you have the permissions set right for the database side.
However make sure if you are accessing any external data such as flat files that the agent is able to access these locations. You may have permissions on your Windows account to access the locations when you run the package in Visual Studio but the agent service running the job requires those permissions too.
If this is not the case can you clarify what your package does and any messages you receive from the catalog reports so I can help further.