Issue with data sources that are created through the Pentaho Admin Console ,Now my MySQL IP has been changes, i did modification in my BI server .
But Pentaho Admin console in not up .!
From logs i found this is having
01:24:40,370 ERROR [Logger] misc-org.pentaho.platform.engine.services.connection.datasource.dbcp.PooledDatasourceSystemListener: PooledDatasourceSystemListener.ERROR_0003 - Unable to pool datasource object: MyLocalDatabase caused by com.mysql.jdbc.exceptions.jdbc4.CommunicationsException: Communications link failure
The last packet sent successfully to the server was 0 milliseconds ago. The driver has not received any packets from the server.
01:24:40,857 WARN [PersistenceEngine] Falling back to built-in config
This MyLocalDatabase database Created at Pentaho Admin Console has to be changed with new IP . Do any one help me to know , which file consists of datasources that are created through the Pentaho Admin Console
My PAC is down, no error in server.log file
Datasource information is kept in the hibernate database.
By default, this is kept in an hypersonic database that is launched when you start the BI server. Check context.xml in webapps/pentaho/META-INF to make sure.
There's a DATASOURCE table in there that stores the data source definitions.
Related
If we need to perform queries in pentaho data integration (IDE), we need to add manually the datasource in simple-jdin/jdbc.properties
MyDatabase/type=javax.sql.DataSource
MyDatabase/driver=org.hsqldb.jdbcDriver
MyDatabase/url=jdbc:hsqldb:hsql://localhost/sampledata
MyDatabase/user=pentaho_admin
MyDatabase/password=password
This works as expected in the ide known as pentaho data integration, spoon or kettle.
But, the same in pentaho server 8.2 does not works.
Steps to reproduce the error
deploy or upload the transformation(.ktr) in the pentaho-server 8.2
add manually the datasource in the server /../pentaho-server/pentaho-solutions/system/simple-jndi/jdbc.properties
executed the transformation using pentaho server web console : run in background or schedule options
error: datasource not found
Alternative
Create the datasource manually using web console of pentaho server instead of manually modification of file : /../pentaho-server/pentaho-solutions/system/simple-jndi/jdbc.properties
Question
simple-jndi/jdbc.properties works in server or is just for development purposes (pdi)?
are the settings available in jdbc.properties file on server can you verify the same by referring to the pentaho data integrator > simple-jndi folder.
also are you able to connect to the database from the server may be using a database client to confirm.
NOTE:whenever you work with databases make sure you relevant library files to connect to the respective database.
From my personal experience, I was not able to make the server pick JNDI connection definitions from the simple-jndi/jdbc.properties file when the Pentaho Server was running with Tomcat.
I was only able to use JNDI on the Pentaho Server by defining JNDI Datasources in the Tomcat configuration files.
I am new to pentaho. Recently i created a job which generates some report. I am using JNDI connection in the report(.prpt). If I am using more than 7 connections inside a report, it will fail to generate report but it's working smoothly with reports which has lesser connections.
Note: If i execute using spoon, it will work fine even report has more JNDI connections but my requirement is to execute job using pentaho kettle API like this ->
https://address/pentaho-di/kettle/executeJob/?job=/home/pentaho/Test/main.kjb&level=Rowlevel
Seeking for help from pentaho experts.
It was actually due to issue in data factory.
Steps to do:
Stop your Pentaho DI Server if it is already running.
Navigate to
server\data-integration-server\tomcat\webapps\pentaho\META-INF and
edit the context.xml file.
In your jdbc/mart JNDI connection change the factory class from,
factory="org.apache.commons.dbcp.BasicDataSourceFactory" to,
factory="org.apache.tomcat.jdbc.pool.DataSourceFactory"
Save and close the file.
Clear tomcat\work and tomcat\temp directories.
Restart your PDI server.
Execute the transformation
I have created a report using SSRS when the report is publish to the SSAS server (Different than DB Server), and I am trying to access the report I got strange error.
An error has occurred during report processing. (rsProcessingAborted)
Cannot create a connection to data source 'DD_BE'. (rsErrorOpeningConnection)
Login failed for user 'NT AUTHORITY\ANONYMOUS
then i go back to check the data source connection.
i set the credential is to use windows authentication (integrated).
here the connection in Data tool is fine
every thing seems fine here then I go to check SSRS Manager to check Data Source.
And I changed the credential and I got different error.
Check the option Use as Windows credentials when connecting to the data source.
I am trying to create an Android Mobile app via the Azure app service and while doing so, I am trying to create a database, but I am getting the error 'there was an error while creating data connection 'SQL data connection'. Deployment to resource group failed'. Please help as soon as possible.
$
This error is due to server limit exceeding the quote associated with your account. Under all resources, check if there's a new SQL server created with a random name- you will need to delete that. Attempt to create again.
(just done- it worked!)
I've faced the exact issue.
First time it failed giving the error - Data Connection Creation Error..!
When Configuring for the first time it didn't work for me.
i.e.,
Creating a Database
Creating a New Server
Entering Server Admin Login
Entering Server Admin Password
I've noted down all the above information.
Tried Second time setting up the Connection String with the above details IT WORKED.
Hope this information helps.
I have Report Builder 3.0 installed on my local PC. I am creating a new report and adding an embedded data source to a SQL Server Analysis Services database.
When I build the connection string of the data source, choosing the server name and the database name, I click 'Test Connection' and receive a message saying "Test connection succeeded". So far, so good.
I close the connection properties with the OK button, and on the Data Source Properties window I click the 'Test Connection' button. This time I get an error saying "The connection either timed out or was lost".
If I ignore the error then I can successfully add a dataset to my report and add data from the dataset in to the report design, but when I try to run the report (on my local PC) I again get an error connecting to the data source.
My best guess is that the connection that succeeded is running under my credentials, whereas the connection that fails is running as some other credential and so needs to use Kerberos delegation to pass my credentials along, but that is only a guess and even if I am correct I am at a loss to understand how to fix the issue - I don't know what other credentials may be being used and I have already set SPN's for the Analysis Services service account.
So it turns out that it was a Kerberos issue as I suspected, and I was also correct that Report Builder was testing the connection using some process running under another authentication context.
It turns out that when setting up Report Builder (and I had forgotten it) that you specify a default SSRS Report Server (see screenshot). It must be that when testing data sources or running reports that it connects to that default Report Server and does the work from there - I was assuming that everything was running locally!
Once I figured that out it was just a case of finding a good guide on how to configure SSRS for Kerberos and everything started working. In my case the only bit that I hadn't already done was to add the <RSWindowsNegotiate/> setting to the AuthenticationTypes in the rsreportserver.config file