any method to export dashboards on pentaho and reused on another pentaho BI server? - pentaho

I have made some dashboards on pentaho BI server. I want to export them and import on another pentaho BI server. Is there any method to realize this? we have the same datasource.

In Pentaho 5 version, you can use User Console to Download the Dashboard files. It will get download in zip format. Once it is done, take these codes to the New BI Server, use the user console and UPLOAD it.
For e.g : If you have a project folder named "PROJECT_XYZ" then open the user console, select the project folder >> on the right hand screen you will find "DOWNLOAD" option. Just click on it and the project folder will get downloaded in .zip format. Take this .zip file and UPLOAD it in the new Server using the user console (of the new server).
Hope it helps :)

For Exporting:
import-export.bat --export --url=http://localhost:8080/pentaho --username=admin --password=admin --charset=UTF-8 --path=/public --file-path=E:/pentaho/shaz.zip --overwrite=true --permission=true -retainOwnership=true
For Importing:
import-export.bat --import --url=http://localhost:8080/pentaho --username=admin --password=admin --charset=UTF-8 --path=/public --file-path=E:/pentaho/anyfile.zip --overwrite=true --permission=true -retainOwnership=true

Related

Can't open the Asset in IMB Watson Studio

I added a .ipynb file in Asset Section of Projects of IBM Watson Studio, but when I try to open the file a pop up displays:
"Preview not available
This file type may not be supported."
I don't know how to fix this and I don't know how to proceed to submit my assignament for Coursera IBM Data Science course.
Please, help
The .ipynb file is for Jupyter notebook. Assuming you are trying to create new notebook to execute some code from that notebook file.
You need to use Add to Project -> Notebook option
Then use "from file" option to upload
For more information, refer to the documentation:- https://dataplatform.cloud.ibm.com/docs?context=cpdaas

How to get pentaho report designer and BI server log files?

We are using Pentaho Report Designer(PRD) Tool and BI server, here in PRD and BI server where and how do we find out log files. We are using Pentaho 7.1.0 version.
When you launch PRD/BI server/PDI for the first time it creates .pentaho and .kettle folder inside users directory.
You will be able to find logs inside
Users\Your username\.pentaho\logs
Users\Your username\.kettle
Check respective folder for more details.

How to restore a database from bak file from azure data studio on Mac

Previously on Mac I use mysql operation studio and I click on database and click restore then browse to my bak file, but now they change to azure data studio and when I repeat the same steps I got this error:
"You must enable preview features in order to use restore"
but I cannot figure out hot to enable that. I have googled and tried few things even open my azure, microsoft account on website but I do not see that option.
Can some one help please !
Go to Azure Data Studios > Settings
Edit settings by clicking on
"new settings editor"
Search for "preview"
Scroll to the bottom and
check "Enable unreleased preview features"
Adding this as an answer as I don't have enough rep to comment. This is in response to the question in the comment for the answer to the OP (If that's not confusing enough!)
This only applies when your sql database is hosted on a Mac/Linux/Docker Container. We don't have any Windows servers in our estate for me to test this on to see if the location of the .bak files is any different.
When you click on the "..." button, it browses to /var/opt/mssql/data on the machine (or docker container) the database is hosted on. This is not an issue if you are backing up & restoring databases on the same host, however, if you're migrating to a new server or just creating a dev/UAT/staging environment, it becomes a problem because you don't have access to var/opt/mssql/data.
This is a bit of a sledgehammer to crack a nut type solution but as I'm working with 2 dev boxes, it doesn't make a lot of difference to me.
To make this easier to understand I'll call the server that hosts the database you have backed up ProdServ & the server you are restoring to DevServ.
On DevServ, at a terminal prompt, navigate to /var/opt and make a note of the current permissions on the mssql directory (mine were drwxrwx---).
$ cd /var/opt
$ ls -la
Google the octal value for your permissions (in my instance, it's 770)
Change the permissions of the data directory to rwxrwxrwx.
$ chmod -R 777 /var/opt/mssql/data
(You will also need to do this on ProdServ if that is also a Unix-based o/s)
Copy the .bak files from ProdServ to DevServ via a method suitable to the environment you're working in.
Windows --> Linux I'd use WinSCP
For Mac to Docker, docker cp <fileToCopy> <container>:<destinationPath> works perfectly fine.
Once the files have been copied over, they will magically appear when you click the "..." button in azure data studio again. Make sure you change the directory permissions back to their original value via the same command. So in my instance, simply
$ chmod -R 770 /var/opt/mssql/data
As an extra note, if you're used to working in MSSMS, the wizard there allows you to create a database from a .bak file, from what I can see, Azure Data Studio does not. You have to create the database first (CREATE DATABASE <databasename>) in a query window, then restore the .bak file to it.
simple add this to your setting file:
"workbench.enablePreviewFeatures": true
You can click on View at the top menu next to Window.
Then select the first option Command Palette
Then in the command textbox type Restore and select Restore
Then the restore window comes up.

Essbase dataload error

I'm on a Windows 10 box using Oracle Essbase Admin Services console 11.1.2.4. When I load data I get an error. My problem is that the dataload.err file will not write to my c:\ drive or any network drive to which I have access. If I manually create the dataload.err file in Windows Explorer and run the load data process, I can watch the dataload.err file disappear. I've tried a different file name and different directories, but it won't save.
Why won't the dataload.err file save? Any help would be greatly appreciated.

How upload file to Pentaho User Console server?

I need :
1) Let the user select a file from his local pc
2) Upload that file to the pentaho server
3) Process the file using a kettle transformation
I tried with a csv data source in Pentaho User Console (PUC) 5.0 but found no way to access it from a .ktr file uploaded to PUC repository. I also try to upload the csv file to a folder and still not able to access it from a .ktr file.
I think this requirement is valid :
Upload a csv data file and .ktr file to PUC folder. The .ktr should be able to read the uploaded csv file when it is executed from PUC
Imagine a simple user, with a csv. Will he be able to upload csv file to linux host using wincsp, filezilla or another ftp tool??
We need to give an easy upload functionality to our user, so after several researching hours (pentaho source code) without one line of Pentaho documentation, I found this test:
https://github.com/pentaho/pentaho-platform/blob/master/extensions/src/test/java/org/pentaho/platform/plugin/services/importer/PlatformImporterTest.java that showed me that a mimetype list should be exist somewhere.
So after search some words in all pentaho folder wiht grep command, I found this file:
/my_apps/pentaho-server-ce-7.1.0.0-12/pentaho-server/pentaho-solutions/system/ImportHandlerMimeTypeDefinitions.xml
With some intuition, I added this xml
<ImportHandler class="org.pentaho.platform.plugin.services.importer.RepositoryFileImportFileHandler">
<MimeTypeDefinitions>
<MimeTypeDefinition mimeType="text/plain" >
<extension>csv</extension>
</MimeTypeDefinition>
</MimeTypeDefinitions>
</ImportHandler>
At the bottom of file:
<tns:ImportHandlerMimeTypeDefinitions xmlns:tns="http://www.pentaho.com/schema/" .....
<ImportHandler ../>
<ImportHandler ../>
<!-- PUT CSV CONFIG HERE -->
</tns:ImportHandlerMimeTypeDefinitions>
Finally, I restarted my pentaho-server-ce-7.1.0.0-12 server and I was able to upload my csv file with this steps :
go to http://localhost:8080/pentaho
click en browse files
select some folder
click in upload (right side)
select csv and ok
Read this csv file from ktr is pending...
I hope this helps