Writing to local file system in Pentaho Server 8 CE - pentaho

I have used in development a Text File Output step in Pentaho 8 CE.
While using the client local installation, output files are written normally to file system.
When I installed the Pentaho Server 8 CE and configured everything to use MySQL as repository, I noticed that system was not writing files to file system.
I suspect that this is because Jackrabbit has been configured to use MySQL as repository following official documentation (https://help.pentaho.com/Documentation/8.2/Setup/Installation/Archive/MySQL_Repository)
Is it possible to configure Jackrabbit so all files use the filesystem?
If so, where in the documentation is this process documented?
Is there any alternative step which forces using local file system?

Quick answers:
No.
See above.
The files can be written, but most likely they’re inside your pentaho-solutions folder inside the server installation. You must use absolute paths when running from the Pentaho repository.

Related

Need help to select correct install file and options for BA\PDI

We want to upgrade our Pentaho suite from 4.1 to 5.0 version. Currently we use Analyzer for ad hoc reporting (which is installed on linux box) and Kettle (which is on windows).Plus this time we want to install repository also so that our kettle jobs are shared centrally.
With this requirement I believe I need to install both BA and DI components, since Analyzer comes under BA. But I can see that BA and DI have separate repository.
Can someone help me explaining which install files I should run and what options should I pick ?. I have 3 files pdi-ee-5.0.2-dist.zip, pdi-ee-client-5.0.1-dist.zip and pdi-ee-server-5.0.2-dist.zip
I would also appreciate if someone can provide me with correct link or video to help me with the install\configuration.Right now I am referring to http://infocenter.pentaho.com/help/index.jsp
You don't have to install anything, you just have to unzip the files and have to run appropriate batch file on windows and .sh file in linux.
For pentaho Data integration you have to run "spoon.bat" on windows and "spoon.sh" on linux
For pentaho BA Server you have to run you have to reach down to biserver-ce foler (i am using community addition) for you it may be like biserver-ee and under that you have to run "start-pentaho.bat" in windows and "start-pentaho.sh" in linux.
you can refer this link..

Using Jenkins as a service on Cloudbees

I am new to Cloudbees and have been trying to find out how I can run an existing Jboss Portal Server based application which we run in our locally hosted CI in Cloudbees infrastructure.
Our stack has the following components
JDK 1.6 JBoss
Portal Server (EPP 4.3)
Oracle Express Edition (XE)
Would appreciate any help from the community to ensure that I dont discard the option of running Jenkins in the cloud on the Cloudbees platform without proper research.
You will have to setup your build job to install and start the adequate runtime
JDK 6 is available as part of CloudBees runtimes, you can then use /private repository to store EPP 4.3 as a zip and expand to /tmp during a pre-build step
Same principle applies to your database, but I'm not sure you can install Oracle XE without user interaction and without being root. I remember doing this myself some years ago on ubuntu and was not as trivial as just unzipping a binary distro.
Is your code tied to this DB ? Or are you using some DB abstraction layer that you could use to test using another DB runtime (mysql / postgres) ?

Where to find Apache Hive Embedded

I believe Apache Hive can be embedded in Java programs. Can somebody please direct me to the page where "Embedded Hive" can be downloaded? I need to embed Hive to be able to run it on Windows, which is where I am developing my application. Further instructions for embedding and code samples will also be useful.
Hive supports embedded mode only in the sense that the RDBMS which stores the meta information for the Hive tables can run locally or on a stand alone server (ref https://cwiki.apache.org/confluence/display/Hive/HiveClient ). Furthermore, hive with it's accompanying database is merely an orchestrator for a string of MapReduce jobs, which requires the Hadoop framework to be running as well.
Use the following class org.apache.hadoop.hive.service.HiveServer.HiveServerHandler make sure that hive/conf is on the classpath. And all the hive jars from hive/lib. This embedded client needs to be run from the same machine where your hive is installed. If your hive-site.xml is using derby than the embedded client will create a .metastore folder, if your hive-site.xml is using standalone db, then the embedded client will communicate with the db directly so make sure it is running.

How to set target MDS in EAR file with Oracle ADF

I want to enable user customization at run time with Oracle ADF. With JDeveloper, if I deploy my application directly to the server, a window permits me to choose a repository previously registered on my Enterprise Manager, as depicted in following image:
In this way my application works great. Now I would to deploy to EAR file (in order to save a copy of my release) but I don't know how to target MDS repository. On the web I've found that maybe the adf-config.xml file has to be modify, but in which way?
have a look at this blog entry, which explains how to do it. Apparently this information needs to be provided upon EAR file deployment
http://andrejusb.blogspot.de/2011/05/target-mds-repository-for-adf.html

Prestashop 1.3.6 to 1.4.1

I have an eshop running PS 1.3.6 version. On my local I've updated to 1.4 first and then to 1.4.1...
Now I would like to update on server... is it possible to just upload files from my local 1.4.1, adjust the settings file and run the update script from 1.4.1 directly (without the middle step to 1.4)?
I can see there are database update scripts for each version, so it should be safe to do it like that, but I want to be sure before I run it on server.... thanks
I am usually doing the major upgrades this way:
Take a snapshot of the current site (tar.gz) & backup database using mysqldump tool (for compatibility);
Download all the files and setup the site on your local server machine using database dump (via mysql command) and downloaded snapshot. Adjust settings if necessary.
Perform an upgrade on your local site, thoroughly test it and test it again with your code & theme.
Repackage your updated files and database (tar.gz & mysqldump) and upload them to the server.
Erase the old site and untar upgraded site to its folder to take its place.
Replace old database with an upgraded one (using mysql command on the server).
Adjust settings if necessary. Test and run it! :)
That should be all. If you're more advanced you could optimize most of the steps. Give me a shout if you need all the useful commands to back up and restore files & DB.