Install Azure U-SQL Extensions to run R/Python scripts locally? - azure-data-lake

We can extend U-SQL scripts with R/Python code in Azure Data Lake Analytics, but how can we do it locally?

Install U-SQL Advanced Analytics extensions in your Data Lake Analytics Account
1.1 Launch your Azure Portal
1.2 Navigate to your Data Lake Analytics Account
1.3 Click Sample Scripts
1.4 Click More and select Install U-SQL Extensions
1.5 Wait until the extensions have finished installing (2GB)
1.6 Have you waited? Then go to your Data Lake Analytics Account
1.7 Navigate to your default Data Lake Store Account
1.8 Click on Data Explorer and verify that a folder /usqlext exists
Get your USQL Local Run path
2.1 Launch your Visual Studio
2.2 Select Tools > Options > Azure Data Lake > General
2.3 Under U-SQL Local Run, find and copy the value for DataRoot
2.4 The value will look like this: C:\Users\username\AppData\Local\USQLDataRoot
Copy U-SQL Advanced Analytics extensions from Azure to your localhost
3.1 Use a powershell script or ... go to the next line
3.2 Launch Microsoft Azure Storage Explorer (great tool, install it)
3.3 Locate your default Data Lake Store, the one of your Data Lake Analytics Account
3.4 Open data explorer and Download the folder /usqlext to your USQL Local Run's path
3.5 The full path should look like this: C:\Users\username\AppData\Local\USQLDataRoot\usqlext
Final step, register all Azure U-SQL Extensions under U-SQL Local Run
4.1 Launch your Visual Studio
4.2 Start a new U-SQL project
4.3 Open the file C:\Users\username\AppData\Local\USQLDataRoot\usqlext\RegisterAll.usql
4.4 Copy the text into your own U-SQL script
4.5 Run it in Local Execution mode
4.6 Once the script finishes...
You will be able to use all the U-SQL Advanced Analytics features (Python, R, Cognitive) on your own machine and explore all the samples in \usqlext\samples!
Have a nive day!

This answer does not apply directly to the OP, but this article is the closest match to the issue that I had. Knowing others may find this article while searching for the solution, I am recording my answer to this thread.
Problem: In the Azure Portal (not locally), if you choose "Install Azure U-SQL Extensions", the job eventually fails with a non-descript error.
What is happening behind the scenes is that all of the files are copying into storage, but the assemblies fail to register. We have to create a job manually to register the assemblies.
Answer:
Open Data Explorer
Navigate to /usqlext
Download the file "RegisterAll.usql". Open in Notepad, and copy the text
Create a new Job. Paste in the text.
Execute the job.
Assemblies will be registered (verify by checking Catalog->master->Assemblies). You can now run the Cognition and Python samples.

Related

How do I update a Google Sheets Add-On

I have made changes to my (already published) Google Sheets Add-On (in the Script Editor).
I have created a new version and pressed "Publish > Deploy as Add-On" which says
You must configure the G Suite Marketplace SDK to publish an add-on.
When I click on the link G Suite Marketplace opens (and I see the currently published products catalog entry) but no way of specifying a new version.
.
When I click OK nothing happens.
The entire Google Sheets add-on deployment workflow is one hot mess. Do not use Publish -> Deploy as Sheets add-on... to update an already-deployed AppScript. It does not work.
Instead, first create a version of the script File -> Manage version. Then go to the G Suite Marketplace SDK page in the Google Cloud Platform console (console.cloud.google.com). Under the Configuration tab, scroll down to find Editor Add-on extensions where Sheets add-on extension should already be checked (if you followed the publication directions earlier and have already successfully published the add-on).
There you will see a field for the script version number. Increment it to the latest version, and Save changes. As far as I can tell this takes effect immediately.
Answer:
Google is currently migrating add-ons from the Chrome Web Store to the G Suite Marketplace, so now need to be published there rather than the legacy marketplace.
Further Information:
In order to publish an Apps Script project as an addon, as specified in your message, the G Suite Marketplace SDK must be configured. You can find an article detailing the migration process here.
You will need to enable the G Suite Marketplace SDK in your Google Cloud Platform (GCP) project. You can follow the information on the Getting Started page for Google Cloud.
References:
All add-ons now in the G Suite Marketplace
Publish to the G Suite Marketplace
Getting Started with Google Cloud
The new GSM publishing flow for Editor Add-ons as described https://developers.google.com/gsuite/add-ons/how-tos/editor-publish-overview is now available.
This means that all updates now needs to be done through GSM. Please make sure to complete all the fields in both the configuration and publish section of the marketplace SDK and click on the 'Publish' button.

Writing to local file system in Pentaho Server 8 CE

I have used in development a Text File Output step in Pentaho 8 CE.
While using the client local installation, output files are written normally to file system.
When I installed the Pentaho Server 8 CE and configured everything to use MySQL as repository, I noticed that system was not writing files to file system.
I suspect that this is because Jackrabbit has been configured to use MySQL as repository following official documentation (https://help.pentaho.com/Documentation/8.2/Setup/Installation/Archive/MySQL_Repository)
Is it possible to configure Jackrabbit so all files use the filesystem?
If so, where in the documentation is this process documented?
Is there any alternative step which forces using local file system?
Quick answers:
No.
See above.
The files can be written, but most likely they’re inside your pentaho-solutions folder inside the server installation. You must use absolute paths when running from the Pentaho repository.

Need help to select correct install file and options for BA\PDI

We want to upgrade our Pentaho suite from 4.1 to 5.0 version. Currently we use Analyzer for ad hoc reporting (which is installed on linux box) and Kettle (which is on windows).Plus this time we want to install repository also so that our kettle jobs are shared centrally.
With this requirement I believe I need to install both BA and DI components, since Analyzer comes under BA. But I can see that BA and DI have separate repository.
Can someone help me explaining which install files I should run and what options should I pick ?. I have 3 files pdi-ee-5.0.2-dist.zip, pdi-ee-client-5.0.1-dist.zip and pdi-ee-server-5.0.2-dist.zip
I would also appreciate if someone can provide me with correct link or video to help me with the install\configuration.Right now I am referring to http://infocenter.pentaho.com/help/index.jsp
You don't have to install anything, you just have to unzip the files and have to run appropriate batch file on windows and .sh file in linux.
For pentaho Data integration you have to run "spoon.bat" on windows and "spoon.sh" on linux
For pentaho BA Server you have to run you have to reach down to biserver-ce foler (i am using community addition) for you it may be like biserver-ee and under that you have to run "start-pentaho.bat" in windows and "start-pentaho.sh" in linux.
you can refer this link..

Using Jenkins as a service on Cloudbees

I am new to Cloudbees and have been trying to find out how I can run an existing Jboss Portal Server based application which we run in our locally hosted CI in Cloudbees infrastructure.
Our stack has the following components
JDK 1.6 JBoss
Portal Server (EPP 4.3)
Oracle Express Edition (XE)
Would appreciate any help from the community to ensure that I dont discard the option of running Jenkins in the cloud on the Cloudbees platform without proper research.
You will have to setup your build job to install and start the adequate runtime
JDK 6 is available as part of CloudBees runtimes, you can then use /private repository to store EPP 4.3 as a zip and expand to /tmp during a pre-build step
Same principle applies to your database, but I'm not sure you can install Oracle XE without user interaction and without being root. I remember doing this myself some years ago on ubuntu and was not as trivial as just unzipping a binary distro.
Is your code tied to this DB ? Or are you using some DB abstraction layer that you could use to test using another DB runtime (mysql / postgres) ?

How to set target MDS in EAR file with Oracle ADF

I want to enable user customization at run time with Oracle ADF. With JDeveloper, if I deploy my application directly to the server, a window permits me to choose a repository previously registered on my Enterprise Manager, as depicted in following image:
In this way my application works great. Now I would to deploy to EAR file (in order to save a copy of my release) but I don't know how to target MDS repository. On the web I've found that maybe the adf-config.xml file has to be modify, but in which way?
have a look at this blog entry, which explains how to do it. Apparently this information needs to be provided upon EAR file deployment
http://andrejusb.blogspot.de/2011/05/target-mds-repository-for-adf.html