Upgrading Hive to allow Update/Delete Transactions within Ambari - hive

I have created a Hadoop Cluster with Ambari 2.1 including Hive. I would like to be able to do Update and Delete queries within Hive, but it looks like I currently have version 0.12.0.2.0 of Hive. I would like to upgrade to 0.13 or 0.14 to enable these transactions, but I am not sure how to do that with an existing installation of Ambari. Any help would be appreciated.

I think you could follow the HDP docs from hortonworks website
Manual Upgrade of HDP
Upgrading Stack - Ambari
Performing upgrade - Hortonworks
Hope this is helpful.
P.S: Upgrades/ Inserts are not supported in 0.13. You will have to have 0.14 or later for the same.

Related

Running impala service alone in docker

I am trying to install impala in a docker container(using MAPR documentstion).In this docker I am running only Impala service and remaining hive,maprfs services will be running on physical node.When starting impala-server(impala daemon) I am getting wearied errors.I just wanted to know whether this kind of installation is possible or not.
Thanks for Help!!
It is possible, but it depends on your Impala and MapR version. Impala 2.2.0 is supported on MapR 5.x. Impala 2.5.0 is supported on MapR 5.1 and later. Check enter link description here before proceeding.

What version of Hive is packed in CDH 5.5.x?

I would like to know the version of Hive that comes along with the CDH 5.5.x version?
I have managed to get the answer. CDH 5.5.x comes with Hive 1.1.0. For more details of various other tools' version check out this link.

Error while executing select query in Hive - how to update Hadoop version

I am currently using hadoop 1.0.3 version. I recently installed Apache Hive to run with it. I was running the select * query which gave me an NoSuchMethodError: org.apache.hadoop.mapred.JobConf.unset
I further found out its a compatibility issue with my current version of hadoop and requires me to upgrade to 1.2 or later.
I am fairly new to hadoop and would like to upgrade my current version to 1.2 or later. How do I go about doing the same.
I could not find any resources online to do so.
Thanks.
Just download hadoop 1.2.x from here and do necessary configuration changes in your new hadoop. Change HADOOP_HOME to point to your new hadoop folder.
NOTE: Change all the environmental variables (including .bashrc) to point to your new hadoop.

Ambari setup on top of Hadoop

I have the following already setup (in order)
A Hadoop 2.5.x - 3 nodes cluster.
Apache Ambari Version 1.7.0
So my question is, Is there anyway I can integrate/configure Amabri to monitor my already setup Hadoop cluster ?
Any suggestions/views are much appreciate.
Currently it seems like that's not possible, there is open JIRA ticket on this issue:
https://issues.apache.org/jira/browse/AMBARI-2852

Can Presto connect to other Hadoop distributions and run queries

I see Presto has plugin only to CDH4. Can I connect to other distributions such as HortonWorks from this and what does it take to do it.
Without a specific plugin, I am running into "path host null" errors when executing queries from Presto. Appreciate your help.
The Presto Hive connector supports multiple versions of Hadoop:
hive-hadoop1: Apache Hadoop 1.x
hive-hadoop2: Apache Hadoop 2.x
hive-cdh4: Cloudera CDH 4
hive-cdh5: Cloudera CDH 5
See the Hive Connector documentation for more details.
Where is the code for the CDH connector in GitHub?
briefly looking at the code in GitHub, i dont see anything specific to CDH , other than the name, in presto / presto-hive-cdh4 /src / main /java - am i looking at the wrong thing?