Is there a Hive 2 standalone metastore solution? - hive

Hive 3 has a standalone metastore which seems to work great... although in order to run the schematool I still had to download (but not run) Hadoop.
Unfortunately Presto only works with Hive 1.x and Hive 2.x, as Hive 3 has default ACID2 tables which Presto does not work with (https://github.com/prestosql/presto/issues/576).
So, I'm trying to understand if I can run the Hive 2.x metastore without the rest of hive (hive server) or Hadoop running. Like... If I install Hadoop and hive but only run the metastore will it be functional, or are there limitations?

Related

NiFi connection to Hive fails

I am trying to insert into Hive version 2.3.2 using NiFi 1.9.2 on Docker. It works well with PutHiveQL processor but always fails with PutHiveStreaming.
The Hive components included with Apache NiFi are not compatible with
Hive 2.x, they are built with Hive 1.2.x. There is a Jira to add Hive
2 support (NIFI-6456) but it is not yet in NiFi.

Running impala service alone in docker

I am trying to install impala in a docker container(using MAPR documentstion).In this docker I am running only Impala service and remaining hive,maprfs services will be running on physical node.When starting impala-server(impala daemon) I am getting wearied errors.I just wanted to know whether this kind of installation is possible or not.
Thanks for Help!!
It is possible, but it depends on your Impala and MapR version. Impala 2.2.0 is supported on MapR 5.x. Impala 2.5.0 is supported on MapR 5.1 and later. Check enter link description here before proceeding.

Apache Hive: How to obtain the runtime metrics of the hive queries executed from the JDBC as well as CLI?

I am trying to get the runtime metrics of hive after executing the hive queries. Are there any API's to obtain these metrics. Please suggest.
The whole intention behind asking this question is to gather metrics for different mapreduce jobs spawned at each stage and the amount of memory and cpu being used for each stage.
Hadoop Distribution: MapR (5.1)
Hive Version: 1.2.0 (Hive Server 2)

Upgrading Hive to allow Update/Delete Transactions within Ambari

I have created a Hadoop Cluster with Ambari 2.1 including Hive. I would like to be able to do Update and Delete queries within Hive, but it looks like I currently have version 0.12.0.2.0 of Hive. I would like to upgrade to 0.13 or 0.14 to enable these transactions, but I am not sure how to do that with an existing installation of Ambari. Any help would be appreciated.
I think you could follow the HDP docs from hortonworks website
Manual Upgrade of HDP
Upgrading Stack - Ambari
Performing upgrade - Hortonworks
Hope this is helpful.
P.S: Upgrades/ Inserts are not supported in 0.13. You will have to have 0.14 or later for the same.

Can Presto connect to other Hadoop distributions and run queries

I see Presto has plugin only to CDH4. Can I connect to other distributions such as HortonWorks from this and what does it take to do it.
Without a specific plugin, I am running into "path host null" errors when executing queries from Presto. Appreciate your help.
The Presto Hive connector supports multiple versions of Hadoop:
hive-hadoop1: Apache Hadoop 1.x
hive-hadoop2: Apache Hadoop 2.x
hive-cdh4: Cloudera CDH 4
hive-cdh5: Cloudera CDH 5
See the Hive Connector documentation for more details.
Where is the code for the CDH connector in GitHub?
briefly looking at the code in GitHub, i dont see anything specific to CDH , other than the name, in presto / presto-hive-cdh4 /src / main /java - am i looking at the wrong thing?