How to configure server side parameters in tableau desktop (with HIVE)? - hive

I am trying to connect tableau desktop to Hadoop Hive connection. I want to change the default hive engine to TEZ . How can i change this parameter through tableau desktop?

One thing which I have figured it out is, you can use :
initial SQL
option to set default engine to TEZ. However problem with this is, it only gets fired when firing first tableau query, for subsequent queries again engine related changes are reverted back.I am in research on this with Tableau, and will post once I hear anything concrete.

Related

Can Apache Ignite update when 3rd party SQL Server database changed something directly?

Can I get some advice on whether it is possible to proceed like the steps below?
SQL Server data is loaded in Ignite Cluster
The data in SQL Server has been changed.
-> Is there any other way to reflect this changed data without reloading the data from SQL Server?
When used as a cache in front of the database, when changes are made directly to the DB without going through the Ignite Cluster, can the already loaded cache data be directly reflected in the Ignite cache?
Is it possible to set only the value to change without loading the data again?
If possible, which part should I set? Please.
I suppose the real question is - how to propagate changes applied to SQL Server first to the Apache Ignite cluster. And the short answer is - you need to do it by yourself, i.e. you need to implement some synchronization logic between the two databases. This should not be a complex task if most of the data updates come through Ignite and SQL Server-first updates are rare.
As for the general approach, you can check for the Change Data Capture (CDC) pattern implementations. There are multiple articles on how you can achieve it using external tools, for sample, CDC Between MySQL and GridGain With Debezium or this video.
It's worth mentioning that Apache Ignite is currently working on its own native implementation of CDC.
Take a look at Ignite's external storage integration, and the read/write through features. See: https://ignite.apache.org/docs/latest/persistence/external-storage
and https://ignite.apache.org/docs/latest/persistence/custom-cache-store
examples here: https://github.com/apache/ignite/tree/master/examples/src/main/java/org/apache/ignite/examples/datagrid/store

Power BI query is not visible in BigQuery query history

More of a curiosity question really. I load data into Power BI report from Google BigQuery (using native Google BigQuery connector in Power BI). All works fine, but for some reason I don't see this query in BigQuery's query history.
Did anyone experience something similar and knows the reason why this happens or how to change that (if at all possible)?
If I do exactly the same thing but using simba ODBC connector, I see this query in BigQuery's query history as expected.
Never seen that before. I am always able to find the query history no matter what 3rd party connection I used. Could you confirm the GCP service-account or auth-account and the GCP project for BQ job query that you used for your native Google BigQuery connector in Power BI?
Please make sure you have the access to the query history of that GCP account in that BQ job project.

SAP Spark Controller not caching data

I've got SparkController 2.0.0 running on a HDP 2.4.3 with Spark 1.6.2
In the configuration I have these parameters configured:
sap.hana.es.enable.cache=true
sap.hana.es.cache.max.capacity=500
sap.hana.hadoop.datastore=Hive
I've got HANA 1.00.122 connected to that Spark Controller, set enable_remote_cache parameter to true in indexserver.ini, and imported one of exposed Hive tables as a virtual table in HANA.
Then I ran select-statements against that virtual table, but every time I see that no cache is created (nothing in the Storage tab of Spark UI), nor it is hit (query runtime doesn't drop, and I see the job going through the same stages every time).
Using the hint "with hint (USE_REMOTE_CACHE)" doesn't help either.
Are there any other settings I forgot to make?
In order to enable remote caching for federated queries to Hive from HANA you must also set the HANA parameter enable_remote_cache = true
For more info see the bottom of this page:
https://help.sap.com/viewer/6437091bdb1145d9be06aeec79f06363/2.0.1.0/en-US/1fcb5331b54e4aae82c0340a8a9231b4.html
Accordingly to SAP, the HANA version for caching to work should be 2.0+.

RestAPI or any other methods for pulling data from Cacti

Is there a way we can get data(like interface utilization data of routers) from Cacti server remotely?
I did come accross some answers which said using rrdtool export options,but i was wondering if there was a way to do it without touching the rrd's?I would be using a linux server to remotely access this data.
There is no RESTAPI for CACTI, however you can pull the details from the database tables

H2 Database: Abort Query

I have a long running select query on an embedded H2 Database and want to allow the user to cancel the query.
How can this be done? I cannot find anything about this.
[UPDATE]
To be more specific I'm running my query using JPA. How can the query be stopped?
H2 supports a query timeout setting. You can set this in the database URL as follows: jdbc:h2:~/db/test;query_timeout=10000. (Maybe this is not the right approach for you, but it might be for others that read this question.)
You can also cancel a query running in another connection (session) using the cancel_session function. But for this you need to enable the multi-threaded mode, which is currently not recommended for production use (it is still experimental in version 1.3.175).
If you have a few queries which might take very long, then don't use JPA for them. Create your own statement so you can cancel it.
Alternatively, get the source code for a JDBC proxy like log4jdbc and add code that allows you to get connections and statements per thread. That would allow you to get the statement after JPA has sent it to the proxy.