Cannot access Query Editor in AWS Redshift - sql

I'm doing an online Data Engineering course and have been given access to an AWS account for practice that's assigned to me.
When I create a Redshift cluster (done using Boto3 library) I wanted to check if my python code successfully inserted data into tables I created programmatically.
I wanted to use the Query Editor (tried both version 1 and 2) to do this online rather than Python. These errors occur when I click on "Connect and Run".
Query Editor v1 error:
I don't know how to resolve this, I'm unable to run queries in the Query Editor.
Query Editor v2 error:
It's worth mentioning that my role only has the following permission policy when I check in IAM:
AmazonS3ReadOnlyAccess
Thanks for your help!!
UPDATE:
I followed instructions from an AWS Redshift page and just added the following permissions to that role:
AmazonRedshiftReadOnlyAccess
AmazonRedshiftQueryEditor
AmazonRedshiftFullAccess
AmazonS3ReadOnlyAccess
However, I'm still getting the same Query Editor errors.

Related

I am trying to create a table in sandbox and continue to receive this error: Unexpected error Tracking number: c7749408936598037

I am new to bigquery doing the data analytics certificate through google. I am trying to use data from that course to create a table and continue to receive this error. I have tried with multiple data sets. I am uploading as csv files, auto detect schema
Unexpected error
Tracking number: c7749408936598037
if you are using public dataset you can not create table.
You need your own GCP account , try for free trial account if not done.
I'm working on the Google Analytics Certificate as well(using Sandbox for BigQuery), I found that the error pops up and if I close out of the table creation window and refresh the page, the table is included under the dataset.
I'm not sure what is causing this, but I was able to load the csv files that I have needed so far. Hope this works for you as well!
As was previously stated above, simply close the "Create Table" window and confirm that's what you want to do, refresh the page in your browser and look under the dataset to find the table you created will actually be there.
Weird bug.
I think you should try to drop the file .csv in google cloud storage.
Create Table --> Create table from Google Cloud Storage --> Select your source path
Try to browse project name.
If still got error than upload it in google spreadsheet and than browse project name.
I did the same thing and its working.

Why am I getting this error from azure sql? Error Code [70527]

I am encountering this error in azure SQL [70527] any suggestions on how to fix this? This is the first time I have encountered this error.
I would appreciate any advice at all, I'm at a loss how to fix this issue, I am not well versed with azure SQL at all
I am assuming you are migrating a database from on premise sql serve to Azure. I suggest you use the migration agent which will highlight all these issues with possible resolution in some cases.
The error code 70527 is related to user login which is highlighted in the following post as well
SQL70527 error in database project
You may have to create the user as per suggestion in the post above using the script method rather than the backup restore.
You can download the migration agent from below link
https://www.microsoft.com/en-us/download/details.aspx?id=53595

Schema published but not seen in BI Server with jpivot

I created an schema in Schema Workbench and publish with no errors, but when I got in BI Server with the standard user admin, I choose New->jpivot, then it display the name of the schema I created but it does not display the cube. For reference the error I get from catalina.out is:
17:11:45,174 ERROR [PentahoDataSourceResolver] PentahoXmlaServlet.ERROR_0002 - IDatasourceService.UNABLE_TO_INSTANTIATE_OBJECT
org.pentaho.platform.api.data.DBDatasourceServiceException: javax.naming.NameNotFoundException: Name [Esquema Salario] is not bound in this Context. Unable to find [Esquema Salario].
Name [Esquema Salario] is not bound in this Context errors usually appear if you use JNDI name, which is not defined on your system. So, I assumed, that this is the name of the datasource, which you reference while publishing Mondrian schema files to BI server.
Xml file with Mondrian schema definition generated by Schema Workbench does not contain any information regarding how to connect to database. So, you need to specify these details when you upload your schema file on BI server (It's done on step 4 below).
But first you have to create the connection itself (steps 1-2):
Create new JDBC datasource:
Define connection parameters:
If cubes still don't appear after these steps, you may republish your cube: follow same steps as in step 1, but select "Analysis" instead of "JDBC" in the end.
Upload the xml file, generated by schema workbench and select the datasource, which you have created on step 2. .
If cube still does not appear - check your log again. If you see the same Name is not bound error, you may try to restart your BI server application (new connections usually get recongnized immediately, but if you had a connection with same name before, than you might need to restart tomcat).
If that does not work, than once again, check log files. I guess, you'll have some different error in this case.
I had the same problem as the OP (blank screen when clicking New View) with the latest version of Pentaho BI server 7.1 (latest at the moment) and even with the 6.0 version one, Pivot4J SNAPSHOT 1.0 plug in version (latest as of today), Schema Workbech 3.14 (latest as of today).
And as, in line with OP, my catalina.out log was also spitting the Name [DatasourceName] is not bound in this Context. Unable to find [DatasourceName].
After several trials and errors I noticed the problem showed up when I checked the "Register the XMLA Data Source" when publishing the schema on Schema Workbench. So to fix the problem I just unchecked it before publishing.
Another way to fix this is going to the Manage Datasources option on the BI server, Import Analysis, choosing the schema created by Schema Workbench, AND manually setting the data source parameter value EnableXmla to false and saving changes. Now the schema should show up when clicking on Create New > Pivot4J view.

Unable to deploy database to Azure

I created ms sql database in SSMS 2012. Connected successfully to Azure and trying to deploy db to the cloud.
Encountering following errors:
Please see screen shot
Numerous Usupported property errors — not supported when used as part of a data package
You're likely using a feature not supported in Azure SQL Database. Please refer to this non supported features list to help you pinpoint the problem:
http://msdn.microsoft.com/en-us/library/azure/ff394115.aspx
This happened with me too. In my case ,i changed the schema of a table after creating once for the first time. After deleting that table database deployed correctly. Usually this error occurs when validating schema fails.
Regards
MAnoj Bojja

Sql 2008 Developer to Sql Azure Migration

Hi My company is deciding for switching its existing application to azure platform (only Sql Part). So we need to upload our db from local to cloud. For migration i came across various tools like
1. cerebrata 's tools
2. SqlAzure Migration wizard
3. Microsoft Sql Data Sync
4. Conventional Script way via management studio.
But all the above tools showed that they have limited capacity. A user cannot work flawlessly on either of the tool.
In cerebrata's tool - the main drawback was its field for Application User Name and Application Key , which my admin havent shared. Also there is manual mapping of fields between azure and local.
Sql Azure Migration wizard - generates scripts and executed too but with lots of error . I was using its version 2.1. Also it very slow. It seems that its a replica of Sql Srvr Mgmt Studio.
Sql Data Sync :- I found it cool as its a MS product but it has limitation too that it only connects with Windows Authentication based local sql server, or you need to explicitly allow the required but. Even after allowing while syncing , I got some Sql Azure Provisioning Error.
4 Sql Srvr Mgmt Studio :- This is most easiest way but requires a lot of manual work to do before actual migration. What i did is that I generated a script of entire db (almost 101123 lines of code for single db) and tried to execute on azure. On the very first time i faced some keyword mismatch error . Finally i removed all line after primary key declaration that With (Padding = Off ....)or something similar and also On Primary then i executed , but still got error on Set Identity Insert On. After doing a lot of hard work in removing unwanted lines waited more than 2 hrs to completed the script remotely, i got no Errors , errors and errors.
So you guys are requested to please suggest me any good alternative stated than above or i am lacking something and can do more with above.
Thanks
Amit Ranjan
I've faced a similar problem recently, running through the options you've listed.
You might give a try to Red-Gate beta for Azure (free for a few months). I found their tools to be quite good for SQL schema and data replication.
Never tried the Azure build myself, though (I migrated tables manually by the time I was told about the offer).