I am currently taking Google's Data Analytics course, I'm enjoying it well and am able to troubleshoot most issues on my own. One issue I keep running into is that I am never able to PASTE any of the queries from the lessons into BigQuery's SQL editor. I'm having no issue copying, but I've never been able to successfully paste a query. I've tried writing them myself; to varying degrees of success. However, at this point I can't keep dancing around my lack of understanding. Any help is appreciated.
I've had exactly the same problem while trying to copy and paste from and to Bigquery while taking Google's Data Analytics course.
This issue of not being able to paste only happens to me when I compose new query and line 1 is blank.
One way to get around it that works for me is to type anything on line 1 and to paste your text on line 2 and onwards. Delete line 1 and you should be able to run it without any issues.
Use Shortcut commands Ctrl+V - or command+V,to paste the text into BigQuery after copying it.
cheers.
Related
Long story short, I am dealing with an excel files, which need to be modified a little bit. As the files are coming on weekly basis, I decided to write a simple program via Access, which will help me to make the process fully automatic.
The first step was to upload the excel file into an Access database. I managed to achieve that by creating a custom function and inside to just use the "DoCMD.TransferSpreadsheet acImport" aproach.
The second step was to create two queries and update the table that I just uploaded. That was also pretty straight forward too.
However, the third step is what I am struggling with. Now when the table is updated I wanted to export it back to .xlsx format. However, when I do that no matter if I do it manually via the "External Data" tab or simply use "DoCMD.TransferText acExport" approach I noticed that a few columns that have a space after the end of the string are trimmed automatically. For example, original:"string ", but after exporting it is changed to "string".
I would be really grateful if someone can tell me how to specify to Access that the space after the string is intended and not done by mistake? Preferably with a VBA solution than having to do it manually. Thank you in advance for the help!
PS: I know that .CSV format would be way better, but sadly I need it to be in a XLSX format.
I've been working on perfecting a couple of my google sheets for work. I need to share it with my coworkers, but I want to find a way to protect my hard work from being copied and repurposed by others.
They need to have edit access, so that eliminates "Disable options to down, print, and copy for commenters and viewers".'
I need to make sure this document doesn't get duplicated a dozen times by everyone the moment I provide them all with edit access. Any advice?
For any of those coming back to this feed, I've created a google script that if someone tries to make a copy of my document, the moment they open their doc, the entire sheet gets wiped (what I like to call 'self destruct').
So it is possible, despite the entire internet and google themselves saying it's not.
I have some code running in NodeJs using the standard googleapis client which has been working fine for months but starting last night when I perform a simple read on a range of cells in one of my spreadsheets I've been getting an internal error returned from the api with the message 'The service is currently unavailable.' I also get the error when attempting the same read through the explorer interface on the Google developer website.
I've now copied all the data to a second spreadsheet which is now, as far as I can see, identical to the first and from that I can both read and write data normally with no problems.
Has anyone come across a similar issue and can explain what the problem is and whether I can avoid it happening again in the future?
I have some scripts running from GSheet getting data from BigQuery. However, in order to make the files run, I need to manually enable the API every time for a given sheet.
So the question is: How to enable API within the code, so that if I share the GSheet or make a copy I don't have to go to the script editor and enable the API from there?
Thanks
I am a huge fan of this particular use of the Google ecosystem, so I'm happy to help get others up and running using GSheets with BigQuery! Hopefully it is working well for you!
When sharing the sheet with others, there is no need to alter anything in the script editor at all. The scripts should run and query BigQuery without issue; this has been my experience at least. The obvious caveat to this is that the users you share it with must have access to the Google Developer Project that the BigQuery instance is associated with.
However, when copying the sheet, I do not believe it is possible to have it replicate the connection. This is because when the file is copied, it becomes associated with a new Google Developer Project. Thus, you have to go into the script editor, then go to Resources > Developers Console Project and change the project listed to the one in which you have BigQuery enabled.
Hopefully this helps! Sorry I don't have better news for you!
I'm trying to load data from my database into an excel file of a standard template. The package is ready and it's running, throwing a couple of validation warnings stating that truncation may occur because my template has fields of a slightly smaller size than the DB columns i've matched them to.
However, no data is getting populated to my excel sheet.
No errors are reported, and when I click preview for my OLE DB source, it's showing me rows of results. None of these are getting populated into my excel sheet though.
You should first make sure that you have data coming through the pipeline. In the arrow connecting your Source task to Destination task (I'm assuming you don't have any steps between), double click and you'll open the Data Flow Path Editor. Click on Data Viewer, then Add and click OK. That will allow you to see what is moving through the pipeline.
Something to consider with Excel is that is prefers Unicode data types to Non-Unicode. Chances are you have a database collation that is Non-Unicode, so you might have to convert the values in a Data Conversion task.
ALSO, you may need to force the package to execute in 32bit runtime. The VS application develops in a 32bit environment, so the drivers you have visibility to are 32bit. If there is no 64bit equivalent, it will break when you try and run the package. Right click on your project and click Properties and under the Debug menu you'll need to change the setting Run64BitRuntime to FALSE.
you dont provide much informatiom. Add a Data View between your source and your excel destination to see if data is passing through. Do do it, just double click the data flow path, select data view and then add a grid.
Run your app. If you see data, provide more details so we can help you
Couple of questions that may lead to an answer:
Have you checked that data is actually passed through the SSIS package at run time?
Have you double checked your mapping?
Try converting within the package so you don't have the truncation issue
If you add some more details about what you're running, I may be able do give a better answer.
EDIT: Considering what you wrote in your comment, I'd defiantly try the third option. Let us know if this doesn't solve the problem.
Just as an assist for anyone else running into this - I had a similar issue and beat my head against the wall for a long time before I found out what was going on. My export WAS writing data to the file, but because I was using a template file as the destination, and that template file had previous data that had been deleted, the process was appending the data BELOW the previously used rows. So, I was writing out three lines of data, for example, but the data did not start until row 344!!!
The solution was to select the entire spreadsheet in my template file, and delete every bit of it so that I had a completely clean sheet to begin with. I then added my header lines to the clean sheet and saved it. Then I ran the data flow task and...ta-daa!!! Perfect export!
Hopefully this will help some poor soul who runs into this same issue in the future!