Headers are not picking up from the Google sheet in Google big query - google-bigquery

I am trying to upload a google sheet to my dataset in the google bigquery. Below are the settings
Uploading configuration snap
I have made the "Header rows to skip" = 1 and =0 as well
Result doesn't pick the header on the top. It shows "string_field_0"
Here is the file with just one header column and 3 values https://docs.google.com/spreadsheets/d/1mnYBiDe-eOauhVa5QbM-BRelW2iwlhQD9lPMR50A9Lk/edit?usp=sharing
What am I doing wrong?

I was getting the same error. Then I realized that I got it because all the columns are in string format. After adding a field in integer format the issue was resolved.

It is important to follow the steps from the official documentation.
We need to take care of:
Retrieving the Google Drive URI by getting shareable link.
Enabling Google Drive access, because accessing data hosted within Google Drive requires an additional OAuth scope. We can use following command to authenticate with Google Drive:
gcloud auth login --enable-gdrive-access
Check if we are granted with required permissions and scopes. Finally, we can create a table.
Keep in mind that for JSON or CSV files, you can check the Auto-detect option to enable schema auto-detect. Schema auto-detection in the web UI is not currently supported for Google Sheets (though it is available using the bq command-line tool and the API).
I hope you find the above pieces of information useful.

Related

How to access Bigquery using key

I was given a key which happens to be a .json file to access a bigquery data but I have no idea where to put it and how I should use it. I tried to go to the bigquery console but I can't seem to find where I can place the key to view their data. I have no experience using bigquery so I tried to search for any tutorials to no avail.
I can assume that you have created service account key with assigned roles (i.e. roles/bigquery.admin) and downloaded a JSON file that contains your key.
You will need to use it only whenever you decide to use BigQuery API by using client libraries, such as Python or Java. As you can see in the documentation, you need to set the environment variable GOOGLE_APPLICATION_CREDENTIALS to the path of the JSON file that contains your service account key to be able to access Bigquery resources.
When using the web UI in the Google Cloud Console, you don't need to use JSON key file. You only need to take care of assigning appropriate roles to the service account you have created. Please take a look for the following documentation.
Additionally, I would like to share with you the introduction to authentication, which is really important.
I hope you find the above pieces of information useful.

Are there APIs for classifying data on Office 365? Specifically OneDrive and Sharepoint Online?

I am looking to build an application that among other things can connect to a customer's O365 account (using their admin credentials), get directory and file and folder info (metadata), classify the files to find sensitive data (like SSN, PII, etc) and remediate problems.
Is there any way - using APIs - to connect to OneDrive, apply a classification profile (for example , give a regular expression that matches SSN) and get a response with the list of files that match that criterion?
Put another way, how can I classify data on OneDrive?
Thanks,
- deepak
You can use Microsoft's Graph API to get the list of files.
You can see an example of using it in a project I wrote
https://github.com/maximrub/SharePointOnlineTasker
You can use the Graph API to get the files info (hashes, names, size and so on..) and you can also use it to download the content.
If you can classify by metadata only (file names etc'..) you won't need to download it

Cloud Files API with Indexable Meta Data

I would like to use a cloud file system that supports adding meta-data that is searchable. I want to use this meta-data to store keys from my application to associate the document.
E.g.
File:
/XYZ/image.png
Meta Data:
Person-Id:12345
Group-Id:23456
Other-Id:3456
I would then like to use a API to search (v.fast) documents by Person-Id or Group-Id. I understand that I could create this table mapping myself (in mysql within the app) but is there a cloud files solution (google drive, rackspace, amazon) that supports this use case already?
Thanks
You can make use of the Directory API of Google to insert data onto the Google Drive which is hosted on cloud.
Send a POST request like this:
POST https://www.googleapis.com/admin/directory/v1/groups/groupKey/members
The require parameters include the groupKey which Identifies the group in the API request. The value can be the group's email address, group alias, or the unique group ID.
You can directly try this with this official documentation.
For inserting a single member visit this official documentation.

How can I obtain an API-key for my Fusion Tables

Yep, newbie question here, but it's bothering me for some days now, trying to read all the docs on google developer site, but I'm spinning in circles.
I've created a Fusion Table and set the access to 'public' and got an ID.
According to Goolge I should have an API key to access the data from a REST-call. Google suggests:
Go to the Google Developers Console.
Select a project, or create a new one.
In the sidebar on the left, expand APIs & auth. Next, click APIs. In the list of APIs, make sure the status is ON for the Fusion Tables API.
In the sidebar on the left, select Credentials.
I can do that all I've got an API-key, but how does this relates to the Fusion Table I've created? Can I use that API key for
this is really simple API key give the ability to do the most of mysql request type SELECT,INSERT,UPDATE,DELETE using GET and POST also PUT request , for GET you can use the navigator for that but the most effective way for your case is the use of curl librairie for php or jaira for java ... so you can send post or put request with a simple script.
So, what you can do with fusion table is automating the process of manipulating data and the option to share those data with someone else.
Edit: procedures changed since this post. Your mileage may vary
Head to the Google Developers Console
Create a project
Under Explore other services click "Enable APIs and get credentials like keys"
Search for Fusion Tables
Enable Fusion Tables API as a service under APIs & Auth --> APIs
You probably want the browser key. Grab the API key.
Happy Mapping...
API-keys are not related to specific Fusion Tables, they are related to projects.
You may use the key to request data from any public and downloadable FusionTable(not only your own Tables ), the key basically is used to identify your project(google-account) .
So when you have problems with requesting data from a public table, check if the table is downloadable too(click on the table-name on top-left->reuse access->allow downloads ).

Hybris hMC login configuration

Forgive me here if this is a complete newbie question. At work, they are (as I am as well) trying to onboard me into using Hybris. While the documentation on the wiki.hybris site is not well placed, most of the information is there. I am however having some trouble finding how to change default HMC logins credential?
When I rebuilt the server, it forced me to reinitialize the database, and thus changed all of the logins. I managed to find the cms login, but I am curious as to where the HMC admin login is stored at?, as it appears to be changed, and I need to find it. I know that it heavily leverages Spring, and I searched the .xml files for a password, but I am not finding what I need.
Any help would be greatly appreciated!
they are exist in different location in .impex files (importer files) , that you initialize the store for the first time , those .impex files got imported to your database
example of one location :
ext-template\yacceleratorinitialdata\resources\yacceleratorinitialdata\import\common\user-groups.impex
as there are multiple starter stores that comes with hybris (accelerator, telcoaccelerator, and powershop b2b) I suggest you to search as text for username or password in all files extention = .impex then change for files that belong to your store .
i did a quick search , not only in .impex file , the hmc admin/nimda seems to changes from
/bin/platform/project.properties
# Login and password for the automatic logging into the hMC
hmc.default.login=admin
hmc.default.password=nimda
hope that helped you .
Thanks
When searching for Out Of The Box impex files, search against ".impex" & ".csv" as many of the impex scripts are written as csv as well. In rare case, you may also find *.txt also catering to impex scripts.
You will not be able to find an impex file where "admin" user credentials are maintained. It is "nimda" by default and may be changed via impex file or simply by HMC under users.
/bin/platform/project.properties
Any property file will have no impact on the user credentials, the mentioned property file just defaults the jsp page with filled in values in the login form. This has nothing to do with the current/changed credentials.