Bigquery load from Avro gives can not convert from long to int - google-bigquery
I am trying to load the avro file from google storage to Big query tables but faced these issue.
Steps i have followed are as below.
Create a dataframe in spark.
Stored these data by writing it into avro.
dataframe.write.avro("path")
Loaded these data into google storage.
Tried to load the data into google bigquery by using following command
bq --nosync load --autodetect --source_format AVRO datasettest.testtable gs://test/avrodebug/*.avro
This command leads to give this error.
Error while reading data, error message: The Apache Avro library failed to read data with the follwing error: Cannot resolve: "long" with "int"
So i even tried to use this command by specifying the schema.
bq --nosync load --source_format AVRO datasettest.testtable gs://test/avrodebug/*.avro C1:STRING, C2:STRING, C3:STRING, C4:STRING, C5:STRING, C6:INTEGER, C7:INTEGER, C8:INTEGER, C9:STRING, C10:STRING, C11:STRING
Here i have only C6,C7 and C8 are having integer values.
Even this also giving the same previous error.
Is there any reason why i am getting error for long to int instead of long to INTEGER
Please let me know is there any way to load these data by casting it.
Related
(InternalError) when calling the SelectObjectContent operation in boto3
I have a series of files that are in JSON that need to be split into multiple files to reduce their size. One issue is that the files are extracted using a third party tool and arrive as a JSON object on a single line. I can use S3 select to process a small file (say around 300Mb uncompressed) but when I try and use a larger file - say 1Gb uncompressed (90Mb gzip compressed) I get the following error: [ERROR] EventStreamError: An error occurred (InternalError) when calling the SelectObjectContent operation: We encountered an internal error. Please try again. The query that I am trying to run is: select count(*) as rowcount from s3object[*][*] s I can't run the query from the console because the file is larger than 128Mb but the code that is performing the operation is as follows: def execute_select_query(bucket, key, query): """ Runs a query against an object in S3. """ if key.endswith("gz"): compression = "GZIP" else: compression = "NONE" LOGGER.info("Running query |%s| against s3://%s/%s", query, bucket, key) return S3_CLIENT.select_object_content( Bucket=bucket, Key=key, ExpressionType='SQL', Expression=query, InputSerialization={"JSON": {"Type": "DOCUMENT"}, "CompressionType": compression}, OutputSerialization={'JSON': {}}, )
"Error while reading data" error received when uploading CSV file into BigQuery via console UI
I need to upload a CSV file to BigQuery via the UI, after I select the file from my local drive I specify BigQuery to automatically detect the Schema and run the job. It fails with the following message: "Error while reading data, error message: CSV table encountered too many errors, giving up. Rows: 2; errors: 1. Please look into the errors[] collection for more details." I have tried removing the comma in the last column, and tried changing options in the advanced section but it always results in the same error. The error log is not helping me understand where the problem is, this is example of the error log entry: 2 019-04-03 23:03:50.261 CLST Bigquery jobcompleted bquxjob_6b9eae1_169e6166db0 frank#xxxxxxxxx.nn INVALID_ARGUMENT and: "Error while reading data, error message: CSV table encountered too many errors, giving up. Rows: 2; errors: 1. Please look into the errors[] collection for more details." and: "Error while reading data, error message: Error detected while parsing row starting at position: 46. Error: Data between close double quote (") and field separator." The strange thing is that the sample CSV data has NO double quote field separator!? 2019-01-02 00:00:00,326,1,,292,0,,294,0,,-28,0,,262,0,,109,0,,372,0,,453,0,,536,0,,136,0,,2609,0,,1450,0,,352,0,,-123,0,,17852,0,,8528,0 2019-01-02 00:02:29,289,1,,402,0,,165,0,,-218,0,,150,0,,90,0,,263,0,,327,0,,275,0,,67,0,,4863,0,,2808,0,,124,0,,454,0,,21880,0,,6410,0 2019-01-02 00:07:29,622,1,,135,0,,228,0,,-147,0,,130,0,,51,0,,381,0,,428,0,,276,0,,67,0,,2672,0,,1623,0,,346,0,,-140,0,,23962,0,,10759,0 2019-01-02 00:12:29,206,1,,118,0,,431,0,,106,0,,133,0,,50,0,,380,0,,426,0,,272,0,,63,0,,1224,0,,740,0,,371,0,,-127,0,,27758,0,,12187,0 2019-01-02 00:17:29,174,1,,119,0,,363,0,,59,0,,157,0,,67,0,,381,0,,426,0,,344,0,,161,0,,923,0,,595,0,,372,0,,-128,0,,22249,0,,9278,0 2019-01-02 00:22:29,175,1,,119,0,,301,0,,7,0,,124,0,,46,0,,382,0,,425,0,,431,0,,339,0,,1622,0,,1344,0,,379,0,,-126,0,,23888,0,,8963,0 I shared an example of a few lines of CSV data. I expect BigQuery to be able to detect the schema and load the data into a new table.
Using BigQuery new WebUI and your input data I did the following: Select a dataset Clicked on create a table Filled the create table form as follow: The table was created and I was able to SELECT 6 rows as expected SELECT * FROM projectId.datasetId.SO LIMIT 1000
Hi , Google big query - bq fail load display file number how to get the file name
I'm running the following bq command bq load --source_format=CSV --skip_leading_rows=1 --max_bad_records=1000 --replace raw_data.order_20150131 gs://raw-data/order/order/2050131/* order.json and getting the following message when loading data into bq . ************************************* Waiting on bqjob_r4ca10491_0000014ce70963aa_1 ... (412s) Current status: DONE BigQuery error in load operation: Error processing job 'orders:bqjob_r4ca10491_0000014ce70963aa_1': Too few columns: expected 11 column(s) but got 1 column(s). For additional help: http://goo.gl/RWuPQ Failure details: - File: 844 / Line:1: Too few columns: expected 11 column(s) but got 1 column(s). For additional help: http://goo.gl/RWuPQ ********************************** The message display only the file number . checked the files content most of them are good . gsutil ls and the cloud console on the other hand display file names . how can I know which file is it according to the file number?
There seems to be some weird spacing introduced in the question, but if the desired path to ingest is "/order.json" - that won't work: You can only use "" at the end of the path when ingesting data to BigQuery.
Error in BQ shell Loading Datastore with write_disposition as Write append
1: I tried to load on an existing table [using Datastore file] 2. Bq Shell asked me to add write_disposition to write append to load to existing table 3. If I do the above, throws an error as follows: load --source_format=DATASTORE_BACKUP --write_disposition=WRITE_append --allow_jagged_rows=None sample_red.t1estchallenge_1 gs://test.appspot.com/bucket/ahFzfnZpcmdpbi1yZWQtdGVzdHJBCxIcX0FFX0RhdGFzdG9yZUFkbWluX09wZXJhdGlvbhiBwLgCDAsSFl9BRV9CYWNrdXBfSW5mb3JtYXRpb24YAQw.entity.backup_info Error parsing command: flag --allow_jagged_rows=None: ('Non-boolean argument to boolean flag',None) I tried allow jagged rows = 0 and allow jagged rows = None, nothing works just the same error. Please advise on this. UPDATE: As Mosha suggested --allow_jagged_rows=false has worked. It should be before --write_disposition=Write_truncate. But this has led to another issue on encoding. Can anyone say what should be the encoding type for DATASTORE_BACKUP?. I tried both --encoding=UTF-8 and --encoding=ISO-8859. load --source_format=DATASTORE_BACKUP --allow_jagged_rows=false --write_disposition=WRITE_TRUNCATE sample_red.t1estchallenge_1 gs://test.appspot.com/STAGING/ahFzfnZpcmdpbi1yZWQtdGVzdHJBCxIcX0FFX0RhdGFzdG9yZUFkbWluX09wZXJhdGlvbhiBwLgCDAsSFl9BRV9CYWNrdXBfSW5mb3JtYXRpb24YAQw.entityname.backup_info Please advise.
You should use "false" (or "true") with boolean arguments, i.e. --allow_jagged_rows=false
Unexpected error while loading data
I am getting an "Unexpected" error. I tried a few times, and I still could not load the data. Is there any other way to load data? gs://log_data/r_mini_raw_20120510.txt.gzto567402616005:myv.may10c Errors: Unexpected. Please try again. Job ID: job_4bde60f1c13743ddabd3be2de9d6b511 Start Time: 1:48pm, 12 May 2012 End Time: 1:51pm, 12 May 2012 Destination Table: 567402616005:myvserv.may10c Source URI: gs://log_data/r_mini_raw_20120510.txt.gz Delimiter: ^ Max Bad Records: 30000 Schema: zoneid: STRING creativeid: STRING ip: STRING update: I am using the file that can be found here: http://saraswaticlasses.net/bad.csv.zip bq load -F '^' --max_bad_record=30000 mycompany.abc bad.csv id:STRING,ceid:STRING,ip:STRING,cb:STRING,country:STRING,telco_name:STRING,date_time:STRING,secondary:STRING,mn:STRING,sf:STRING,uuid:STRING,ua:STRING,brand:STRING,model:STRING,os:STRING,osversion:STRING,sh:STRING,sw:STRING,proxy:STRING,ah:STRING,callback:STRING I am getting an error "BigQuery error in load operation: Unexpected. Please try again." The same file works from Ubuntu while it does not work from CentOS 5.4 (Final) Does the OS encoding need to be checked?
The file you uploaded has an unterminated quote. Can you delete that line and try again? I've filed an internal bigquery bug to be able to handle this case more gracefully. $grep '"' bad.csv 3000^0^1.202.218.8^2f1f1491^CN^others^2012-05-02 20:35:00^^^^^"Mozilla/5.0^generic web browser^^^^^^^^ When I run a load from my workstation (Ubuntu), I get a warning about the line in question. Note that if you were using a larger file, you would not see this warning, instead you'd just get a failure. $bq show --format=prettyjson -j job_e1d8636e225a4d5f81becf84019e7484 ... "status": { "errors": [ { "location": "Line:29057 / Field:12", "message": "Missing close double quote (\") character: field starts with: <Mozilla/>", "reason": "invalid" } ]
My suspicion is that you have rows or fields in your input data that exceed the 64 KB limit. Perhaps re-check the formatting of your data, check that it is gzipped properly, and if all else fails, try importing uncompressed data. (One possibility is that the entire compressed file is being interpreted as a single row/field that exceeds the aforementioned limit.) To answer your original question, there are a few other ways to import data: you could upload directly from your local machine using the command-line tool or the web UI, or you could use the raw API. However, all of these mechanisms (including the Google Storage import that you used) funnel through the same CSV parser, so it's possible that they'll all fail in the same way.