使用BigQuery分析Dataproc serverless任务
update the script according to your project, e.g. project_id
, region
./get_dataproc_batches.sh > data.json
go build main.go
now you should see main
executable in the project directory
./main --input data.json --output data_new.json
Go to the BigQuery Console
- create the dataset if there is none
- create table by
upload
filedata_new.json
, let BigQuery auto detect the schema
Click Run in Colab Enterprise
then import the notebook into BigQuery.
Update the TABLE_ID
accordingly to which table you uploaded to from previous
step.
(optional) you may need to change the code accordingly to your job configurations.