DataFlow Bulk Decompress Cloud Storage Files
Dec 8, 2021
Cloud storage files are needed to decompress. DataFlow has a API template to do it. Here is a showcase. I want to unzip cloud files based on a date on url and then store into another cloud storage. For example, I input “2020–01–01” and unzup this date file.
PROJECT_ID
: the Cloud project ID where you want to run the Dataflow jobJOB_NAME
: a unique job name of your choiceLOCATION
: the regional endpoint where you want to deploy your Dataflow job—for example,us-central1
VERSION
: the version of the template that you want to use
Here is the example after input.
Before do this job, you need access token for your cloud storage authentication OAuth 2.0.
Refer to <https://www.youtube.com/watch?v=tGDn3V-mIOM&t=1097s>