DataFlow Bulk Decompress Cloud Storage Files

Kenneth Law
Dec 8, 2021

Cloud storage files are needed to decompress. DataFlow has a API template to do it. Here is a showcase. I want to unzip cloud files based on a date on url and then store into another cloud storage. For example, I input “2020–01–01” and unzup this date file.

  • PROJECT_ID: the Cloud project ID where you want to run the Dataflow job
  • JOB_NAME: a unique job name of your choice
  • LOCATION: the regional endpoint where you want to deploy your Dataflow job—for example, us-central1
  • VERSION: the version of the template that you want to use

Here is the example after input.

Before do this job, you need access token for your cloud storage authentication OAuth 2.0.

Refer to <https://www.youtube.com/watch?v=tGDn3V-mIOM&t=1097s>

--

--