I have the following scenario:
- Pipeline A looks up table A in BigQuery, does some computation and returns a list of column names.
- This list of columns names is used as the BigQuery schema for output of pipeline B.
Can you please let me know what is the best option to achieve this?
Can pipeline A use TextIO to write the list of column names to temporary or staging location files which are then read by the Pipeline executor to define the schema for pipeline B. If this approach looks fine, can you please let me know if there is a Dataflow utility to read files from temporary or staging location or if the GCS API should be used.