In the Dataflow Monitoring Interface for Beam Pipeline Executions, there is a time duration specified in each of the Transformation boxes (see https://cloud.google.com/dataflow/pipelines/dataflow-monitoring-intf).
For bounded data, I understood this is the estimated time it would take for the transformation to be completed. However, for unbounded data as in my streaming case, how do I interpret this number?
Some of my transforms have a duration significantly higher than the others, and this means that the transform takes more time. But what are the other implications regarding how this uneven distribution affects my execution, especially if I have a windowing functions going on?
Also, is this related to autoscaling? For e.g. do more workers get spun up if the time taken for execution exceeds certain thresholds? Or does autoscaling depend on data volume at the input?