In Google cloud platform - Dataflow, when streaming unbounded PCollection (say from PubSub topic using PubSubIO), is there an efficient way to start and stop the beam pipeline in Dataflow? (Example running at start of day and ending at end of day) Is the only way to have a scheduler to have a Cron App engine service and to start the above pipeline job and then stop the job? Just looking at if there are any other options out there.
Also, in case if I choose windowing for the unbounded PCollection(say from PubSub), is there a way to have the files written in a configurable directory say. hourly directory for every window? I see it creates one file for every window.