I am using bidirectional LSTM in the many-to-one setting (sentiment analysis task) with tflearn. I want to understand how does tflearn aggregate representations from the forward and backward LSTM layers before sending it to the softmax layer to get probabilistic output? For instance, in the following diagram, how are concat and aggregate layers usually implemented?
Is there any documentation available on this?
Thank you!