I had two worker role instances out of six total in a cloud service stop by themselves, apparently for no reason. In my logs I discovered that a third instance went down one hour earlier (to the minute), for 25 minutes, and then started again. What could have caused this to happen?
The cloud service is four different worker role projects, two of them has two instances each. They all work with Service Bus and either Storage Table, SQL Database or both. The instances that stopped were all from different projects. They have worked fine before and this particular version has been running until now for six days without any problems. I checked Windows Azure Service Dashboard and everything looks fine there. They were not in recycling mode like when there's an unhandled error.
I have uploaded a new version (I was going to anyway, almost the same) and put two instances for each role. Nothing has stopped so far.
In the management portal I don't see a way that someone could do that without stopping the whole service (all instances). Does anyone have an idea to how this could happen?