I would like to calculate a confidence interval for the RMSE of a machine learning regression in the out-of-sample test set predictions.
My train set is the first 80% of the sample, and the "out-of-sample" test set is the last 20% of the sample. I treat the RMSE of the test set predictions as the out-of-sample performance, and would like to calculate a CI of this RMSE.
One idea I had was to resample the train set among the first 80%, but use the same test set each iteration. This would seem to represent the CI of RMSE on the test set across different possible training scenarios. However, it would not account for possible variation in the test set.
Is this approach sensible? Is there a better way to address my question? Thanks!