This paper points out the Limitations of the evaluation method in self-supervised learning (SSL) and proposes a new evaluation framework to improve it. The existing fixed benchmark-based evaluation deviates from the ultimate goal of AI research, "solving all possible tasks", and makes researchers spend a lot of effort to find various evaluation tasks. In this paper, we define the probabilistic space of all possible subtasks by introducing task distribution and task priors. This allows us to evaluate the average performance and variance of the model for all possible subtasks. This is expected to evaluate the model performance in all possible subtasks and especially contribute to the advancement of self-supervised learning research.